SEO is also about making your search engine result relevant to the user's search query so more people click the result when it is shown in search. In this process, snippets of text and meta data are optimized to ensure your snippet of information is appealing in the context of the search query to obtain a high CTR (click through rate) from search results.
hey james - congrats on your success here. just a question about removing crummy links. for my own website, there are hundreds of thousands of backlinks in webmaster tools pointing to my site. The site has no penalties or anything  - the traffic seems to be growing every week. would you recommend hiring someone to go through the link profile anyway to remove crummy links that just occur naturally?
Description meta tags are important because Google might use them as snippets for your pages. Note that we say "might" because Google may choose to use a relevant section of your page's visible text if it does a good job of matching up with a user's query. Adding description meta tags to each of your pages is always a good practice in case Google cannot find a good selection of text to use in the snippet. The Webmaster Central Blog has informative posts on improving snippets with better description meta tags18 and better snippets for your users19. We also have a handy Help Center article on how to create good titles and snippets20.
This is such a great article – so many things I want to try. Question: when you talk about creating ‘snippet bait’ for featured paragraph snippets, where is the best place to add that to your content – in the beginning, all throughout? Also, you said lists should be formatted with header tags – what about paragraph snippets? Thanks for all the great advice!
For our client: We were lucky enough to remove most from the prior agency outreach, we also went directly to many webmasters in which we wanted to remove links. We did not use the Disavow tool as it was not around when we completed this link cleanup, but we all know it has been said that if you are going to use the Disavow Tool to use it with caution. is a certified Google Partner, which means the search engine has certified our knowledge and experience of Google applications and tools. In addition, we have access to specialized Google training and products, and a dedicated Google representative. Whether you are in ecommerce, technology, retail, dentistry, or B2B, you will benefit working with
Part of what makes SEO unique, fun and challenging is that we get to be like detectives on the web. We have to be good at pulling data from different places and getting that data no matter what it takes sometimes. Some new ways have come about lately for doing so. We’re going to walk step by step through nine specific ways you can collect data more effectively.
It is well known by now that Google has said site speed is a small ranking factor (about 1%). This is tiny, but it’s rare for Google to say that anything has a definite effect on rankings. So it makes so much sense to follow this advice. Plus, users love fast and responsive sites. They will feel more in control of their experience, consume your amazing content more efficiently and convert better.
With our exclusive community forums you benefit from the experience and wisdom of hundreds of industry professionals who have boosted the sales, profitability, and overall performance of thousands of websites. With search, when you increase traffic you increase profits. This is true for mom and pop websites and large corporations. The traffic scale is only limited by your ambitions and the quality of the information you consume.
Regarding RankBain, my own assumption is that user signals are part of the training data RankBrain gets (even though Paul Haahr does not confirm that in the talk at SMX or the discussion afterwards). If that is true, then RankBrain will see your high CTR and maybe TOS, might try to figure out what pattern causes them and MIGHT try to change it’s own algorithm in a way that ranks results LIKE YOURS higher.
Loved the bit on the Youtube optimization and how to get the words to catch people and keep them engaged. My average time on my site at the moment is 1min 19 seconds 🙁 So dwell time is going to be my goal so that I can increase my DA from 16 🙂 goal is 25 so I have a long way to go — but hoping it will come. Podcasts is an interesting thought – have never thought about doing one.
That helped explain some of the organic traffic loss, but knowing that this client had gone through a few website redesigns, I wanted to make sure that all redirects were done properly. Regardless of whether or not your traffic has changed, if you’ve recently done a website redesign where you’re changing URLs, it’s smart to look at your top organic landing pages from before the redesign and double check to make sure they’re redirecting to the correct pages.
Keep resources crawlable. Blocking page resources can give Google an incomplete picture of your website. This often happens when your robots.txt file is blocking access to some or all of your page resources. If Googlebot doesn't have access to a page's resources, such as CSS, JavaScript, or images, we may not detect that it's built to display and work well on a mobile browser. In other words, we may not detect that the page is "mobile-friendly," and therefore not properly serve it to mobile searchers.
That helped explain some of the organic traffic loss, but knowing that this client had gone through a few website redesigns, I wanted to make sure that all redirects were done properly. Regardless of whether or not your traffic has changed, if you’ve recently done a website redesign where you’re changing URLs, it’s smart to look at your top organic landing pages from before the redesign and double check to make sure they’re redirecting to the correct pages.
First, I will show you a quick snapshot of the traffic uplift, which yielded an additional 400,000 unique visitors from organic search traffic on a monthly basis. Then I will explain to you the steps in which we took to get the client to this level. I have also tried to keep this quite general so everyone can adapt their own situation to this case study.
We want to see landing pages that came from organic searches, so first we need to add to this dataset the parameter “Medium” which is how Analytics identifies channels. To do this, use the drop down above the table of data and locate the option for “Medium”. The table below should refresh and now you will have a second column of data showing the channel for each landing page.
Google claims their users click (organic) search results more often than ads, essentially rebutting the research cited above. A 2012 Google study found that 81% of ad impressions and 66% of ad clicks happen when there is no associated organic search result on the first page.[2] Research has shown that searchers may have a bias against ads, unless the ads are relevant to the searcher's need or intent [3]
SEO is not an appropriate strategy for every website, and other Internet marketing strategies can be more effective like paid advertising through pay per click (PPC) campaigns, depending on the site operator's goals. Search engine marketing (SEM), is practice of designing, running, and optimizing search engine ad campaigns.[55] Its difference from SEO is most simply depicted as the difference between paid and unpaid priority ranking in search results. Its purpose regards prominence more so than relevance; website developers should regard SEM with the utmost importance with consideration to PageRank visibility as most navigate to the primary listings of their search.[56] A successful Internet marketing campaign may also depend upon building high quality web pages to engage and persuade, setting up analytics programs to enable site owners to measure results, and improving a site's conversion rate.[57] In November 2015, Google released a full 160 page version of its Search Quality Rating Guidelines to the public,[58] which now shows a shift in their focus towards "usefulness" and mobile search. In recent years the mobile market has exploded, overtaking the use of desktops as shown in by StatCounter in October 2016 where they analysed 2.5 million websites and 51.3% of the pages were loaded by a mobile device [59]. Google has been one of the companies that have utilised the popularity of mobile usage by encouraging websites to use their Google Search Console, the Mobile-Friendly Test, which allows companies to measure up their website to the search engine results and how user-friendly it is.
Website owners recognized the value of a high ranking and visibility in search engine results,[6] creating an opportunity for both white hat and black hat SEO practitioners. According to industry analyst Danny Sullivan, the phrase "search engine optimization" probably came into use in 1997. Sullivan credits Bruce Clay as one of the first people to popularize the term.[7] On May 2, 2007,[8] Jason Gambert attempted to trademark the term SEO by convincing the Trademark Office in Arizona[9] that SEO is a "process" involving manipulation of keywords and not a "marketing service."
To avoid undesirable content in the search indexes, webmasters can instruct spiders not to crawl certain files or directories through the standard robots.txt file in the root directory of the domain. Additionally, a page can be explicitly excluded from a search engine's database by using a meta tag specific to robots (usually ). When a search engine visits a site, the robots.txt located in the root directory is the first file crawled. The robots.txt file is then parsed and will instruct the robot as to which pages are not to be crawled. As a search engine crawler may keep a cached copy of this file, it may on occasion crawl pages a webmaster does not wish crawled. Pages typically prevented from being crawled include login specific pages such as shopping carts and user-specific content such as search results from internal searches. In March 2007, Google warned webmasters that they should prevent indexing of internal search results because those pages are considered search spam.[46] Organic traffic