9. Troubleshooting and adjustment. In your first few years as a search optimizer, you’ll almost certainly run into the same problems and challenges everyone else does; your rankings will plateau, you’ll find duplicate content on your site, and you’ll probably see significant ranking volatility. You’ll need to know how to diagnose and address these problems if you don’t want them to bring down the effectiveness of your campaign.
The Featured Snippet section appearing inside the first page of Google is an incredibly important section to have your content placed within. I did a study of over 5,000 keywords where HubSpot.com ranked on page 1 and there was a Featured Snippet being displayed. What I found was that when HubSpot.com was ranking in the Featured Snippet, the average click-through rate to the website increased by over 114%.

You should optimize your site to serve your users' needs. One of those users is a search engine, which helps other users discover your content. Search Engine Optimization is about helping search engines understand and present content. Your site may be smaller or larger than our example site and offer vastly different content, but the optimization topics we discuss below should apply to sites of all sizes and types. We hope our guide gives you some fresh ideas on how to improve your website, and we'd love to hear your questions, feedback, and success stories in the Google Webmaster Help Forum1.

Tablet - We consider tablets as devices in their own class, so when we speak of mobile devices, we generally do not include tablets in the definition. Tablets tend to have larger screens, which means that, unless you offer tablet-optimized content, you can assume that users expect to see your site as it would look on a desktop browser rather than on a smartphone browser.
Direct traffic is defined as visits with no referring website. When a visitor follows a link from one website to another, the site of origin is considered the referrer. These sites can be search engines, social media, blogs, or other websites that have links to other websites. Direct traffic categorizes visits that do not come from a referring URL.
This is such a great article – so many things I want to try. Question: when you talk about creating ‘snippet bait’ for featured paragraph snippets, where is the best place to add that to your content – in the beginning, all throughout? Also, you said lists should be formatted with header tags – what about paragraph snippets? Thanks for all the great advice!
The response rate here was huge because this is a mutually beneficial relationship. The bloggers get free products to use within their outfits (as well as more clothes for their wardrobe!) and I was able to drive traffic through to my site, get high-quality backlinks, a load of social media engagement and some high-end photography to use within my own content and on product pages.

If you are using Responsive Web Design, use meta name="viewport" tag to tell the browser how to adjust the content. If you use Dynamic Serving, use the Vary HTTP header to signal your changes depending on the user-agent. If you are using separate URLs, signal the relationship between two URLs by tag with rel="canonical" and rel="alternate" elements.
Second and last, keywords. I use SEO Ultimate on WordPress and I sometimes doubt about what words and sentences to put in the “Tags” box. Does it need to be very specific or broad words are good to have too ? Like if I’m writing about Porsche beating the previous record on the Nurburgring with the new GT2 RS, can I put “Nurburgring” and “Porsche GT2 RS” as tags or is better to just keep “Porsche record nurburgring” and specific tags like this one ?
Webmasters and content providers began optimizing websites for search engines in the mid-1990s, as the first search engines were cataloging the early Web. Initially, all webmasters needed only to submit the address of a page, or URL, to the various engines which would send a "spider" to "crawl" that page, extract links to other pages from it, and return information found on the page to be indexed.[5] The process involves a search engine spider downloading a page and storing it on the search engine's own server. A second program, known as an indexer, extracts information about the page, such as the words it contains, where they are located, and any weight for specific words, as well as all links the page contains. All of this information is then placed into a scheduler for crawling at a later date.
As an Internet marketing strategy, SEO considers how search engines work, the computer programmed algorithms which dictate search engine behavior, what people search for, the actual search terms or keywords typed into search engines, and which search engines are preferred by their targeted audience. Optimizing a website may involve editing its content, adding content, doing HTML, and associated coding to both increase its relevance to specific keywords and to remove barriers to the indexing activities of search engines. Promoting a site to increase the number of backlinks, or inbound links, is another SEO tactic. By May 2015, mobile search had surpassed desktop search.[3] In 2015, it was reported that Google is developing and promoting mobile search as a key feature within future products. In response, many brands are beginning to take a different approach to their Internet marketing strategies.[4]
So if you think about it, SEO is really just a process of proving to search engines that you are the best site, the most authoritative, the most trusted, the most unique and interesting site that they can offer to their customer - the searcher. Get people to talk about you, produce good quality content, get people to link to you, and Google will be more confident that you are the best result that they can offer to their searchers, and that’s when you will start ranking on the first page of Google.
Tablet - We consider tablets as devices in their own class, so when we speak of mobile devices, we generally do not include tablets in the definition. Tablets tend to have larger screens, which means that, unless you offer tablet-optimized content, you can assume that users expect to see your site as it would look on a desktop browser rather than on a smartphone browser.
The first step to digging into organic traffic is to analyze what content on your website is performing best in this area. For obvious reasons, the homepage is almost certainly the landing page for most organic traffic, but the other top pages are often revealing. To view this data, we’re going to head over to the Behaviour section in the Analytics sidebar, then choose Site Content and finally Landing Pages.

Black hat SEO attempts to improve rankings in ways that are disapproved of by the search engines, or involve deception. One black hat technique uses text that is hidden, either as text colored similar to the background, in an invisible div, or positioned off screen. Another method gives a different page depending on whether the page is being requested by a human visitor or a search engine, a technique known as cloaking. Another category sometimes used is grey hat SEO. This is in between black hat and white hat approaches, where the methods employed avoid the site being penalized, but do not act in producing the best content for users. Grey hat SEO is entirely focused on improving search engine rankings.
To sum up all of this information, even organic traffic, like direct traffic, has some gray areas. For the most part, though, organic traffic is driven by SEO. The better you are ranking for competitive keywords, the more organic traffic will result. Websites that consistently create content optimized for search will see a steady increase in organic search traffic and improved positioning in the search results. As a marketer, it is important to look at your keywords and high-ranking pages to identify new SEO opportunities each month.  
Search engine optimization (SEO) is the process of affecting the online visibility of a website or a web page in a web search engine's unpaid results—often referred to as "natural", "organic", or "earned" results. In general, the earlier (or higher ranked on the search results page), and more frequently a website appears in the search results list, the more visitors it will receive from the search engine's users; these visitors can then be converted into customers.[1] SEO may target different kinds of search, including image search, video search, academic search,[2] news search, and industry-specific vertical search engines. SEO differs from local search engine optimization in that the latter is focused on optimizing a business' online presence so that its web pages will be displayed by search engines when a user enters a local search for its products or services. The former instead is more focused on national or international searches.
Website owners recognized the value of a high ranking and visibility in search engine results,[6] creating an opportunity for both white hat and black hat SEO practitioners. According to industry analyst Danny Sullivan, the phrase "search engine optimization" probably came into use in 1997. Sullivan credits Bruce Clay as one of the first people to popularize the term.[7] On May 2, 2007,[8] Jason Gambert attempted to trademark the term SEO by convincing the Trademark Office in Arizona[9] that SEO is a "process" involving manipulation of keywords and not a "marketing service."
Though a long break is never suggested, there are times that money can be shifted and put towards other resources for a short time. A good example would be an online retailer. In the couple of weeks leading up to the Christmas holidays, you are unlikely to get more organic placement than you already have. Besides, the window of opportunity for shipping gifts to arrive before Christmas is ending, and you are heading into a slow season.

Though a long break is never suggested, there are times that money can be shifted and put towards other resources for a short time. A good example would be an online retailer. In the couple of weeks leading up to the Christmas holidays, you are unlikely to get more organic placement than you already have. Besides, the window of opportunity for shipping gifts to arrive before Christmas is ending, and you are heading into a slow season.
The length of a title tag that Google will show will vary (it’s based on pixels, not character counts) but in general 55-60 characters is a good rule of thumb here. If possible you want to work in your core keyword, and if you can do it in a natural and compelling way, add some related modifiers around that term as well. Keep in mind though: the title tag will frequently be what a searcher sees in search results for your page. It’s the “headline” in organic search results, so you also want to take how clickable your title tag is into account.
When Googlebot crawls a page, it should see the page the same way an average user does15. For optimal rendering and indexing, always allow Googlebot access to the JavaScript, CSS, and image files used by your website. If your site's robots.txt file disallows crawling of these assets, it directly harms how well our algorithms render and index your content. This can result in suboptimal rankings.
would it be easier to set up 2 separate GMAIL Accounts with 2 separate analytics accounts for 2 different web sites? Or is it ok to use 1 GMAIL account to manage 2 sites under 1 Analytics accounts and just have 2 properties inside of it? Take into consideration that it’s a local business doing services (no store front) and might need Adwords etc. Also take into consideration Search console , not sure how it influences Analytics /sites verifications
Regarding Link Detox, links it diagnoses as Toxic are generally fine as they're either not indexed by Google or have malware/viruses/etc., but I recommend a manual review of any diagnosed as Suspicious. I used it recently to get started cleaning up our backlinks and some legit sites and blogs were under Suspicious simply because they didn't have many links pointing to them.
Unfortunately, Google has stopped delivering a lot of the information about what people are searching for to analytics providers. Google does make some of this data available in their free Webmaster Tools interface (if you haven’t set up an account, this is a very valuable SEO tool both for unearthing search query data and for diagnosing various technical SEO issues).

In short, press request alerts are requests for sources of information from journalists. Let's say you're a journalist putting together an article on wearable technology for The Guardian. Perhaps you need a quote from an industry expert or some products that you can feature within your article? Well, all you need to do is send out a request to a press service and you can wait for someone to get back to you.
Like I said at the beginning, building organic traffic is hard. Anything that promises a shortcut to an avalanche of traffic will more than likely lead to a penalty down the road. Embrace the daily grind of creating great content that helps users and provides a solution to what they’re looking for. In the end that will drive more organic traffic than any shortcut ever will.
It increases relevancy: Siloing ensures all topically related content is connected, and this in turn drives up relevancy. For example, linking to each of the individual yoga class pages (e.g. Pilates, Yoga RX, etc) from the “Yoga classes” page helps confirm—to both visitors and Google—these pages are in fact different types of yoga classes. Google can then feel more confident ranking these pages for related terms, as it is clearer the pages are relevant to the search query.
Earlier in the comment stream, there was a brief discussion about page load time/website speed and its effect on page ranking. I have tried to find unbiased information about which hosting company to use when starting a blog or a small WordPress sites, keeping in mind the importance of speed. This endeavor has been harder than expected as most hosting review sites have some kind of affiliate relationship with the hosting companies they review.
Visual assets aren’t regular images you might pull from a Google Image search. Instead, these are unique diagrams or infographics you’ve created specifically for your epic content. These kinds of diagrams or infographics explain a theory, communicate a point, or showcase data in exciting and interesting ways—and gain attention (and links) because of it.
And finally, the other really important bucket is authority. Google wants to show sites that are popular. If they can show the most popular t-shirt seller to people looking to buy t-shirts online, that’s the site they want to show. So you have to convince Google - send them signals that your site is the most popular site for the kind of t-shirts that you sell. Fill this bucket by building a fan base. Build a social network, get people to link to you, get people to share your t-shirt pages on their social network saying ‘I want this!’, get people to comment, leave testimonials, show pictures of themselves wearing the product or using the product, Create a fan-base and then rally them to link to you and talk about you. That’s how you prove to Google that you are trustworthy and authoritative.
Brian, I have become a huge fan of yours over the last 3 months – I make it a point to read one of your articles every 2 weeks to help boost my Life & Career Coach website. Today I was reminded of a few points I had put off and will now implement (such as creating videos), and I learned cool new things such as LSI keywords and great tips about how to keep prospects longer on my site.

The leading search engines, such as Google, Bing and Yahoo!, use crawlers to find pages for their algorithmic search results. Pages that are linked from other search engine indexed pages do not need to be submitted because they are found automatically. The Yahoo! Directory and DMOZ, two major directories which closed in 2014 and 2017 respectively, both required manual submission and human editorial review.[39] Google offers Google Search Console, for which an XML Sitemap feed can be created and submitted for free to ensure that all pages are found, especially pages that are not discoverable by automatically following links[40] in addition to their URL submission console.[41] Yahoo! formerly operated a paid submission service that guaranteed crawling for a cost per click;[42] however, this practice was discontinued in 2009.

Search engines may penalize sites they discover using black hat methods, either by reducing their rankings or eliminating their listings from their databases altogether. Such penalties can be applied either automatically by the search engines' algorithms, or by a manual site review. One example was the February 2006 Google removal of both BMW Germany and Ricoh Germany for use of deceptive practices.[53] Both companies, however, quickly apologized, fixed the offending pages, and were restored to Google's list.[54]
This is such a great article – so many things I want to try. Question: when you talk about creating ‘snippet bait’ for featured paragraph snippets, where is the best place to add that to your content – in the beginning, all throughout? Also, you said lists should be formatted with header tags – what about paragraph snippets? Thanks for all the great advice!
Thank you Brian, this is awesome! About publishing studies, how do you gather all this unique data? How did you get access to behind-the-scenes data from 1.3M videos to analyze? We recently published an infograpghic on a client’s blog but it’s just data we quoted from other sites, not unique. I wonder if you can get your own stats when you have a small site.
Indeed a great post about website traffic. Nowadays It is very much hard for a blogger to drive targeted traffic to their website and without having targeted traffic, We can never drive customer and sales. Getting website traffic is the most important thing for any website. To have high website traffic, We must have to write high quality content which is very much important to hold the readers on our website for long period of time. We have to write engagging content which can help readers. I am glad that You can covered an amazing article on website traffic. Will definitely follow what you said in this article. Thanks for sharing it with us. :D
One of the reasons for a traffic drop can also be due to your site losing links. You may be seeing a direct loss of that referral traffic, but there could also be indirect effects. When your site loses inbound links, it tells Google that your site isn't as authoritative anymore, which leads to lower search rankings that in turn lead to traffic drops (because fewer people are finding your site if it's not ranked as highly and more).
As of 2009, there are only a few large markets where Google is not the leading search engine. In most cases, when Google is not leading in a given market, it is lagging behind a local player. The most notable example markets are China, Japan, South Korea, Russia and the Czech Republic where respectively Baidu, Yahoo! Japan, Naver, Yandex and Seznam are market leaders.
The majority of new websites created today are built upon WordPress. QuickSprout.com is built on WordPress and most of my readers have site using this popular CMS. I find most people get the basics pretty good for WordPress and SEO, but we’re going to take things a bit further in this section. You’ll find a detailed walkthrough of setting up the SEO for WordPress plugin, improving your WordPress speed and performance, creating a custom author page and more.
As of 2009, there are only a few large markets where Google is not the leading search engine. In most cases, when Google is not leading in a given market, it is lagging behind a local player. The most notable example markets are China, Japan, South Korea, Russia and the Czech Republic where respectively Baidu, Yahoo! Japan, Naver, Yandex and Seznam are market leaders.
Some search engines have also reached out to the SEO industry, and are frequent sponsors and guests at SEO conferences, webchats, and seminars. Major search engines provide information and guidelines to help with website optimization.[18][19] Google has a Sitemaps program to help webmasters learn if Google is having any problems indexing their website and also provides data on Google traffic to the website.[20] Bing Webmaster Tools provides a way for webmasters to submit a sitemap and web feeds, allows users to determine the "crawl rate", and track the web pages index status.
Regarding RankBain, my own assumption is that user signals are part of the training data RankBrain gets (even though Paul Haahr does not confirm that in the talk at SMX or the discussion afterwards). If that is true, then RankBrain will see your high CTR and maybe TOS, might try to figure out what pattern causes them and MIGHT try to change it’s own algorithm in a way that ranks results LIKE YOURS higher.
This should be rephrased to: “Satisfying the needs of the searcher in depth”. Followed by an explanation of how different types of content satisfy different needs, but each should do that in an outstanding way. In depth content is great when a searcher was looking for that and often when the intent is not clear from query and context (context as in the context in which the searcher does their search).
So, you have downloaded your links profiles on a CSV and you now have an extensive list of all your linked domains. If you have been doing SEO for 8+ years like me you can probably just know from analysis which links are bad from a TLD and URL point of view. If you do not know too much you can use tools such as Link Detox: http://www.linkdetox.com/ to complete analysis of your link profile. I would always consult the advice of an expert SEO in this instance because it is easy for these tools to mistake good and bad links.
The world is mobile today. Most people are searching on Google using a mobile device. The desktop version of a site might be difficult to view and use on a mobile device. As a result, having a mobile ready site is critical to your online presence. In fact, starting in late 2016, Google has begun experiments to primarily use the mobile version of a site's content41 for ranking, parsing structured data, and generating snippets.
First, I will show you a quick snapshot of the traffic uplift, which yielded an additional 400,000 unique visitors from organic search traffic on a monthly basis. Then I will explain to you the steps in which we took to get the client to this level. I have also tried to keep this quite general so everyone can adapt their own situation to this case study.
For a long time, digital marketers summed up the properties of direct and organic traffic pretty similarly and simply. To most, organic traffic consists of visits from search engines, while direct traffic is made up of visits from people entering your company URL into their browser. This explanation, however, is too simplified and leaves most digital marketers short-handed when it comes to completely understanding and gaining insights from web traffic, especially organic and direct sources.
×