I have always believed in good quality content, well structured and written in a way that isn’t just about promotional talk. Thanks for sharing this information with us, it’s always helpful to have everything written in a concise manner so we can remind ourselves now and again of what it takes to increase organic traffic. As an SEO consultant myself I come across websites all the time that are messy and still using tactics that have long been out of date. Having a successful website is all about quality content and links. I like it that you stated what the problem was and then how you fixed it for the client. Great article.
This community is full of opportunities if you're a fashion-based retailer. One of the major advantages is the fact that they add links to each of the products that they feature within their outfits - the links go directly to product pages. This is the holy grail for ecommerce SEO, and the traffic those links will bring through will convert at a very high rate.

The length of a title tag that Google will show will vary (it’s based on pixels, not character counts) but in general 55-60 characters is a good rule of thumb here. If possible you want to work in your core keyword, and if you can do it in a natural and compelling way, add some related modifiers around that term as well. Keep in mind though: the title tag will frequently be what a searcher sees in search results for your page. It’s the “headline” in organic search results, so you also want to take how clickable your title tag is into account.
5. Link building. In some respects, guest posting – one popular tactic to build links, among many other benefits – is just content marketing applied to external publishers. The goal is to create content on external websites, building your personal brand and company brand at the same time, and creating opportunities to link back to your site. There are only a handful of strategies to build quality links, which you should learn and understand as well.
When would this be useful? If your site has a blog with public commenting turned on, links within those comments could pass your reputation to pages that you may not be comfortable vouching for. Blog comment areas on pages are highly susceptible to comment spam. Nofollowing these user-added links ensures that you're not giving your page's hard-earned reputation to a spammy site.
If you are using Responsive Web Design, use meta name="viewport" tag to tell the browser how to adjust the content. If you use Dynamic Serving, use the Vary HTTP header to signal your changes depending on the user-agent. If you are using separate URLs, signal the relationship between two URLs by tag with rel="canonical" and rel="alternate" elements.
Structured data21 is code that you can add to your sites' pages to describe your content to search engines, so they can better understand what's on your pages. Search engines can use this understanding to display your content in useful (and eye-catching!) ways in search results. That, in turn, can help you attract just the right kind of customers for your business.
Black hat SEO attempts to improve rankings in ways that are disapproved of by the search engines, or involve deception. One black hat technique uses text that is hidden, either as text colored similar to the background, in an invisible div, or positioned off screen. Another method gives a different page depending on whether the page is being requested by a human visitor or a search engine, a technique known as cloaking. Another category sometimes used is grey hat SEO. This is in between black hat and white hat approaches, where the methods employed avoid the site being penalized, but do not act in producing the best content for users. Grey hat SEO is entirely focused on improving search engine rankings.
Link text is the visible text inside a link. This text tells users and Google something about the page you're linking to. Links on your page may be internal—pointing to other pages on your site—or external—leading to content on other sites. In either of these cases, the better your anchor text is, the easier it is for users to navigate and for Google to understand what the page you're linking to is about.

That's why it's necessary to always stay abreast of developments in the SEO world, so that you can see these algorithm updates coming or you can determine what to do once they’ve been released. The WordStream blog is a great resource for SEO updates, but we also recommend Search Engine Land and Search Engine Roundtable for news on updates. Glenn Gabe of G-Squared Interactive is also a great resource for analyzing the causes and impact of algorithm updates.
That helped explain some of the organic traffic loss, but knowing that this client had gone through a few website redesigns, I wanted to make sure that all redirects were done properly. Regardless of whether or not your traffic has changed, if you’ve recently done a website redesign where you’re changing URLs, it’s smart to look at your top organic landing pages from before the redesign and double check to make sure they’re redirecting to the correct pages.
By 2004, search engines had incorporated a wide range of undisclosed factors in their ranking algorithms to reduce the impact of link manipulation. In June 2007, The New York Times' Saul Hansell stated Google ranks sites using more than 200 different signals.[25] The leading search engines, Google, Bing, and Yahoo, do not disclose the algorithms they use to rank pages. Some SEO practitioners have studied different approaches to search engine optimization, and have shared their personal opinions.[26] Patents related to search engines can provide information to better understand search engines.[27] In 2005, Google began personalizing search results for each user. Depending on their history of previous searches, Google crafted results for logged in users.[28]

While the title tag is effectively your search listing’s headline, the meta description (another meta HTML element that can be updated in your site’s code, but isn’t seen on your actual page) is effectively your site’s additional ad copy. Google takes some liberties with what they display in search results, so your meta description may not always show, but if you have a compelling description of your page that would make folks searching likely to click, you can greatly increase traffic. (Remember: showing up in search results is just the first step! You still need to get searchers to come to your site, and then actually take the action you want.)
One of the reasons for a traffic drop can also be due to your site losing links. You may be seeing a direct loss of that referral traffic, but there could also be indirect effects. When your site loses inbound links, it tells Google that your site isn't as authoritative anymore, which leads to lower search rankings that in turn lead to traffic drops (because fewer people are finding your site if it's not ranked as highly and more).
Great guide. 1 thing i would like to mention ( If i may ) is that the importance of having a secure domain ( SSL ) cant be overstated. A recent Semrush survey revealed that over 65% of websites ranking top 3 organically, all had HTTPS domains. If Rankbrain is going to look at bounce rate as a signal then i can’t see any bigger factor than this in terms of having an effect once a user lands on a website, particularly as Google is going to make it crystal clear if a domain is secure or not.
Google recommends that all websites use https:// when possible. The hostname is where your website is hosted, commonly using the same domain name that you'd use for email. Google differentiates between the "www" and "non-www" version (for example, "www.example.com" or just "example.com"). When adding your website to Search Console, we recommend adding both http:// and https:// versions, as well as the "www" and "non-www" versions.
The leading search engines, such as Google, Bing and Yahoo!, use crawlers to find pages for their algorithmic search results. Pages that are linked from other search engine indexed pages do not need to be submitted because they are found automatically. The Yahoo! Directory and DMOZ, two major directories which closed in 2014 and 2017 respectively, both required manual submission and human editorial review.[39] Google offers Google Search Console, for which an XML Sitemap feed can be created and submitted for free to ensure that all pages are found, especially pages that are not discoverable by automatically following links[40] in addition to their URL submission console.[41] Yahoo! formerly operated a paid submission service that guaranteed crawling for a cost per click;[42] however, this practice was discontinued in 2009.
By 2004, search engines had incorporated a wide range of undisclosed factors in their ranking algorithms to reduce the impact of link manipulation. In June 2007, The New York Times' Saul Hansell stated Google ranks sites using more than 200 different signals.[25] The leading search engines, Google, Bing, and Yahoo, do not disclose the algorithms they use to rank pages. Some SEO practitioners have studied different approaches to search engine optimization, and have shared their personal opinions.[26] Patents related to search engines can provide information to better understand search engines.[27] In 2005, Google began personalizing search results for each user. Depending on their history of previous searches, Google crafted results for logged in users.[28]
A few links down and I've noticed that Brian has a link from WordPress.org. Not bad! Turns out that his content has been referenced within one of WordPress's codex posts. If I were to reach out and offer some additional insight, citing one of my articles, there's a chance I could bag a similar link, especially considering they have a 'Useful Resources' section.

Search engines may penalize sites they discover using black hat methods, either by reducing their rankings or eliminating their listings from their databases altogether. Such penalties can be applied either automatically by the search engines' algorithms, or by a manual site review. One example was the February 2006 Google removal of both BMW Germany and Ricoh Germany for use of deceptive practices.[53] Both companies, however, quickly apologized, fixed the offending pages, and were restored to Google's list.[54]

hey james - congrats on your success here. just a question about removing crummy links. for my own website, there are hundreds of thousands of backlinks in webmaster tools pointing to my site. The site has no penalties or anything  - the traffic seems to be growing every week. would you recommend hiring someone to go through the link profile anyway to remove crummy links that just occur naturally?
For our client: We rolled out a successful implementation of rel="author" for the three in-house content writers the company had. The client had over 300+ articles made by these content writers over the years and it was possible to implement rel="author" for all the aged articles. I advise anyone who has a large section of content to do so as it will only benefit the website. We were also in the process of rolling out further schema markup to the site's course content as it only has a benefit for CTR.
This topic seems actually quite controversial. Google answered the question by what could be taken as a denial. But their answer was kind of open to interpretations. And on the other hand, there are studies (one of them from Moz) that showed linking out has an impact. So, how can you be so assertive? Is it something that comes out from your own experiments?
After adjusting that, the table refreshes again and I’m looking at a month-by-month summary of my Organic Traffic. Hover your mouse over any single month dot to view a summary of that month’s numbers. In this particular example, we can see that recently there’s been an increase in Organic Traffic. January had 6,630 organic sessions, February (short month) had 5,982 and then March came in strong with 7,486 organic sessions. This information lets us know that something on the site is performing better than usual in March. In most cases, this means that either interest in a topic has increased or the website has begun to rank better in the search engines for specific keywords. In the next section we’ll begin to break this down further.
Inclusion in Google's search results is free and easy; you don't even need to submit your site to Google. Google is a fully automated search engine that uses web crawlers to explore the web constantly, looking for sites to add to our index. In fact, the vast majority of sites listed in our results aren't manually submitted for inclusion, but found and added automatically when we crawl the web. Learn how Google discovers, crawls, and serves web pages.3
If you are serious about improving search traffic and are unfamiliar with SEO, we recommend reading this guide front-to-back. We've tried to make it as concise as possible and easy to understand. There's a printable PDF version for those who'd prefer, and dozens of linked-to resources on other sites and pages that are also worthy of your attention.
Optimization techniques are highly tuned to the dominant search engines in the target market. The search engines' market shares vary from market to market, as does competition. In 2003, Danny Sullivan stated that Google represented about 75% of all searches.[63] In markets outside the United States, Google's share is often larger, and Google remains the dominant search engine worldwide as of 2007.[64] As of 2006, Google had an 85–90% market share in Germany.[65] While there were hundreds of SEO firms in the US at that time, there were only about five in Germany.[65] As of June 2008, the marketshare of Google in the UK was close to 90% according to Hitwise.[66] That market share is achieved in a number of countries.
Google claims their users click (organic) search results more often than ads, essentially rebutting the research cited above. A 2012 Google study found that 81% of ad impressions and 66% of ad clicks happen when there is no associated organic search result on the first page.[2] Research has shown that searchers may have a bias against ads, unless the ads are relevant to the searcher's need or intent [3]
To prevent users from linking to one version of a URL and others linking to a different version (this could split the reputation of that content between the URLs), focus on using and referring to one URL in the structure and internal linking of your pages. If you do find that people are accessing the same content through multiple URLs, setting up a 301 redirect32 from non-preferred URLs to the dominant URL is a good solution for this. You may also use canonical URL or use the rel="canonical"33 link element if you cannot redirect.

Search engine optimization (SEO) is often about making small modifications to parts of your website. When viewed individually, these changes might seem like incremental improvements, but when combined with other optimizations, they could have a noticeable impact on your site's user experience and performance in organic search results. You're likely already familiar with many of the topics in this guide, because they're essential ingredients for any web page, but you may not be making the most out of them.
Q re CTR: what’s the best way to study it? I just looked at the search console in Google Analytics and am perplexed. If I just look at the content part of GA, my top page has 12K uniques from google in past 30 days. But if I look at search console part, it says 222 clicks for past 30 days. I see a CTR there, but since there is such a discrepancy between the two counts for visits/clicks, I’m not sure what to think.
Mobile traffic: In the Groupon experiment mentioned above, Groupon found that both browser and device matter in web analytics’ ability to track organic traffic. Although desktops using common browsers saw a smaller impact from the test (10-20 percent), mobile devices saw a 50 percent drop in direct traffic when the site was de-indexed. In short, as mobile users grow, we are likely to see direct traffic rise even more from organic search traffic.
×