hey james - congrats on your success here. just a question about removing crummy links. for my own website, there are hundreds of thousands of backlinks in webmaster tools pointing to my site. The site has no penalties or anything  - the traffic seems to be growing every week. would you recommend hiring someone to go through the link profile anyway to remove crummy links that just occur naturally?
If both page are closely related (lots of topical overlap), I would merge the unique content from the lower ranking article into the top ranking one, then 301 redirect the lower performing article into the top ranking one. This will make the canonical version more relevant, and give it an immediate authority boost. I would also fetch it right away, do some link building, and possibly a little paid promotion to seed some engagement. Update the time stamp.

I would also advise to continue doing what works. If something you have rolled out generates great traffic and links bring out a new version of the content, for example the 2012 version worked effectively bring out the 2013 version of the content. Another effective strategy is to make the piece of content into an evergreen article which you add to over time so it is always up to date.
The first step to digging into organic traffic is to analyze what content on your website is performing best in this area. For obvious reasons, the homepage is almost certainly the landing page for most organic traffic, but the other top pages are often revealing. To view this data, we’re going to head over to the Behaviour section in the Analytics sidebar, then choose Site Content and finally Landing Pages.
Another tip you can use is just reach out to the prior agency and say something like the following: “We realise you were using link networks for our website which has resulted in a Google penalty and loss in business. Can you please remove my website from any link network you have built?”. If the prior agency is decent, they will remove the links from the network.
In regards to the “read More” button on mobile. Isn’t the page loaded asynchronously in the code (which google bots look at), meaning that the whole page is in FACT already loaded int he background, just not in the frontend, meaning google can read the content without being stopped by the button? Making it only a UI thing. How sure are you on the statement that mobile first will have an issue with this?

First, I will show you a quick snapshot of the traffic uplift, which yielded an additional 400,000 unique visitors from organic search traffic on a monthly basis. Then I will explain to you the steps in which we took to get the client to this level. I have also tried to keep this quite general so everyone can adapt their own situation to this case study.

Regarding Link Detox, links it diagnoses as Toxic are generally fine as they're either not indexed by Google or have malware/viruses/etc., but I recommend a manual review of any diagnosed as Suspicious. I used it recently to get started cleaning up our backlinks and some legit sites and blogs were under Suspicious simply because they didn't have many links pointing to them.

Second and last, keywords. I use SEO Ultimate on WordPress and I sometimes doubt about what words and sentences to put in the “Tags” box. Does it need to be very specific or broad words are good to have too ? Like if I’m writing about Porsche beating the previous record on the Nurburgring with the new GT2 RS, can I put “Nurburgring” and “Porsche GT2 RS” as tags or is better to just keep “Porsche record nurburgring” and specific tags like this one ?
Robots.txt is not an appropriate or effective way of blocking sensitive or confidential material. It only instructs well-behaved crawlers that the pages are not for them, but it does not prevent your server from delivering those pages to a browser that requests them. One reason is that search engines could still reference the URLs you block (showing just the URL, no title or snippet) if there happen to be links to those URLs somewhere on the Internet (like referrer logs). Also, non-compliant or rogue search engines that don't acknowledge the Robots Exclusion Standard could disobey the instructions of your robots.txt. Finally, a curious user could examine the directories or subdirectories in your robots.txt file and guess the URL of the content that you don't want seen.

Having large groups of content that all revolve around the same topic will build more relevance around keywords that you're trying to rank for within these topics, and it makes it much easier for Google to associate your content with specific topics. Not only that, but it makes it much easier to interlink between your content, pushing more internal links through your website.
Problems donating? | Other ways to give | Frequently asked questions | We never sell your information. By submitting, you are agreeing to our donor privacy policy. The Wikimedia Foundation is a nonprofit, tax-exempt organization. If you make a recurring donation, you will be debited by the Wikimedia Foundation until you notify us to stop. We'll send you an email receipt for each payment, which will include a link to easy cancellation instructions.
You may not want certain pages of your site crawled because they might not be useful to users if found in a search engine's search results. If you do want to prevent search engines from crawling your pages, Google Search Console has a friendly robots.txt generator to help you create this file. Note that if your site uses subdomains and you wish to have certain pages not crawled on a particular subdomain, you'll have to create a separate robots.txt file for that subdomain. For more information on robots.txt, we suggest this Webmaster Help Center guide on using robots.txt files13.
Simply great and agree with your all subject...! I like the way you explained. Each heading are awesome Create the best quality content and consistently, Long tail keyword is better, Guest blog for SEO is dead, and Aha....Do not anger Google. conclusion is awesome. Hard work and Patient is best practice to see the good results in any field. Really useful and helpful post indeed. Thank you.
Obviously that doesn’t make any sense, as no tool developer would have the capabilities to deliver actual LSI keyword research. Something like LSI optimization does not exist. Even Google using LSI in it’s algorithm is pure speculation. Anyone who makes such claims should take a long hard look at the LSI tutorial by Dr. E Garcia (and then stop making those claims, obviously). This is the only part I can find: http://www.360doc.com/content/13/1124/04/9482_331692838.shtml
Optimise for your personas, not search engines. First and foremost, write your buyer personas so you know to whom you’re addressing your content. By creating quality educational content that resonates with you>r ideal buyers, you’ll naturally improve your SEO. This means tapping into the main issues of your personas and the keywords they use in search queries. Optimising for search engines alone is useless; all you’ll have is keyword-riddled nonsense.

The first step to digging into organic traffic is to analyze what content on your website is performing best in this area. For obvious reasons, the homepage is almost certainly the landing page for most organic traffic, but the other top pages are often revealing. To view this data, we’re going to head over to the Behaviour section in the Analytics sidebar, then choose Site Content and finally Landing Pages.
Hi Matt, realizing now how difficult it is to run a blog, trying to promote it and carry on with your daily activities. I would say it's a full time job. Once you thing you done learning about something, something else is coming :). My blog is about preparing for an ironman so I need to add the training on top of it. Thanks a lot for sharing this article with us so we can keep focus!!!
Regarding Link Detox, links it diagnoses as Toxic are generally fine as they're either not indexed by Google or have malware/viruses/etc., but I recommend a manual review of any diagnosed as Suspicious. I used it recently to get started cleaning up our backlinks and some legit sites and blogs were under Suspicious simply because they didn't have many links pointing to them.
Having large groups of content that all revolve around the same topic will build more relevance around keywords that you're trying to rank for within these topics, and it makes it much easier for Google to associate your content with specific topics. Not only that, but it makes it much easier to interlink between your content, pushing more internal links through your website.

What we look for in a list like this is to identify the pages that are performing well so we can continue to capitalize on those. In this example, we see that the inventory pages are getting significant traffic, which is great, but we also see that the Team page and the Service page are both also ranking well. With this information in mind, we should revisit these pages to ensure that they are structured with the right content to perform as the visitor’s first page view, possibly their first glimpse at your business.


Place strategic search phrases on pages. Integrate selected keywords into your website source code and existing content on designated pages. Make sure to apply a sug­gested guideline of one to three keywords/phrases per content page and add more pages to complete the list. Ensure that related words are used as a natural inclu­sion of your keywords. It helps the search engines quickly determine what the page is about. A natural approach to this works best. In the past, 100 to 300 words on a page was recommended. Many tests show that pages with 800 to 2,000 words can outperform shorter ones. In the end, the users, the marketplace, content and links will determine the popularity and ranking numbers.
Looking at the keyword rankings and organic landing pages provided a little bit of insight into the organic traffic loss, but it was nothing definitive. Because of this, I moved to the on-page metrics for further clarity. As a disclaimer, when I talk about on-page metrics, I’m talking about bounce rate, page views, average page views per session, and time on site.

Thanks for the comment Slava good too see your team is on top of things and happy you liked the post. The website in the case listed was a client who has taken on an agency who was doing lower quality SEO work which was affecting the site such as the huge link network and a strategy which only revolved around mainly head terms. We saw no long tail integration from the old agency's strategy, so we were able to yield great results to begin with. The clients site has 100's of high quality articles which we were able to re optimize and update as noted. Further to this they had a large index of high quality pages to work from. Sure enough the points listed above were key elements to a far wider strategy which could be 100's of points. I just wanted to include some of the biggest wins and easy to implement points.  
Mobile traffic: In the Groupon experiment mentioned above, Groupon found that both browser and device matter in web analytics’ ability to track organic traffic. Although desktops using common browsers saw a smaller impact from the test (10-20 percent), mobile devices saw a 50 percent drop in direct traffic when the site was de-indexed. In short, as mobile users grow, we are likely to see direct traffic rise even more from organic search traffic.
The Google, Yahoo!, and Bing search engines insert advertising on their search results pages. The ads are designed to look similar to the search results, though different enough for readers to distinguish between ads and actual results. This is done with various differences in background, text, link colors, and/or placement on the page. However, the appearance of ads on all major search engines is so similar to genuine search results that a majority of search engine users cannot effectively distinguish between the two.[1] How To Grow a New Website to Over 50,000 Organic Visits Per Month
×