How to recover your site from Penguin and Panda updates

by hadsek
34 replies
  • SEO
  • |
Thought to share this. Used it to recover some friends 'sites affected by Panda and immuned them from Penguin. This advice is aimed at people who listen and act on good faith. I want to help and stop them from steering off course, worring about offsite stuff. As a reminder the penguin update was initially called the over optimization filter. Isn't that explicit enough?

Yes, excessive link getting practices have some effect but only in conjunction with a poorly designed site (slow loading, too many ads vs content, duplicate pages, seozied writing style). Fix the site and Judge Google will set it free.

And please be responsible. There are many who are desperately seeking this help and can be discouraged by confusing arguments and comments. If you disagree, how about taking a problematic page from your site and apply the steps below to it ? Let it run for a week or 2, then come back here to make your feeback-based comment.

As of to date, Penguin is still in test and feedback mode. This strategy is intended for the long run no matter what the algorithm's name.

This formula holds onsite factors accountable for a long list of reasons.

I started out the experiment by putting top 7 different-niche sites under the microscope.
I was convinced offsite factors have no impact (with the exception of blatant violations the kind you get a GWT notification for).

My first visit to the top 7 sites revealed they're all implicated in a blackhat seo, some on a large scale (hundreds of forum profile links, bookmarks, comments,etc).

Which is not a secret.

Every site that has attempted to rank commercially on google during the last 10 years has been engaged(in some form or other) in unatural link building and self-promotion(fake reviews, manipulated alexa ranking, etc).

What's more, in almost all niches sites in 5th and 6th positions have a higher link profile than the top 4, which told me google is simply not counting thousands of their links. Big G is not penalizing them, just ignores a portion of their links it considers shady.

When comparing those top sites to the ones that were competing with them and dropped, I noticed the following pattern. Those that dropped have a combination of the following:

1 - A higher keyword density (%7 +).

2 - A higher ratio of duplicate pages.

3 - A higher ratio of thin content above the fold.

4 - Duplicate Rss feeds.


It has to be all the above combined, not seperately. If a site only has keyword stuffing, only a specific-keyword penalty is applied not even the page itself.

Also, those top sites have:

1 - A high number of sandboxed duplicates (the ex -50 penalty).

2 - A higher number of ranking pages(apparently due to their traffic ratio).

Where did these findings leave me ?

They confirmed to me that:

1- Offsite factors are not that of a major issue as many believe.
2- Duplicate content is punishable across the board and is a factor.
3- Disparity in keyword density (keyword stuffing) is a big ranking factor.

This finding agrees with google's repeated assertion that they care more about content retrieved to their users. Duplicate and keyword-stuffed content hurt their users' experience. Links do NOT.


Working on 1 & 2 above resulted in freeing wounded sites from the sandbox, in particular the homepage. Internal pages recovered partially (resurfaced on the serps but ranked poorly).

I wondered why did homepages recovered successfuly but internal pages did not. First thing that jumped to my mind is that homepages are usually UNIQUE. Turned out (at least how I saw it), site-wide penalty still held even for cleaned internal pages so long as their sister pages are not cleaned yet. In other words, if you clean 5 out of 30 pages, the 5 will continue to perform poorly till all 30 are fixed. That's a Panda penalty refined.


To cut it short, for those trying to recover from Panda/Penguin, try this:

1- Reduce keyword density. I helped recover some sites from Panda where site owners almost gave up by this very method exclusively. Here's how you do it. First Things First:

Create a 3 columns/rows table with these entries:

Keyword | Count | Density

Now Visit one of these sites (or google for yours). I used these, their results vary but still helpful:

Keyword Density Checker - Keyword Cloud
Live Keyword Analysis

Paste your url, then copy and paste the keyword, count and density results in the table you created. Take notice of the Density ratio only, forget the other stuff.

Now go back to the checker and paste your best competitor's url (those that are still ranking).

Do you see the difference ?

Now try to match that keyword density count by reducing yours. From my observation, google compares your ratio to your market's average. 2% limit is recommended, 3% is the maximum. The lower you set it, the upper you will go in ranking. As you do this, you will notice your article is getting "more" natural (OMG, am writing for readers for the first time

Did you match it ? If yes, congrats. If not, go back to work.

Don't forget to remove any keyword stuffing in the link's title tag and img tag (google counts those too in the ratio calculation). No more than 1 or 2 words. While you're at it, remove above the fold ads (site-wide). Also check any reciprocal linking. A decent site you linked back to 2 years ago may be is now a spam/malware site popping up a download box(someone's site got filtered out simply due to this).

When you're done, go to your sitemap and update the date.

Now log to your GWT, select the wounded site, look to your left and click on Diagnostics, then then click on Fetch as Googlebot. Enter the url (the root should already be there, just add the rest). Then click submit (make sure you only submit that url not the url and its linked urls option).

Submit a reconsideration request:

docs.google.com/spreadsheet/viewform?formkey=dEVxdmdRWFJRTjRoLWZVTHZkaTBQbkE6M Q

Now wait for 3-5 days and the magic will happen (either automatically or manually, assuming that was your only issue)

If no joy yet, you need to work on duplicate pages. Google for "duplicate page checker" or use this one below:

Similar Page Checker - Duplicate content checker

If you find any, insert the rel=canonical tag in the duplicate page to point to the one you prefer ranked. Some dyanmic database-driven pages will simply not recover under these algorithms. Most those will be considered similar no matter how much you tweaked them unique. Googlebot goes beyond unique meta data, 10% different body content and analyse those pages also based on their template structure. A change in the city or product name does not make them unique. What you can do is focus on ranking those that are accepted and indexed. Hopefully you get some yahoo referrals from the others or try to recuperate their lost traffic thru DailyMotion videos.

By this stage, your waiting time will be around a week or two. Past this, you definitely have an offsite seo penalty. You need to do some whitehat link mixology to offset the negative ratio. Even if you don't have a penalty, these will boost your rankings. Proceed as follows:


1- Do some LOCAL directory linking (1 pr5 link from these can drag your site back 20 pages up). By local, I mean country-specific. Don't waste your time on any other.

2 - Do MANUAL social bookmarking (SU and Digg are the best - write real reviews, google ignores 1 line comment of "Great read, thank you for the share).

3 - Do some article marketing. I recommend hubpages but they should be real articles that are highly DIFFERENT from those already indexed. Penguin can detect rehashed information and will drop the article even it got indexed the first time.

4 - Do not hurt your baby article with backlinking in its probationary period(first 2 weeks).
Only tweet and facebook it on your own accounts. Then make sure someone else social bookmark it on SU and Digg. Video would be a grea extra. That's it. No Forum profiles, no Angela & Paul and no Fiverr.

5 - Tweet your links more with variations of same achor text. Do same thing with Fb.

6 - Blog commenting on AUTHORITY blogs in the same niche. Blog commenting does boost your site ranking when credited. But only comments on authority blogs in the same niche will count. 50.000 automated comments will NOT.

Why is commenting important in recovery and ranking ?

It creates you AUTHOR CREDIBILITY with google who is more interested in you as a blogger than your post or link. It is also a social signal of your interactivity. Engaging in commenting validates you as an interesting source of information. This formula would be lame if I kept any how-to. So am gonna share with you the exact commenting technique I deployed. And, of course, I tested it before implementing it.

How did I test it ?

I set up a free blog and decided to rank it on twitter, facebook and commenting only. I used twitter and facebook at first with another blog that did not cut it. So here you go:

1 - Share your article's link and a good snippet on Twitter MANUALLY (No url Shortner).
2 - Ping your twitter'rss. Do not log out from Twitter.
3 - Share same article's link and a good snippet on facebook fan page MANUALLY (No url Shortner).
4 - Ping your fb fan page'rss. Log out from fb.
5 - Go back to twitter. If you are following good content providers, look for the best tweets and
retweet them. If not, type your niche in the search box on top, follow some good brains and
retweet their stuff.
6 - Follow the tweet link and post a comment signing as Twitter.
7 - Post a good comment with NO LINK whatsoever(your twitter's url should be loaded along your
photo when you signed in).

Now google will follow your twitter address back to your original tweeted post (and those retweets).
Google will credit your Twitter profile in its own way (pretty valuable).

See the difference ?

You get credit for yourself as a twitterer (thru your followed twitter profile). In other words, you're reconciling with google and establishing "author credibility" . As a consequence, your entire blog will rank. By the way, the 2 weeks old test blog is sitting now on page 1 (6th position).

Before commenting, always remember that google assumes you're a bot till proven innocent. Therefore, to make the comment count, prove you're not a bot by referencing words, phrases or concepts from the blog post. Basically, do what automated commenting can't. Instead of :

Great article, thanx

try something like:

Did not realize that Personal Branding can be that beneficial to one's business. Thanks for the heads up

Not only the latter comment will likely be approved by the blog owner, googlebot will realize non-humans cannot relate intelligently to a blog post they don't read. Also, when was the last time
you heard a blog owner disapprove a Twitter signed-in comment ?

I'd like to remind you this recovery method is a per-page process. Recovering the homepage (which is extremely important) will not recover internal pages if they,too, have issues.

Lastly, take a break to refuel then simply rinse and repeat the above steps and you should be able to clean everything.

The process above is based on actual work done on wounded established sites that DID recover.

Hope this helps.

N.B.

Since Penguin algo is still under tweaking, there may be some false positives as well as floating spam. Google has a form where you can report those poor pages that should be caught but weren't:

https://accounts.google.com/ServiceL...?hl%3Den&hl=en
#panda #penguin #recover #site #updates
  • Profile picture of the author marylou
    GREAT WORK, THANX
    {{ DiscussionBoard.errors[6191026].message }}
  • Profile picture of the author GeorgR.
    1- Reduce keyword density. I helped recover some sites from Panda where site owners almost gave up by this very method exclusively. Here's how you do it. First Things First:

    Create a 3 columns/rows table with these entries:

    Keyword | Count | Density
    This is exactly the kind of "help" i complained about in another post.

    Is this pure speculation or is your advice that on-page changes are needed based on REAL observed results? If so..where is the studies showing that exactly those changes are actually helping to "fight" penguin let alone prevent it...

    If penguin is purely adressing link diversity and is ONLY a change based on off-site factors, you can sit and tweak your content, titles and keywords until you are blue in the face. You would not even address what Penguin is about since on-page, keyword density etc. is not even a factor there.

    Did you personally already observe that reducing KW density improved your rankings after Penguin or are you just blowing hot air and like to speculate?
    Signature
    *** Affiliate Site Quick --> The Fastest & Easiest Way to Make Affiliate Sites!<--
    -> VISIT www.1UP-SEO.com *** <- Internet Marketing, SEO Tips, Reviews & More!! ***
    *** HIGH QUALITY CONTENT CREATION +++ Manual Article Spinning (Thread Here) ***
    Content Creation, Blogging, Articles, Converting Sales Copy, Reviews, Ebooks, Rewrites
    {{ DiscussionBoard.errors[6191118].message }}
    • Profile picture of the author hadsek
      @ georgr

      You even quoted my answer already. I have updated my post above.

      I don't know how can someone fix something based on SPECULATION. If you noticed, I joined the forum 4 years ago but only wrote 12 posts. So obviously, am not into blowing hot air and like to speculate.

      I did the dirty work first, reaped the results, then came here to share how I achieved them. It took me 3 hectic months of testing, writing down the notes, comparing, redoing, tweaking and re-testing. I had even reversed and recreated the issue (wax on, wax off).

      I recovered two of my own sites and everyday am still bringing 3 pages from the sandbox right to page 20, 8, 3 then page 2 where they have been previously. These pages had only two or three natural inbound links so an offsite issue is out of the equation.

      Quoting you:
      [If penguin is purely adressing link diversity and is ONLY a change based on off-site factors..]


      from my observations, that is not the case at all. And if you re-read the post, I have addressed the issue both on-site and offsite, onsite being the major issue that if corrected google seems to forgive minimal offsite factors.

      And as a disclaimer, that was my personal experiment that worked for the sites I worked on by mainly fixing keyword density and internal duplicates. But if anyone is trying to recover or rank a duplicate, I'd say forget it.
      {{ DiscussionBoard.errors[6192602].message }}
  • Profile picture of the author momo3
    WARNING - I AM NOT CLAIMING TO BE AN SEO PRO:

    But I like the OP's post. Because I feel Penguin has to be a combination of Onsite and Offsite overoptimziation.

    I feel google would be foolish to base it all on inbound anchor text.

    I feel that maybe sites that were overopitmized on site got hit, and sites that were overoptimized offsite (too many same-text anchor) were hit.. and some sites with a little of both were hit.
    {{ DiscussionBoard.errors[6191369].message }}
  • Profile picture of the author momo3
    Hadsek -- please explain what you mean by your final sentence:

    "And as a disclaimer, that was my personal experiment that worked for the sites I worked on by mainly fixing keyword density and internal duplicates. But if anyone is trying to recover or rank a duplicate, I'd say forget it. "


    what do you mean by duplicates
    {{ DiscussionBoard.errors[6192739].message }}
    • Profile picture of the author hadsek
      Hi momo3,

      Duplicates are similar webpages common on dynamically generated product pages where only location, product name or color is different. This can also be a page that you deliberately write to rank for a variation of the same keyword but with a slightly different or spun content. Lately, google seems to include into this category external rehashed content (like a squidoo lens with a link to your site that is similar to another hub with your link trail) .
      {{ DiscussionBoard.errors[6192927].message }}
  • Profile picture of the author socialbookmark
    Thanks for good tips. There are many members talk about Penguin and Panda updates in WF and these tips can help them. In my opinion Over optimization is a wrong seo technique that avoiding it, fixes many problems.
    Signature

    I love warriorforum. Computer Tutorials

    {{ DiscussionBoard.errors[6192875].message }}
  • Profile picture of the author momo3
    Hadsek --

    I have many pages that I did write around certain terms. They are NOT too stuffy but they are kind of redundant.

    I wouldn't say they are duplicate content.. but maybe equal to writing articles like

    Dog Food
    Cheap Dog Food
    Dog Chow
    Doggie Chow

    etc.

    I did this with about ~70 articles on my site.

    I think this could be the cause of my problem. Thoughts?
    {{ DiscussionBoard.errors[6193493].message }}
    • Profile picture of the author hadsek
      Momo3,

      I have no doubt that is a problem. Try pasting two pages you think they're similar on this tool:

      webconfs.com/similar-page-checker.php

      then work on keeping similarity ratio below 90%. They can survive at 95% but the lower the rate, the more unique they become.
      {{ DiscussionBoard.errors[6194067].message }}
    • Profile picture of the author hadsek
      thesmarterbrain,

      Thank you in advance. I would love the feedback.
      Good luck !
      {{ DiscussionBoard.errors[6194074].message }}
  • Profile picture of the author dp40oz
    I think you are severely underestimating the impact of offpage factors. Yes there were onpage factors in Penguin but the majority of people were affected because of offpage links.
    {{ DiscussionBoard.errors[6194152].message }}
  • Profile picture of the author johnben1444
    Here is what someone is doing and its working wonders too since the 2012 updates and penguin.

    Step 1) Buy high PR domains from 2-3 yrs ago on godaddy auctions in bulk (he seemingly has bought 133 so far)

    Step 2) Redirect high PR domains to a fresh site with tons of content you backdate to make your site seem older with no outbound links

    Step 3) Blast a few million blogs, profiles, guestbooks to these 133 domains

    This guy is now on page one for a 5-6 million/yr term, while authority sites with PRs of 4 and 5 and thousands of pages of hand written content are on page 5/6 or worse

    This strategy is interesting... and seems to work very well (it's also very easy to do)
    Post March changes, I've seen this 301 redirection technique work true miracles... but this site in particular is amazing for a bunch of reasons.

    The guy is also doing this across 5 mega niches at once, and his various sites are ranking on all of them (all the same backdated post technique)
    I'm talking about some of the SEO dream terms like:
    e-cigarettes

    Terms that can easily rotate $20-30k/day in profit for sites, and yes his site is on page 1 for e-cigs now. He has no affiliate links, stores, etc set up so I truly wonder what he's doing. I'm guessing he has 1-2 months to capitalize before his site is gone.

    This would never, ever have worked before February... yet...
    Thoughts?
    Signature
    Grow your social media account, Spotify Streams, YT Views & IG Followers & More
    Software & Mobile APP Developer
    Buy Spotify, Facebook Bot & IG M/S Method
    {{ DiscussionBoard.errors[6194194].message }}
    • Profile picture of the author kindsvater
      Originally Posted by johnben1444 View Post

      Step 2) Redirect high PR domains to a fresh site with tons of content you backdate to make your site seem older with no outbound links
      Yeah, that will really fool Google - because Google will backdate its index thinking your page has been around for years just because you put the number 2006 on it.

      Step 3) Blast a few million blogs, profiles, guestbooks to these 133 domains
      Nothing like trying to avoid the Penguin penalty by doing exactly what Google says it is penalizing.

      And the rest of us, we so appreciate all the spam.

      .
      {{ DiscussionBoard.errors[6197340].message }}
  • Profile picture of the author momo3
    Jeez JohnBen1444 - that is intense. I've heard of that technique too..
    But it sounds like they will catch onto it?
    {{ DiscussionBoard.errors[6196858].message }}
  • Profile picture of the author John Williamson
    OP, interesting finds, but correlation does not imply causation, my friend, especially with such a small sample.
    Signature
    The Google Adwords Keyword Tool is hiding your valuable keywords!
    OFFLINERS, Start using this simple technique and these 6 "weapons" today to get more clients and skyrocket your conversions! - FREE, no opt-in.
    Make some money by helping me market this idea.
    {{ DiscussionBoard.errors[6197374].message }}
  • Profile picture of the author discuss4u
    Hadsek, do you include the meta description and meta keyword when you check the keyword density?

    And how about the side bars / nevigation bar on the templates. Do you also include that when you check the density? Or you just check the article body?

    Thanks!
    {{ DiscussionBoard.errors[6199109].message }}
    • Profile picture of the author hadsek
      Discuss4u,

      Yes, title, meta description, any url title, alt all count in keyword density. When playing with it on a page that runs adsense, you will see how adsense bot is serving you ads based on the dominant keyword. This is also a great way to cure adsense irrelevancy. In many cases, you may be giving higher density to the wrong keyword.

      Navigation bars are a major cause for duplicate penalty, try to reduce them to a minum.I'd put an ad there instead.
      {{ DiscussionBoard.errors[6199217].message }}
      • Profile picture of the author hadsek
        SamDermot,

        Yes natural and manual linkbuilding still a factor by google's official position. But updates, social interaction(social media, forum post, guest blog post, yelp, craiglist), domain age, dofollow links from high pr sites in the same niche are even more important. Once a site achieves some good standing it can go a long time without any link building.
        {{ DiscussionBoard.errors[6199243].message }}
  • Profile picture of the author SamDermot
    Banned
    If we work on content only then how can we rank or compete for some competitive keywords.

    If I don't have to do link building and I have to use content with proper density.

    Andy I'm trying to rank for the keyword "seo"

    and other sites are also trying to rank for the same keyword, so that is the main technique in content so it would compete from others?
    {{ DiscussionBoard.errors[6199140].message }}
  • Profile picture of the author skid106
    First of all, excellent research Hadsek. Thank you for all of that.

    I'm seeing an interesting thing (which a few others have also seen and commented on in other forums). Here's an example for one of my sites, which I resurrected at the end of March :

    mysite.com (170 backlinks - built over 10 months - all pointing to the home page as it was originally a thin 'sniper site' of about 5-7 pages). Domain is now 1yr old.

    From April 1st, I started putting on 2 product niche specific Amazon-type reviews per day - getting as far as 17 reviews (site in total had 22 posts and the home 'page'). Content was my own take on Amazons descriptions / specifications / reviews, plus comments / descriptions of other 'specialist' sites. 600-700 words per review.

    I went for easy to rank, low search volume, product names - targeting "brand/model" and "brand/model review" for each post (main focus being on "Brand/model" keyword).

    I was ranking in 48 hours, getting as far as p2 (some p1) of Google in 7-10 days - with no backlinking to any of those posts (competition was very weak). The posts were natural writing with a density of no more than 1.5%.

    On April 26th (when Penguin seemed to hit my neck of the woods), everything went south and, since then, my posts are still ranking (but no higher than #35), but for a combination of either :

    (a) the wrong keyword or
    (b) just the category section (there are 5 obvious categories in this niche).

    Just to clarify point (a), this is what I'm seeing :

    Search Term : "Brand Model AAA", shows url mysite.com/brand-model-bbb at #35 in the SERPS.

    Search Term : "Brand Model BBB", shows url mysite.com/brand-model-ccc at #48 in the SERPS.

    Obviously on page ccc, there is no mention at all of model BBB, so I can't see how this url is ranking at all, whilst the actual keyword (search term) url is no-where to be seen ?

    This is happening for every single one of my posts for the search terms "Brand/model" and / or "Brand/Model Review". The homepage keyword is also throwing up one of the inner page urls as the best position for that search term.

    I'm using SEOBook Rank Checker, which is telling me where I'm ranking and for which URL. This is also telling me that the highest ranking url for many of the same inner page search terms under Yahoo & Bing is........ "About Us" ?!?!?

    Any ideas Hadsek ??
    {{ DiscussionBoard.errors[6212089].message }}
    • Profile picture of the author Mantasmo
      Originally Posted by skid106 View Post

      First of all, excellent research Hadsek.
      This is an insult to every person who understands the meaning of the word research.

      Pure speculation, especially when there's a MASSIVE chance that Penguin is rolled out in iterations, just like Panda. Keep on analyzing, bros.
      {{ DiscussionBoard.errors[6212296].message }}
      • Profile picture of the author OmarNegron
        Originally Posted by bnetwork View Post

        This is an insult to every person who understands the meaning of the word research.

        Pure speculation, especially when there's a MASSIVE chance that Penguin is rolled out in iterations, just like Panda. Keep on analyzing, bros.
        Yep. I mean it is a good overview but no one really knows the answer and there is massive speculation going on.

        -Omar
        {{ DiscussionBoard.errors[6212633].message }}
      • Profile picture of the author Oranges
        Originally Posted by bnetwork View Post

        This is an insult to every person who understands the meaning of the word research.

        Pure speculation, especially when there's a MASSIVE chance that Penguin is rolled out in iterations, just like Panda. Keep on analyzing, bros.
        This!

        G! also said that penguin update is like Panda, so in every 2-3 months it will refresh and update then there will be whole new set of speculations.
        Signature

        {{ DiscussionBoard.errors[6224172].message }}
  • Profile picture of the author plsearch
    Yea this is pure speculation, there's tons of evidence of sites with keyword anchor text that is over 60% money keywords having problems too.

    Out of 50 sites we work on, many improved that have keyword stuffing and duplicate pages (ecommerce). The ones that dropped had been very aggressive with low quality high volume links.
    {{ DiscussionBoard.errors[6212199].message }}
  • Profile picture of the author plsearch
    https://plus.google.com/u/1/11254407...ts/hHaXg8Rs5Lf

    link profiles are cause from what Randfish has found so far
    {{ DiscussionBoard.errors[6212328].message }}
  • Profile picture of the author rizoalbert
    I did a great research on penguin.It will be helpful.
    {{ DiscussionBoard.errors[6217959].message }}
  • Profile picture of the author puvanahim
    I got some good idea
    Thanks
    {{ DiscussionBoard.errors[6222472].message }}
  • Profile picture of the author gpromotional
    Good ideas
    {{ DiscussionBoard.errors[6224204].message }}
  • Profile picture of the author WinsonYeung
    awesome post! this is indeed one of the rare threat I can find about telling you in steps what to do to recover from the Penguin update
    Signature
    [WSO of The Day] Discount How To Generate 172.56% Positive Return OR build your List for FREE!

    "Case Study: Discover You Can Make $1371.66 With A Simple Blog Post by Clicking Here"
    {{ DiscussionBoard.errors[6225369].message }}
  • Profile picture of the author deliverer
    Great and helpful. This gives the newbies a BIG leap to this crazy but paying world; AM.
    {{ DiscussionBoard.errors[6231247].message }}
  • Profile picture of the author dotgirish
    Originally Posted by hadsek View Post

    Thought to share this. Used it to recover some friends 'sites affected by Panda and immuned them from Penguin. This advice is aimed at people who listen and act on good faith. I want to help and stop them from steering off course, worring about offsite stuff. As a reminder the penguin update was initially called the over optimization filter. Isn't that explicit enough?

    Yes, excessive link getting practices have some effect but only in conjunction with a poorly designed site (slow loading, too many ads vs content, duplicate pages, seozied writing style). Fix the site and Judge Google will set it free.

    And please be responsible. There are many who are desperately seeking this help and can be discouraged by confusing arguments and comments. If you disagree, how about taking a problematic page from your site and apply the steps below to it ? Let it run for a week or 2, then come back here to make your feeback-based comment.

    As of to date, Penguin is still in test and feedback mode. This strategy is intended for the long run no matter what the algorithm's name.

    This formula holds onsite factors accountable for a long list of reasons.

    I started out the experiment by putting top 7 different-niche sites under the microscope.
    I was convinced offsite factors have no impact (with the exception of blatant violations the kind you get a GWT notification for).

    My first visit to the top 7 sites revealed they're all implicated in a blackhat seo, some on a large scale (hundreds of forum profile links, bookmarks, comments,etc).

    Which is not a secret.

    Every site that has attempted to rank commercially on google during the last 10 years has been engaged(in some form or other) in unatural link building and self-promotion(fake reviews, manipulated alexa ranking, etc).

    What's more, in almost all niches sites in 5th and 6th positions have a higher link profile than the top 4, which told me google is simply not counting thousands of their links. Big G is not penalizing them, just ignores a portion of their links it considers shady.

    When comparing those top sites to the ones that were competing with them and dropped, I noticed the following pattern. Those that dropped have a combination of the following:

    1 - A higher keyword density (%7 +).

    2 - A higher ratio of duplicate pages.

    3 - A higher ratio of thin content above the fold.

    4 - Duplicate Rss feeds.


    It has to be all the above combined, not seperately. If a site only has keyword stuffing, only a specific-keyword penalty is applied not even the page itself.

    Also, those top sites have:

    1 - A high number of sandboxed duplicates (the ex -50 penalty).

    2 - A higher number of ranking pages(apparently due to their traffic ratio).

    Where did these findings leave me ?

    They confirmed to me that:

    1- Offsite factors are not that of a major issue as many believe.
    2- Duplicate content is punishable across the board and is a factor.
    3- Disparity in keyword density (keyword stuffing) is a big ranking factor.

    This finding agrees with google's repeated assertion that they care more about content retrieved to their users. Duplicate and keyword-stuffed content hurt their users' experience. Links do NOT.


    Working on 1 & 2 above resulted in freeing wounded sites from the sandbox, in particular the homepage. Internal pages recovered partially (resurfaced on the serps but ranked poorly).

    I wondered why did homepages recovered successfuly but internal pages did not. First thing that jumped to my mind is that homepages are usually UNIQUE. Turned out (at least how I saw it), site-wide penalty still held even for cleaned internal pages so long as their sister pages are not cleaned yet. In other words, if you clean 5 out of 30 pages, the 5 will continue to perform poorly till all 30 are fixed. That's a Panda penalty refined.


    To cut it short, for those trying to recover from Panda/Penguin, try this:

    1- Reduce keyword density. I helped recover some sites from Panda where site owners almost gave up by this very method exclusively. Here's how you do it. First Things First:

    Create a 3 columns/rows table with these entries:

    Keyword | Count | Density

    Now Visit one of these sites (or google for yours). I used these, their results vary but still helpful:

    Keyword Density Checker - Keyword Cloud
    Live Keyword Analysis

    Paste your url, then copy and paste the keyword, count and density results in the table you created. Take notice of the Density ratio only, forget the other stuff.

    Now go back to the checker and paste your best competitor's url (those that are still ranking).

    Do you see the difference ?

    Now try to match that keyword density count by reducing yours. From my observation, google compares your ratio to your market's average. 2% limit is recommended, 3% is the maximum. The lower you set it, the upper you will go in ranking. As you do this, you will notice your article is getting "more" natural (OMG, am writing for readers for the first time

    Did you match it ? If yes, congrats. If not, go back to work.

    Don't forget to remove any keyword stuffing in the link's title tag and img tag (google counts those too in the ratio calculation). No more than 1 or 2 words. While you're at it, remove above the fold ads (site-wide). Also check any reciprocal linking. A decent site you linked back to 2 years ago may be is now a spam/malware site popping up a download box(someone's site got filtered out simply due to this).

    When you're done, go to your sitemap and update the date.

    Now log to your GWT, select the wounded site, look to your left and click on Diagnostics, then then click on Fetch as Googlebot. Enter the url (the root should already be there, just add the rest). Then click submit (make sure you only submit that url not the url and its linked urls option).

    Submit a reconsideration request:

    docs.google.com/spreadsheet/viewform?formkey=dEVxdmdRWFJRTjRoLWZVTHZkaTBQbkE6M Q

    Now wait for 3-5 days and the magic will happen (either automatically or manually, assuming that was your only issue)

    If no joy yet, you need to work on duplicate pages. Google for "duplicate page checker" or use this one below:

    Similar Page Checker - Duplicate content checker

    If you find any, insert the rel=canonical tag in the duplicate page to point to the one you prefer ranked. Some dyanmic database-driven pages will simply not recover under these algorithms. Most those will be considered similar no matter how much you tweaked them unique. Googlebot goes beyond unique meta data, 10% different body content and analyse those pages also based on their template structure. A change in the city or product name does not make them unique. What you can do is focus on ranking those that are accepted and indexed. Hopefully you get some yahoo referrals from the others or try to recuperate their lost traffic thru DailyMotion videos.

    By this stage, your waiting time will be around a week or two. Past this, you definitely have an offsite seo penalty. You need to do some whitehat link mixology to offset the negative ratio. Even if you don't have a penalty, these will boost your rankings. Proceed as follows:


    1- Do some LOCAL directory linking (1 pr5 link from these can drag your site back 20 pages up). By local, I mean country-specific. Don't waste your time on any other.

    2 - Do MANUAL social bookmarking (SU and Digg are the best - write real reviews, google ignores 1 line comment of "Great read, thank you for the share).

    3 - Do some article marketing. I recommend hubpages but they should be real articles that are highly DIFFERENT from those already indexed. Penguin can detect rehashed information and will drop the article even it got indexed the first time.

    4 - Do not hurt your baby article with backlinking in its probationary period(first 2 weeks).
    Only tweet and facebook it on your own accounts. Then make sure someone else social bookmark it on SU and Digg. Video would be a grea extra. That's it. No Forum profiles, no Angela & Paul and no Fiverr.

    5 - Tweet your links more with variations of same achor text. Do same thing with Fb.

    6 - Blog commenting on AUTHORITY blogs in the same niche. Blog commenting does boost your site ranking when credited. But only comments on authority blogs in the same niche will count. 50.000 automated comments will NOT.

    Why is commenting important in recovery and ranking ?

    It creates you AUTHOR CREDIBILITY with google who is more interested in you as a blogger than your post or link. It is also a social signal of your interactivity. Engaging in commenting validates you as an interesting source of information. This formula would be lame if I kept any how-to. So am gonna share with you the exact commenting technique I deployed. And, of course, I tested it before implementing it.

    How did I test it ?

    I set up a free blog and decided to rank it on twitter, facebook and commenting only. I used twitter and facebook at first with another blog that did not cut it. So here you go:

    1 - Share your article's link and a good snippet on Twitter MANUALLY (No url Shortner).
    2 - Ping your twitter'rss. Do not log out from Twitter.
    3 - Share same article's link and a good snippet on facebook fan page MANUALLY (No url Shortner).
    4 - Ping your fb fan page'rss. Log out from fb.
    5 - Go back to twitter. If you are following good content providers, look for the best tweets and
    retweet them. If not, type your niche in the search box on top, follow some good brains and
    retweet their stuff.
    6 - Follow the tweet link and post a comment signing as Twitter.
    7 - Post a good comment with NO LINK whatsoever(your twitter's url should be loaded along your
    photo when you signed in).

    Now google will follow your twitter address back to your original tweeted post (and those retweets).
    Google will credit your Twitter profile in its own way (pretty valuable).

    See the difference ?

    You get credit for yourself as a twitterer (thru your followed twitter profile). In other words, you're reconciling with google and establishing "author credibility" . As a consequence, your entire blog will rank. By the way, the 2 weeks old test blog is sitting now on page 1 (6th position).

    Before commenting, always remember that google assumes you're a bot till proven innocent. Therefore, to make the comment count, prove you're not a bot by referencing words, phrases or concepts from the blog post. Basically, do what automated commenting can't. Instead of :

    Great article, thanx

    try something like:

    Did not realize that Personal Branding can be that beneficial to one's business. Thanks for the heads up

    Not only the latter comment will likely be approved by the blog owner, googlebot will realize non-humans cannot relate intelligently to a blog post they don't read. Also, when was the last time
    you heard a blog owner disapprove a Twitter signed-in comment ?

    I'd like to remind you this recovery method is a per-page process. Recovering the homepage (which is extremely important) will not recover internal pages if they,too, have issues.

    Lastly, take a break to refuel then simply rinse and repeat the above steps and you should be able to clean everything.

    The process above is based on actual work done on wounded established sites that DID recover.

    Hope this helps.

    N.B.

    Since Penguin algo is still under tweaking, there may be some false positives as well as floating spam. Google has a form where you can report those poor pages that should be caught but weren't:
    Its rare to see these kind of logical explanation in WF these days, thanks for the details. Fed up with the advices like no backlinking, no spinning, end of SEO kind of statements in this forum. Hope fully we will come back to the old glory of sharing useful information .

    Regarding keyword density I would say %7 is a bit more , I prefer to stick to less than %4 or write naturally , which again would be less than 4%.

    On off page, varying anchor text is not BS , and it works like magic. For one specific case i did a few directory submission with raw url as anchor text and I see a nice SERP improvement as of now. I am just watching those as directory submission takes more time to approve, which is good in one way (it looks natural)
    {{ DiscussionBoard.errors[6342825].message }}
    • Profile picture of the author RWhite
      Do not follow this advice off "on page" factors being the reason for penguin. EVERYONE knows that panda is onsite and penguin is OFFSITE.

      This is a wanker of a piece. Makes no sense...and totally goes against what real authorities in the industry have said.

      Give me thanks if I prevented you from wasting your time with these strategies.
      {{ DiscussionBoard.errors[6386446].message }}
  • Profile picture of the author Mary Popowa
    Thanks for good explanation and step-by-step tips, will be very helpful for many people.
    Signature

    You don't have to be great to start, but you have to start to be great ~Zig Ziglar~

    {{ DiscussionBoard.errors[8319511].message }}
  • Profile picture of the author fpdeziner
    Great work man! it is very helpful. After the panda and penguin this is a best post I've read on warrior forum. Vailid points..
    {{ DiscussionBoard.errors[8327714].message }}

Trending Topics