How Not to Optimize Your Google Site Ranking

8 replies
  • SEO
  • |
Note: Before I begin, let me say that I am quite sure that not everyone will agree with what that I have written below. And that’s fine, because that’s exactly what a forum is for: discussing various ideas, opinions and viewpoints, and debating the pros and cons. That being said, let me say that these tips are my own opinions based on my own experience, and therefore may be different than the experiences of other members of this forum.


How Not to Optimize Your Google Site Ranking
Most internet marketers know the things you can do to increase your site’s ranking, so I’m going to present some things you shouldn’t do – that is, things that can actually decrease your site’s PageRank.

Avoiding Things Google Doesn’t Like

·Long and complicated URLs. The shorter, and more straightforward, a page’s URL, the better. Google doesn’t like long URLs, nor does it like URLs that contain special characters. To that end, don’t use codes like &id= as a parameter in your URLs; similarly, avoid the use of dynamic URLs. Google indexes only static URLs, not those that are dynamically generated. If your site does include dynamic pages, use a URL rewrite technique to turn your dynamic URLs into static ones. (Before anyone asks, I do not know how to do this. I have a friend that does that sort of thing for me. You can find tutorials on the internet that teach you how, though, such as the one located at Search Engine-Friendly URLs [PHP & MySQL Tutorials] )

·Splash pages. Here’s the thing about splash pages: Nobody likes ‘em. Visitors don’t like them, because they make it take longer to get into your site. And Google doesn’t like them, either, because they don’t include the types of internal links and menus found on a true home page; nor do they include much, if any, text content. (If there are no links to other pages from the first page that GoogleBot encounters, it won’t know about the other pages on your site.) So, you should make your site’s first page a traditional home page, not a splash page; both your visitors and the searchbots will be much happier.

·High keyword density. Although keywords are good, too many of them are bad. (Google assumes that you are including keywords just to get a high PageRank, instead of creating unique content for your users.) So don’t include too many keywords (especially repeated keywords) in your content or <META> code. If a page has too high a keyword density, Google will categorize your page as a doorway page – and penalize you accordingly.

·Hidden text. This is when you disguise keywords or links by making them the same or similar color as the page background, using a tiny font size, or hiding them within the HTML code itself. Many webmasters think this is a clever way to stuff lots of keywords onto a page without looking as if they are doing so. Unfortunately, it’s hard to trick a searchbot; GoogleBot sees hidden text just as easily as it does the regular text on your page – which means they will see all the keywords you’re trying to stuff. And, as may you know, searchbots don’t like keyword stuffing.

·Duplicate content. This is another false trick that too-savvy webmasters sometimes employ, to their own detriment – duplicating content on a site, in the hopes of increasing the number of hits in search engine results. This may be done by putting the same content on multiple pages, or via the uses of multiple domains or subdomains. Unfortunately, duplicating content is a bad idea, as Google utilizes duplicate content filters that will identify and remove duplicate sites from their search results. When you duplicate your content, you run the risk that your main site or page will be filtered out – while the subsidiary content remains.

·Bad outbound links. You know that the quality of your inbound links matters, but did you know that the quality of your outbound links can also affect your search rankings? Google considers the sites you link to as part of its page-ranking process – at least when it comes to overtly low-quality sites. For that reason, you don’t want to link to a site that’s been dropped or banned from a search engine’s index, as you could get punished simply by association.

·Images, videos and animations. This is a simple concept – searchbots read text (including HTML code); they can’t view images, videos or Flash animations. Which means, of course, that creating an image- or animation-heavy page renders that page virtually invisible to Google’s spider. Your page might look great to visitors, but if GoogleBot can’t see it, it won’t get indexed. When in doubt, go with a text-based approach.

·Big pages. This is another reason to go easy on web page images: GoogleBot doesn’t like big pages. More specifically, it doesn’t like long load times. If it takes too much time for all the elements on a page to load, GoogleBot will give up and move on to the next page in its queue – which, obviously, is bad for you.

·JavaScript code. As I stated earlier, Google’s spider reads text and HTML code – well, some HTML code. GoogleBot will ignore JavaScript code in your HTML, which means anything you have in a script won’t be indexed. This is particularly troublesome if you use a JavaScript menu, as the searchbots may not see all the internal links you have in your menus.

·Too much code. Speaking of HTML code, don’t overdo it, either. Having more code than you do actual text on your page will cause GoogleBot to give up before your entire page is crawled. You should avoid employing too many code-heavy effects, such as nested tables or JavaScript effects. If your important text is buried under hundreds of lines of code, you will be at a disadvantage compared to a well-optimized site.

·Invalid HTML. One last thing when it comes to the coding of your site: Don’t create messy code. This is one instance where neatness counts; invalid HTML can confuse GoogleBot’s searchbot and cause it to miss important content.

Deliberate Practices to Avoid

In the previous section I talked about things you might accidentally do that can adversely affect your PageRank rating. There are also some practices that sneaky web designers deliberately do to increase their page rank; Google does take issue with these practices, and can ban you from their index if you are caught.

To that end, here are some of the more nefarious “outlawed” optimization practices:

*Google bombing. Sometimes called Google washing or link bombing, this is an attempt to increase your PageRank by having a large number of sites link to a page by using identical anchor text. For example, you might register several domains and have them all link to a single site using the same anchor text for the links. Searching for the term used in the link anchor text will return the linked-to-site high in the search results. (Google bombing occurs often in blogs, where a site owner will “bomb” multiple blog postings with replies linking to the owner’s site.)

*Keyword stuffing. This is when you insert hidden, random text on a page to increase the keyword density, and thus increase the apparent relevancy of a page. For example, if your page is about trains, you might insert several lines of invisible text at the bottom of the page repeating the keyword train, over and over, or include multiple instances of the word train in a <META> tag. In the past, some search engines simply counted how often a keyword appeared on a page to determine relevance; today, however, Google employs algorithms to detect keyword stuffing. (A related technique is meta tag stuffing, where keywords are stuffed into HTML meta tags.)

*Doorway pages. Doorway pages are also known as gateway pages, landing pages, and entry pages. These are web pages that are low in actual content, and instead stuffed with repeating keywords and phrases designed to increase the page’s search rank. Doorway pages typically require visitors to click a “click here to enter” link to enter the main website; in other instances, visitors to a doorway page are quickly redirected to another page.

*Link farms. This is a group of pages that all link to one another. The purpose of a link farm is to increase the number of links to a given site; since PageRank is at least partially driven by the number of linked-to pages, using a link farm can make it appear as though a large number of sites are linking to a given site.

*Mirror websites. This is the hosting of multiple websites, all with the same content, but using different URLs. The goal is to increase the likelihood that any one (or more) of the mirror sites will appear on Google’s search results page.

*Cloaking. This is an attempt to mislead Google by serving up a different page to the GoogleBot crawler than will be seen by human visitors. This is sometimes used for code swapping, where one page is optimized to get a high ranking, and then swapped out for another page with different content.

*Scraper sites. This is a site that “scrapes” results pages from Google and other search engines to create phony content for a website. A scraper site is typically full of clickable ads.

The Bottom Line

Google makes it easy to get listed in its web index, but getting listed is just the start. Making sure that your listing appears high in Google’s search results requires a lot of hard work, a bit of technical savvy and, even more importantly, useful and relevant content. Too many webmasters get hung up on arcane search optimization techniques, when they should be focusing more on creating sites that more visitors actually want to visit. If you spend your time, instead, on creating a compelling and well-designed website, a high Google ranking will, more often than not, naturally follow.
#google #optimize #ranking #site
  • Profile picture of the author tomcam
    All good stuff, dude. This is a pretty strong overview of what appears to be Google's current thinking.

    Minor disagreement: By "messy code" I think you mean invalid HTML, with missing closing brackets and so on. Valid HTML is valid, doesn't matter if it looks good or not. In fact there are HTML obfuscators that rearrange the code to look "messy" but which keep the code valid. That way spiders can read it but people can't. It slows down people who are trying to steal the HTML.
    {{ DiscussionBoard.errors[1690781].message }}
    • Profile picture of the author TheJedi
      Originally Posted by tomcam View Post

      All good stuff, dude. This is a pretty strong overview of what appears to be Google's current thinking.

      Minor disagreement: By "messy code" I think you mean invalid HTML, with missing closing brackets and so on. Valid HTML is valid, doesn't matter if it looks good or not. In fact there are HTML obfuscators that rearrange the code to look "messy" but which keep the code valid. That way spiders can read it but people can't. It slows down people who are trying to steal the HTML.
      I have to agree. I believe your term of "invalid HTML" does makes it a bit clearer. Thank you for pointing that out.

      I thought that HTML obfuscators were designed to help protect your web pages from robots, spiders and other collection utilities, such as the automatic scrapers and email harvesters, from indexing or gathering information from your page, as well as deterring code thieves, of course. So you are saying then that the pages can still be indexed? Or maybe it just depends in which software one uses?

      EDIT: I went back and changed the original text to read 'Invalid HTML' instead of 'messy code' because it makes my point more clearly.
      Signature

      {{ DiscussionBoard.errors[1690817].message }}
      • Profile picture of the author JohnMcCabe
        Your comment on bloated code is a good one. I have an old site that had a lot of javascript and css stuff in the header - you had to scroll two or three times to find the actual body content.

        I put most of that stuff in a couple of external files, one a remote .js and one .css, reducing the bloated mass to two lines. My rankings improved with no other changes.

        The side benefit is that I can tweak the code and the style by modifying one page each, rather than having to do a search and replace on all pages...

        Another tip, for those who still format pages using tables...

        Don't put the entire page into one big table. Nothing will be visible until the entire table finishes loading, including any images and resources drawn from other sites (like ads).

        If you have to keep using tables, break the page up into multiple tables, so that the page has something visible immediately.
        {{ DiscussionBoard.errors[1692200].message }}
      • Profile picture of the author tomcam
        Originally Posted by TheJedi View Post

        I thought that HTML obfuscators were designed to help protect your web pages from robots, spiders and other collection utilities, such as the automatic scrapers and email harvesters, from indexing or gathering information from your page, as well as deterring code thieves, of course. So you are saying then that the pages can still be indexed? Or maybe it just depends in which software one uses?
        Obfuscators are for deterring code thieves, period. By definition they do not change the function of the code. They can be indexed just as easily as non-obfuscated pages.
        {{ DiscussionBoard.errors[1725176].message }}
  • Profile picture of the author kevinfahey
    Excellent read.... they dropped one of my sites from PR2 to PR1 and the only reason I can think of it I placed a banner at the top of the page with an affiliate link =?idxxxxx.
    The only thing I changed in weeks... Would I be right in saying this?
    Signature

    Need Help Building Your Business? Online IM Support

    {{ DiscussionBoard.errors[1692240].message }}
    • Profile picture of the author TheJedi
      Originally Posted by kevinfahey View Post

      Excellent read.... they dropped one of my sites from PR2 to PR1 and the only reason I can think of it I placed a banner at the top of the page with an affiliate link =?idxxxxx.
      The only thing I changed in weeks... Would I be right in saying this?
      PR2 to PR1 is not a drop. That is a front page ranking. Congrats!
      Signature

      {{ DiscussionBoard.errors[2371560].message }}
  • Profile picture of the author TheJedi
    The order of the links in your page does make a difference to the GoogleBot. It assumes that the most important links are located at the top of the page, and then it works its way down to the least important (at least according to the GoogleBot's sense of reasoning), so perhaps having the affiliate link (to an external site) at the very top of the page is not such a good idea.
    Signature

    {{ DiscussionBoard.errors[1694593].message }}
  • Profile picture of the author lexilexi
    Super good article, many thanks. One tip to add, if you have pages that host "videos and not much else" you can maybe provide a text transcript of the video if appropriate, in order to add some text content to the page.
    Signature

    "If there is no door, it becomes necessary to break out through the wall."

    {{ DiscussionBoard.errors[2372130].message }}

Trending Topics