How to get around Google's spam filters - part
Google tries to keep its search results as clean as possible.
For that reason, they have a variety of spam filters in their
ranking algorithm that try to remove low quality web sites.
If your web site gets caught by one of these filters, it
will be very difficult to get high Google rankings. In this
article series, we're taking a look a the 15 most common
Google spam filters and we'll tell you how you can get around
Google's -30 filter, the Google bomb filter and
the page load filter
Google seems to apply the -30 filter to web sites that
use spammy SEO methods. If Google finds out that your web
pages or similar spam techniques then your rankings will
drop by 30 spots.
The Google bomb filter seems to be applied to web sites
that get too many identical links in a short time period.
If a web site gets many links with exactly the same link
text then Google will downrank the page because such an
unnatural linking behavior indicates a manipulation attempt.
The page load filter is not exactly a filter. Nevertheless,
it can affect your Google rankings. If your web site takes
too long to load then the search engine spiders will time
out and continue with the next web site in the list. That
means that your web site won't be indexed and that it won't
appear in Google's result pages.
How to get around these filters
If the -30 filter has been applied to your web site then
you must remove the spam elements from your web site. After
removing the spam elements from your site, send
a reinclusion request to Google.
It is very important that all spam elements have been
removed from your site before contacting Google. Otherwise,
the reinclusion request won't work. Use white-hat
SEO methods to optimize your web pages.
If the Google bomb filter has been applied to your web
site then you also have to file a reinclusion request.
However, it is better to avoid that Google applies that
filter to your site. Try to get high quality
inbound links with similar but varying link texts.
These links will tell search engines that your web site
is relevant to a special topic.
If you want to avoid that a slow loading web page prevents
search engine spiders from indexing your page, make sure
that you have a reliable
web host. If your web host offers 99% uptime then this
means that your web site can be down for nearly 4 days
per year. If search engines try to index your site when
it is down then it will be removed from the index.
Next week, we'll take a look at three more Google filters
that can cause ranking problems for your web site.