Search Engine Facts
Search Engine Facts

Read our back issues

May 2017

December 2009

November 2009

October 2009

September 2009

August 2009

July 2009

June 2009

May 2009

April 2009

March 2009

February 2009

January 2009

December 2008

November 2008

October 2008

September 2008

August 2008

July 2008

June 2008

May 2008

April 2008

March 2008

February 2008

January 2008

December 2007

November 2007

October 2007

September 2007

August 2007

July 2007

June 2007

May 2007

April 2007

March 2007

February 2007

January 2007

December 2006

December 2006

November 2006

October 2006

September 2006

August 2006

July 2006

June 2006

May 2006

April 2006

March 2006

February 2006

Januray 2006

December 2005

November 2005

October 2005

September 2005

August 2005

July 2005

June 2005

May 2005

August 2005

March 2005

February 2005

January 2005

December 2004

November 2004

October 2004

September 2004

August 2004

July 2004


» Archive
All about software products and antivirus solutions.
Good deals and offers on computers & hardware.
AVG Antivirus offers top security solutions.

Home   Contact   Privacy policy    Partner sites

Are search engine spammers exploiting your sitemaps file?

A recent thread in a webmaster forum indicated that some search engine spammers might exploit the new XML sitemaps files. Has your sitemaps file been abused by spammers? Can using a sitemaps file harm your search engine rankings?

What is a sitemaps XML file?

The big search engines (Google, Yahoo, MSN and Ask) introduced the Sitemaps protocol earlier this year.

In its simplest form, a sitemap is an XML file that lists URLs for a site along with additional metadata about each URL: when it was last updated, how often it usually changes, how important it is, relative to other URLs in the site, etc.

That information helps search engines to more intelligently crawl your site. The Sitemaps protocol is a standard that makes it easier to create a sitemap that can be parsed by all search engines.

How can such a file harm your rankings?

Some webmasters reported problems with duplicate content after adding a sitemaps XML file to their web sites.

The content of their websites appeared on dubious websites that had nothing to do with the original sites. The content of the original websites had been duplicated on many other sites. The result was that the original sites might have received ranking penalties due to duplicate content.

What happened?

Some search engine spammers used the sitemaps XML files to easily find contents for their scraper sites.

A scraper site is a website that pulls all of its information from other websites using automated tools. The scraper software pulls different contents from other websites to create new web pages that are designed around special keywords. The scraped pages usually show AdSense ads with which the spammers hopes to make money.

The new sitemaps XML files make it very easy for scraper tools to find content rich pages. Although the original intention of the sitemaps files was to inform search engines about every single page of your web site, they can also be used to inform spam bots about your pages.

What can you do to avoid problems with your sitemaps file?

One possible solution is not to use any sitemaps file at all. In that case, scraper bots can still parse your web pages through the normal links on your web pages but that would be more difficult for them than using your sitemaps file.

Another solution is to set up a sitemaps file and delete as soon as search engines have indexed that file.

Do not use free sitemap generator tools. You don't know what they will do with your data and they might even use it to create scraper sites with your content.

Unfortunately, there's not much that you can do to stop spammers from abusing your content. Use a tool such as CopyScape to find sites that have duplicated your content.

Copyright - Internet marketing and search engine ranking software

Home   Contact   Privacy policy    Partner sites
May 2007 search engine articles