Maile Ohye on the Google Webmaster Central blog has written an excellent article detailing “duplicate content issues caused by URL parameters” issue:
How can URL parameters, like session IDs or tracking IDs, cause duplicate content? When user and/or tracking information is stored through URL parameters, duplicate content can arise because the same page is accessible through numerous URLs. It’s what Adam Lasnik referred to in “Deftly Dealing with Duplicate Content” as “store items shown (and — worse yet — linked) via multiple distinct URLs.” In the example below, URL parameters create three URLs which access the same product page.
Why should you care? When search engines crawl identical content through varied URLs, there may be several negative effects:
2. Search results may display user-unfriendly URLs (long URLs with tracking IDs, session IDs)
* Offsets branding efforts
How we help users and webmasters with duplicate content: We’ve designed algorithms to help prevent duplicate content from negatively affecting webmasters and the user experience.
2. We select what we think is the “best” URL to represent the cluster in search results.
3. We then consolidate properties of the URLs in the cluster, such as link popularity, to the representative URL.
If you find you have duplicate content as mentioned above, can you help search engines understand your site? First, no worries, there are many sites on the web that utilize URL parameters and for valid reasons. But yes, you can help reduce potential problems for search engines by:
2. Submitting a Sitemap with the canonical (i.e. representative) version of each URL. While we can’t guarantee that our algorithms will display the Sitemap’s URL in search results, it’s helpful to indicate the canonical preference.
How can you design your site to reduce duplicate content? Because of the way Google handles duplicate content, webmasters need not be overly concerned with the loss of link popularity or loss of PageRank due to duplication. However, to reduce duplicate content more broadly, we suggest:
2. Use a cookie to set the affiliateID and trackingID values.
If you follow this guideline, your webserver logs could appear as:
127.0.0.1 – – [19/Jun/2007:14:40:45 -0700] “GET /product.php?item=swedish-fish HTTP/1.1” 200 74
And the session file storing the raw cookie information may look like:
How can we better assist you in the future? We recently published ideas from SMX Advanced on how search engines can help webmasters with duplicate content. If you have an opinion on the topic, please join our conversation in the Webmaster Help Group (we’ve already started the thread).
Google, Search Engine, Duplicate Content, URL, Sitemaps, Crawling, Site indexing