Duplicate content due to csref
-
Hi,
When i go trough my page, i can see that alot of my csref codes result in duplicate content, when SeoMoz run their analysis of my pages.
Off course i get important knowledge through my csref codes, but im quite uncertain of how much it effects my SEO-results.
Does anyone have any insights in this? Should i be more cautios to use csref-codes or dosent it create problems that are big enough for me to worry about them.
-
Yes, to set up rel-canonical properly, every page that could conceivably be tagged with a csref= parameter should have a self-referencing canonical. The tags are easy to set up, in theory, but once you get into a large site and/or CMS, setting them up on dozens or hundreds of pages can be tricky. Ultimately, it's a more effective approach that has some other benefits (like scooping up stray duplicates that may have been created by other URL parameters), but it really depends on your development resources and how complex your site is.
-
Hi,
Thanks for quick and competent reply.
I guess the reason that google only have registred 8 pages, is that many of the pages we have csref on are campaign pages, and they havent "lived" for so long yet.
As i understand you, there are two ways to proceed with this. One beeing informing in Google Webmaster Tools that google should ignore the csref-parameter, and the other beeing the canonical links.
The first is quite straight forward i guess, its just a matter of registrering in Google Webmaster Tools, that all URLs with www.tryg.dk as main domain, should not be followed by Google.
The latter im not that sure of how to proceed with, its a matter of registrering every page with csref with a cannoical link? Or how is the best way to proceed with that.
-
The good news is that you only seem to have about 8 of these pages in the Google index. You can use this query on Google to see them:
site:www.tryg.dk inurl:csref
Ideally, I'd use the canonical tag on those pages to strip out the parameter and de-index any duplicates, but across the site that can be tricky. You could also tell Google Webmaster Tools to ignore the csref parameter via parameter handling - it's not quite as robust a solution, but it's a lot easier to implement.
-
Hi,
Thanks for your reply.
It is excactly a URL generated based on our tracking codes, f.eg. when i look in the list of duplicated content on our page here in SeoMoz, i get the following URLs:
http://www.tryg.dk/om-tryg/fakta-om-tryg/samarbejdspartnere/index.html?csref=Disclaimer_Nordea http://www.tryg.dk/om-tryg/fakta-om-tryg/samarbejdspartnere/index.html?csref=Bundmenu_Om_Tryg_Vores_Partnere
http://www.tryg.dk/om-tryg/fakta-om-tryg/samarbejdspartnere/index.html
The latter beeing the "original" page for this, and the two above beeing page URLs generated by URLs with csrefs, which are generated by our tracking via Omniture.
So my question is how i make sure that it do not have a negative effect on our SEO.
-
Apologies, but I'm not familiar with the csref parameter - could you tell me what information it passes or give me a sample URL (you can make it generic and mask your domain info)?
It sounds like some kind of tracking code, in which case it can definitely start to create duplicate content issues. You could probably use the rel=canonical tag to make Google "collapse" those pages, or you could tell Google to ignore the parameter in Google Webmaster Tools. Neither should impact your tracking.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content issue
Hi, A client of ours has one URL for the moment (https://aalst.mobilepoint.be/) and wants to create a second one with exactly the same content (https://deinze.mobilepoint.be/). Will that mean Google punishes the second one because of duplicate content? What are the recommendations?
Technical SEO | | conversal0 -
WordPress Duplicate Content Caused By Categories
Hello, We have a wordpress blog that has around 250 categories. Due to our platform we have a hierarchy structure for 3 separate stores. For example iPhone > Apps > Books. Placing a blog post in the books category automatically places it into iPhone and iPhone/Apps category, causing 3 instances of any blog post in this category. Is this an issue? I have seen 2 schools of thought on categories, 1 index follow and 2 noindex follow. I know some of our categories get indexed, but with so many, maybe it is better to noindex them. We also considered reducing our categories to 10 to 12 and use tags to provide the indexed site navigation as follows: Reviews (category) iPhone Book App, iPhone App Store (tags) but this seems a little redundant? Anyone want to take this on? thank you Mike
Technical SEO | | crazymikesapps10 -
GWT Duplicate Content and Canonical Tag - Annoying
Hello everyone! I run an e-commerce site and I had some problems with duplicate meta descriptions for product pages. I implemented the rel=canonical in order to address this problem, but after more than a week the number of errors showing in google webmaster tools hasn't changed and the site has been crawled already three times since I put the rel canonical. I didn't change any description as each error regards a set of pages that are identical, same products, same descriptions just different length/colour. I am pretty sure the rel=canonical has been implemented correctly so I can't understand why I still have these errors coming up. Any suggestions? Cheers
Technical SEO | | PremioOscar0 -
Duplicate Content Issue
SEOMOZ is giving me a number of duplicate content warnings related to pages that have an email a friend and/or email when back in stock versions of a page. I thought I had those blocked via my robots.txt file which contains the following... Disallow: /EmailaFriend.asp Disallow: /Email_Me_When_Back_In_Stock.asp I had thought that the robot.txt file would solve this issue. Anyone have any ideas?
Technical SEO | | WaterSkis.com0 -
Does turning website content into PDFs for document sharing sites cause duplicate content?
Website content is 9 tutorials published to unique urls with a contents page linking to each lesson. If I make a PDF version for distribution of document sharing websites, will it create a duplicate content issue? The objective is to get a half decent link, traffic to supplementary opt-in downloads.
Technical SEO | | designquotes0 -
How to prevent duplicate content in archives?
My news site has a number of excerpts in the form of archives based on categories that is causing duplicate content problems. Here's an example with the nutrition archive. The articles here are already posts, so it creates the duplicate content. Should I nofollow/noindex this category page along with the rest and 2011,2012 archives etc (see archives here)? Thanks so much for any input!
Technical SEO | | naturalsociety0 -
What to do about similar content getting penalized as duplicate?
We have hundreds of pages that are getting categorized as duplicate content because they are so similar. However, they are different content. Background is that they are names and when you click on each name it has it's own URL. What should we do? We can't canonical any of the pages because they are different names. Thank you!
Technical SEO | | bonnierSEO0 -
Root domain not resolving to www. Duplicate content?
Hi, I'm working with a domain that stays on the root domain if the www is not included. But if the www is included, it stays with the www. LIke this: example.com
Technical SEO | | HardyIntl
or
www.example.com Of course, they are identical and both go to the same IP. Do search engines consider that to be duplicate content? thanks,
michael0