Duplicate content & canonicals
-
Hi,
Working on a website for a company that works in different european countries.
The setup is like this:
www.website.eu/nl
www.website.eu/be
www.website.eu/fr
...You see that every country has it's own subdir, but NL & BE share the same language, dutch...
The copywriter wrote some unique content for NL and for BE, but it isn't possible to write unique for every product detail page because it's pretty technical stuff that goes into those pages.
Now we want to add canonical tags to those identical product pages. Do we point the canonical on the /be products to /nl products or visa versa?
Other question regarding SEOmoz: If we add canonical tags to x-pages, do they still appear in the Crawl Errors "duplicate page content", or do we have to do our own math and just do "duplicate page content" minus "Rel canonical" ?
-
Hey Joris,
As of now it will most likely see it as duplicate content, because technically it still is duplicate content to a crawler bot, they won't know your intentions or target audience for each subfolder. The only way you could get around our crawler seeing it as duplicate is by blocking rogerbot with robots.txt or meta robots from that subfolder. Then there is putting up relconanoicals, which is the best way.
Hope this sheds some light on the duplicate content issues.
Best,
Nick
SEOmoz -
Thanks Robert!
-
Will do!
-
Now, that was a good question. Why not send a quick email to help@SEOmoz.org and just ask if there is a way to circumvent? LMK please.
-
Hi Robert,
Thx for your quick answer, I will make sure that in Google Webmaster Tools we say that the /be is for Belgium and the /nl for The Netherlands, but the duplicate content will still show up in our reports in SEOmoz, no?
-
First question is: Have you thought of using the .cc instead of the sub directory? Rand speaks to the .fr issue in his WBF mentioned by iBiz Leverage.
As to canonical to avoid duplicate content, you shouldn't have a duplicate content issue even with the two languages so long as you set your country target for each. But, read or watch the WBF by Rand as it is full of info on this subject and domain auth, etc.
-
I have same problem and found this URL: http://www.youtube.com/watch?v=Ets7nHOV1Yo
Here is also another link from SEOmoz; i think this is most helpful: http://www.seomoz.org/blog/international-seo-where-to-host-and-how-to-target-whiteboard-friday
Hope this can help.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Large site with content silo's - best practice for deep indexing silo content
Thanks in advance for any advice/links/discussion. This honestly might be a scenario where we need to do some A/B testing. We have a massive (5 Million) content silo that is the basis for our long tail search strategy. Organic search traffic hits our individual "product" pages and we've divided our silo with a parent category & then secondarily with a field (so we can cross link to other content silo's using the same parent/field categorizations). We don't anticipate, nor expect to have top level category pages receive organic traffic - most people are searching for the individual/specific product (long tail). We're not trying to rank or get traffic for searches of all products in "category X" and others are competing and spending a lot in that area (head). The intent/purpose of the site structure/taxonomy is to more easily enable bots/crawlers to get deeper into our content silos. We've built the page for humans, but included link structure/taxonomy to assist crawlers. So here's my question on best practices. How to handle categories with 1,000+ pages/pagination. With our most popular product categories, there might be 100,000's products in one category. My top level hub page for a category looks like www.mysite/categoryA and the page build is showing 50 products and then pagination from 1-1000+. Currently we're using rel=next for pagination and for pages like www.mysite/categoryA?page=6 we make it reference itself as canonical (not the first/top page www.mysite/categoryA). Our goal is deep crawl/indexation of our silo. I use ScreamingFrog and SEOMoz campaign crawl to sample (site takes a week+ to fully crawl) and with each of these tools it "looks" like crawlers have gotten a bit "bogged down" with large categories with tons of pagination. For example rather than crawl multiple categories or fields to get to multiple product pages, some bots will hit all 1,000 (rel=next) pages of a single category. I don't want to waste crawl budget going through 1,000 pages of a single category, versus discovering/crawling more categories. I can't seem to find a consensus as to how to approach the issue. I can't have a page that lists "all" - there's just too much, so we're going to need pagination. I'm not worried about category pagination pages cannibalizing traffic as I don't expect any (should I make pages 2-1,000) noindex and canonically reference the main/first page in the category?). Should I worry about crawlers going deep in pagination among 1 category versus getting to more top level categories? Thanks!
Moz Pro | | DrewProZ1 -
Duplicate content report - question on best practice
Hello all, New to MOZ Pro and SEO - so lots to get my head round! I’m working through the Duplicate Content section of the Crawl report and am not sure what the best practice is for my situation. Background: We are a reference guide for luxury hotels around the world, but the hotels that are featured on the site vary year on year. When we add a new hotel page, it sets up the url as ourwebsite.com/continent/country/regionORcity/hotel. When the hotels come off, I redirect their URL to the country or region where we have other hotels. Example: http://www.johansens.com/europe/switzerland/zermatt/ The hotel in Zermatt has come off the site, showing 0 results on this landing page. Question: My duplicate content report is showing a number of these regional pages that are displaying the copy “0 places - Region’ because the hotel has come off, but the landing page is still live. Should I redirect the regional page back to the main country page? And then if I add a new hotel to the site from that region in the future, simply remove the redirect? Should I also delete the page? Any tips would be much appreciated!
Moz Pro | | CN_Johansens0 -
Duplicate content nightmare
Hey Moz Community I ran a crawl test, and there is a lot of duplicate content but I cannot work out why. It seems that when I publish a post secondary urls are being created depending on some tags and categories. Or at least, that is what it looks like. I don't know why this is happening, nor do I know if I need to do anything about it. Help? Please.
Moz Pro | | MobileDay0 -
Duplicate content on SearchResults.asp
hi guys. I'm currently working through the reported crawl errors in Moz Analytics, but an unsure what to do about some of them. for example... Searchresults.asp?search=frankie+says+relax is showing as having duplicate page content and page title as SearchResults.asp?searching=Y&sort=13&search=Frankie+Says+Relax&show=24 There's all sorts of searchresults.asp page being flagged. Is this something i can safely ignore or is it something i should endeavour to rectify? I'm also getting errors reported on shoppingcart.asp pages as well as pindex.asp (product index). I'm thinking i should maybe add disallow/ shoppingcart.asp to my robots text file, but am unsure as to whether i should be blocking robots from the search results pages and product index (which is essentially a secondary sitemap). Any advice would be greatly appreaciated. Thanks, Dave 🙂
Moz Pro | | giddygrafix0 -
Duplicate content in crawl despite canonical
Hi! I've had a bunch of duplicate content issues come up in a crawl, but a lot of them seem to have canonical tags implemented correctly. For example: http://www.alwayshobbies.com/brands/aztec-imports/-catg=Fireplaces http://www.alwayshobbies.com/brands/aztec-imports/-catg=Nursery http://www.alwayshobbies.com/brands/aztec-imports/-catg=Turntables http://www.alwayshobbies.com/brands/aztec-imports/-catg=Turntables?page=0 Aztec http://www.alwayshobbies.com/brands/aztec-imports/-catg=Turntables?page=1 Any ideas on what's happening here?
Moz Pro | | neooptic0 -
Duplicate Content
My website is hosted by Hubspot. With each blog I write I can tag them to be listed in a specific category. As an example, one blog article my have three tags or categories that it fits in. Seomoz is seeing this as a duplication of content. in other words, if you go to the different category pages the same article would be listed on all three pages, even though it is just one article. However, I only have 36 duplicate content warnings and I have 150 blog articles, each having 2 or 3 tags (categories.), so there should be many more than 36 duplications. Is this something that affects my seo, or should I just ignore the problem and check these warnings as fixed? Thanks,
Moz Pro | | Rong
Ron0 -
Ask a QuestionIs it possible to duplicate a report from one campaign to another one ?
Hi, Maybe it's easy to do and I've not seen it but I wonder wether or not it's possible to duplicate a report I've personnalised from one campaign to another one. I'd don't want to rewrite it and i'd like to have the same structure within all my reports ! Thanks,
Moz Pro | | Spleen0 -
302 (Temporary Redirect) & HTTPS - how to improve for SEO?
Hello, I have a crawl diagnostic warning that says 302 (temporary Redirect). My site is set up with HTTPS - is that the main reason I am getting this warning? I don't have a strong need for higher security SSL on my site, which is what I understand the main purpose of HTTPS to be so, to improve SEO, would I be better off getting rid of this and going with HTTP? Or would 301 redirects be a better option? Thanks for your help! Niall
Moz Pro | | NiallTom0