Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Geotargeting duplicate content to different regions - href and canonical tag confusion
-
If you duplicate content onto a sub-folder for say a new US geotargeted site (to target kw spelling differences) and, in addition to GWT geotargeting settings, implement the 'Canonical' and 'Hreflang' tags on these new pages to show G different region and language version (en-us). Then does the original/main site similar pages also need to have canonical and href tags ?
The main/original sites page I don't really want to target a specific country (although existing signals (hosting etc) will be UK (primary target of main site) but pages show up in other country searches too (which we want).
Im presuming fine to leave the original/main site as it currently is although wording in google blog/webmaster central articles etc are a bit confusing hence why im asking for anyone elses opinion/input on this.
Also is there are any benefit (or just best practice) to use 'www.example.com/en-us/...' in the subdirectory URL as opposed to just 'www.example.com/us/'
many thanks in advance to any commentators
-
Many thanks Gianluca !!
-
Hi,
I suggest you both to give a read to this post by DejanSEO, which is quite clear and - IMHO - points to the right interpretation of a somehow confused best practice.
-
Thats what i thought originally but getting confised when i read this page: http://googlewebmastercentral.blogspot.com.au/2011/12/new-markup-for-multilingual-content.html
Specifically this bit:
Annotating pages as substantially similar content
Optionally, for pages that have substantially the same content in the same language and are targeted at multiple countries, you may use the rel="canonical" link element to specify your preferred version. We’ll use that signal to focus on that version in search, while showing the local URLs to users where appropriate. For example, you could use this if you have the same product page in German, but want to target it separately to users searching on the Google properties for Germany, Austria, and Switzerland.
And read in conjunction with this article:
Specifically this bit:
The Effect Of Combining Canonical Tags & Hreflang Tags
Not forgetting that the canonical tags should only be used with content in the same language, when would we use both?
Well firstly, the use of both would involve what I usually call world languages such as English, Spanish, French or Portuguese. These languages are used in many countries and, whilst there are variations between the use of these languages in those countries, the variations are sometimes small.
Additionally, multinational publishers often save costs by using one version of the language for all countries speaking that general language, thus ignoring the regional variations. In other words, for Spain and Mexico, Google is presented with exactly the same content, letter for letter.
The canonical acknowledges that this is the same content. The Hreflang tag identifies which URL should be displayed in different sets of results.
So, in other words, canonical + Hreflang = same content + different URL.
Google knows the content is the same, but displays the correct URL for the Google domain search (e.g. google.com.mx will see the relevant URLs for Mexico displayed in the results).
-
With canonical tag it is a one way road:
You have Page A and Page B with the same content but you want to point out Page A
Page B has a canonical to Page A:
Page B will disappear from the Search Results transferring all the link juice that it has gained to Page A
If you have the same content in different languages then you should use hreflang telling search engines that the two are the same but in other language:
Page A and Page B will have both the following in their headers
This way you will not Geo-Target but Language-Target the two pages ;-)
-
thanks Istvan
but what about whether its a requirement, or suggested best practice, that if you have tags (say canonical) on one set of duplicate pages then you must also add to the other similar/dupe pages (on original site).
Can you have one but not the other without it causing issues or do you need both to stop duplicate issues ?
-
Sorry for responding late, but I somehow forgot to answer this one.
So basically I would consider putting HREFLANG to all of the pages (US, original and any other language). Please note that HREFLANG is connected to optimizing the same content on different languages and not for geo-targeting mainly.
The best example would be Belgium. You can have content in French and in Dutch, still you are optimizing for the same region.
-
Thanks Itsvan, its a good answer and further information! What im really trying to establish though is if its ok to ONLY add canonical & href tags to the US focused subdirectory site ? Do they need to be added to the main site too or can I leave them off (since dont want to geotarget the main site) ? Im confused by wording on google articles/bogs etc on this subject. Since think they say that if you put the tags on a duplicate page you need to also put tags with alternative region/lang tags on the corresponding dupe content page (although i dont want to since want to leave main site free of specific geotargeting). In other words is it a technical requirement/necessity to have tags on both sets of dupe content ?
-
Hi danarchism,
This is what we have on a quite big website:
1. Main site is geo-targeted for a specific country
2. sub-folders of the site are geo-targeted for other countries
3. On each Page in the header we have the HREFLANG to the other 9 languages we use on the site.
Still when we talk about SERP impressions we have many times overlays (Such as the Geo-Targeted content to the Netherlands will appear in the Google.be or Geo-Targeted content to Germany appears in Google.At).
I hope this helped,
Istvan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content, although page has "noindex"
Hello, I had an issue with some pages being listed as duplicate content in my weekly Moz report. I've since discussed it with my web dev team and we decided to stop the pages from being crawled. The web dev team added this coding to the pages <meta name='robots' content='max-image-preview:large, noindex dofollow' />, but the Moz report is still reporting the pages as duplicate content. Note from the developer "So as far as I can see we've added robots to prevent the issue but maybe there is some subtle change that's needed here. You could check in Google Search Console to see how its seeing this content or you could ask Moz why they are still reporting this and see if we've missed something?" Any help much appreciated!
Technical SEO | | rj_dale0 -
Removing a canonical tag from Pagination pages
Hello, Currently on our site we have the rel=prev/next markup for pagination along with a self pointing canonical via the Yoast Plugin. However, on page 2 of our paginated series, (there's only 2 pages currently), the canonical points to page one, rather than page 2. My understanding is that if you use a canonical on paginated pages it should point to a viewall page as opposed to page one. I also believe that you don't need to use both a canonical and the rel=prev/next markup, one or the other will do. As we use the markup I wanted to get rid of the canonical, would this be correct? For those who use the Yoast Plugin have you managed to get that to work? Thanks!
Technical SEO | | jessicarcf0 -
Query Strings causing Duplicate Content
I am working with a client that has multiple locations across the nation, and they recently merged all of the location sites into one site. To allow the lead capture forms to pre-populate the locations, they are using the query string /?location=cityname on every page. EXAMPLE - www.example.com/product www.example.com/product/?location=nashville www.example.com/product/?location=chicago There are thirty locations across the nation, so, every page x 30 is being flagged as duplicate content... at least in the crawl through MOZ. Does using that query string actually cause a duplicate content problem?
Technical SEO | | Rooted1 -
Handling of Duplicate Content
I just recently signed and joined the moz.com system. During the initial report for our web site it shows we have lots of duplicate content. The web site is real estate based and we are loading IDX listings from other brokerages into our site. If though these listings look alike, they are not. Each has their own photos, description and addresses. So why are they appear as duplicates – I would assume that they are all too closely related. Lots for Sale primarily – and it looks like lazy agents have 4 or 5 lots and input the description the same. Unfortunately for us, part of the IDX agreement is that you cannot pick and choose which listings to load and you cannot change the content. You are either all in or you cannot use the system. How should one manage duplicate content like this? Or should we ignore it? Out of 1500+ listings on our web site it shows 40 of them are duplicates.
Technical SEO | | TIM_DOTCOM0 -
Is duplicate content ok if its on LinkedIn?
Hey everyone, I am doing a duplicate content check using copyscape, and realized we have used a ton of the same content on LinkedIn as our website. Should we change the LinkedIn company page to be original? Or does it matter? Thank you!
Technical SEO | | jhinchcliffe0 -
Duplicate content and http and https
Within my Moz crawl report, I have a ton of duplicate content caused by identical pages due to identical pages of http and https URL's. For example: http://www.bigcompany.com/accomodations https://www.bigcompany.com/accomodations The strange thing is that 99% of these URL's are not sensitive in nature and do not require any security features. No credit card information, booking, or carts. The web developer cannot explain where these extra URL's came from or provide any further information. Advice or suggestions are welcome! How do I solve this issue? THANKS MOZZERS
Technical SEO | | hawkvt10 -
Block Quotes and Citations for duplicate content
I've been reading about the proper use for block quotes and citations lately, and wanted to see if I was interpreting it the right way. This is what I read: http://www.pitstopmedia.com/sem/blockquote-cite-q-tags-seo So basically my question is, if I wanted to reference Amazon or another stores product reviews, could I use the block quote and citation tags around their content so it doesn't look like duplicate content? I think it would be great for my visitors, but also to the source as I am giving them credit. It would also be a good source to link to on my products pages, as I am not competing with the manufacturer for sales. I could also do this for product information right from the manufacturer. I want to do this for a contact lens site. I'd like to use Acuvue's reviews from their website, as well as some of their product descriptions. Of course I have my own user reviews and content for each product on my website, but I think some official copy could do well. Would this be the best method? Is this how Rottentomatoes.com does it? On every movie page they have 2-3 sentences from 50 or so reviews, and not much unique content of their own. Cheers, Vinnie
Technical SEO | | vforvinnie1 -
Robots.txt and canonical tag
In the SEOmoz post - http://www.seomoz.org/blog/robot-access-indexation-restriction-techniques-avoiding-conflicts, it's being said - If you have a robots.txt disallow in place for a page, the canonical tag will never be seen. Does it so happen that if a page is disallowed by robots.txt, spiders DO NOT read the html code ?
Technical SEO | | seoug_20050