Hreflang tag implentation
-
Hi,
We've had hreflang tags implemented on our site for a few weeks now, and while we are seeing some improvements for the regional subfolders I wanted to double check I had the tags implemented correctly (a couple of examples are below). However while the regional subfolder sites are now ranking instead of the US site for some keywords, some key search terms are still returning the US site. Could this be due to incorrect implementation for that specific page?
Due to complications with using Magento we're implementing the tags in the site maps. Also magento appears to be inserting a rel canonical tag automatically for each page and self referencing e.g. On www.example.com/uk/security-cameras (one of the pages we're having issues with) the canonical tag is http://www.example.com/uk/security-cameras" />. Is this an issue?
Any advice would be appreciated. Thanks.
<url><loc>http://www.example.com/uk/dvrs-kits</loc>
<lastmod>2014-07-23</lastmod>
<changefreq>daily</changefreq>
<priority>0.5</priority></url>
<url><loc>http://www.example.com/uk/dvrs-kits/1080p</loc>
<lastmod>2014-07-23</lastmod>
<changefreq>daily</changefreq>
<priority>0.5</priority></url> -
Thanks.. it took me a while but I did realise that one eventually.
-
Thanks Thomas. That is how our sitemap now reads. It just seems odd that when you access map via a browser the hreflang tags don't appear...
-
The answers you received here in the thread are correct:
your sitemap implementation is wrong, because it commits a classic mistake: self-referencing hreflang annotations are missing.
In other words, you must declare that the URL http://www.example.com/uk/dvrs-kits is the one Google must show to English speaking users in the UK.
So you must add this:
Seeing the changes taking place in the SERPs is not immediate.
-
So I will test the site map agin after you made the XML sitemap did you tell Google about language and country targeting? Use tell Google Webmaster tools to Geo target the subfolders? http://www.google.com/webmasters/ Check your listing all over the world http://www.isearchfrom.com/ https://support.google.com/webmasters/answer/2620865?hl=en&ref_topic=2370587 You need the site map to look like this. <url><loc>http://www.example.com/english/</loc> **This is the URL we want to be indexed** <xhtml:link rel="alternate" <strong="">This tells Google there is an alternate version of the URL hreflang="de" **This tells Google the language we’re targeting with the alternate version (German)** href="http://www.example.com/deutsch/" **This tells Google the URL of the alternate version** /> <xhtml:link rel="alternate" <strong="">This tells Google that there is an alternate version of the URL hreflang="de-ch" **This tells Google the language and country we’re targeting with the alternate version (German and Switzerland)** href="http://www.example.com/schweiz-deutsch/" **This tells Google the URL of the alternate version** /> <xhtml:link rel="alternate" <strong="">This tells Google that there is an alternate version of the URL hreflang="en" **This tells Google the language we’re targeting with the alternate version (English)** href="http://www.example.com/english/" **This tells Google the URL of the alternate version** /></xhtml:link></xhtml:link></xhtml:link></url>
Important note - the example above is only for one URL which has three alternate versions. The following code is an example of a full XML sitemap which includes three URLs, each with three alternate versions.
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9" xmlns:xhtml="http://www.w3.org/1999/xhtml"><url> <loc>http://www.example.com/english/</loc> </url> <url> <loc>http://www.example.com/deutsch/</loc> </url> <url> <loc>http://www.example.com/schweiz-deutsch/</loc> </url></urlset> I will use Deep Crawl to look for hreflang it site maps. You might want to use the header way if it is not working. Thomas
-
I've updated the site maps but the hreflang tags still aren't showing when you view the site map...
-
I think I might see the problem. I'm only referencing the alternative pages with the hreflang tag, I've not included the site itself. it should be:
<url><loc>http://www.swann.com/uk/dvrs-kits</loc>
<lastmod>2014-07-23</lastmod>
<changefreq>daily</changefreq>
<priority>0.5</priority></url> -
This is the beginning of my sitemap for the UK subfolder, as far as I can see we're following both the case studies you posted and the Google advice on the structure. But as you point out when you view the sitemap online it's not showing the hreflang tags.
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9" xmlns:xhtml="http://www.w3.org/1999/xhtml"><url><loc>http://www.swann.com/uk/dvrs-kits</loc>
<lastmod>2014-07-23</lastmod>
<changefreq>daily</changefreq>
<priority>0.5</priority></url>
<url><loc>http://www.swann.com/uk/dvrs-kits/1080p</loc>
<lastmod>2014-07-23</lastmod>
<changefreq>daily</changefreq>
<priority>0.5</priority></url>
<url><loc>http://www.swann.com/uk/dvrs-kits/960h</loc>
<lastmod>2014-07-23</lastmod>
<changefreq>daily</changefreq>
<priority>0.5</priority></url></urlset> -
-
- Here is the next test
- Use this link to see in the the site
- http://crawl.blueprintmarketing.com/projects/reports/69930?ro=991e4a1c6d5086bd0bd4d0965e3b6037ed69b692
- http://cl.ly/3G2h3u464400
- Files
- http://cl.ly/3G2h3u464400
- http://searchengineland.com/how-to-implement-the-hreflang-element-using-xml-sitemaps-123030
- https://search.nerdydata.com/search/#!/searchTerm=http://www.swann.com//searchPage=1/sort=pop
- sitemap
- http://cl.ly/text/1d2E2h3P022J
- http://cl.ly/text/0t0T3T2p210h
- https://cl.ly/Xjb3
-
Thanks for running this.
The deeptrawl says that there are no hreflang tags in place. If you view the sitemap.xml files on the site you can't see any there either. However they're definitely in the files I'm giving to my colleague to upload.
I matched the structure of the tag to the two case studies you put in your first post...
-
Here is crawl 1
- http://cl.ly/3u0Z3N0S3m1B
- Interactive URL
- http://crawl.blueprintmarketing.com/projects/reports/69929?ro=0de594f6c4b262001d6d234c282d8ade7e42c020
- http://crawl.blueprintmarketing.com/report_grid/trend/69929/pages_without_hreflang
- &
- http://crawl.blueprintmarketing.com/report_grid/69929?repname=pages_without_hreflang
- I am still running a 2ed full crawl it will post in the AM
Tom
-
You could use the site map however I've had better luck with other methods. If you would like me to I am more than happy to run deep crawl on your site and figure out the problem with your current setup via site map
I have Deepcrawl & will use it to check.
I am referencing this from
http://moz.com/blog/hreflang-behaviour-insights
"Section 4: Tools for the serious International SEO
Essentials:
- Reliable rank tracker that can localize: Advanced Web Ranking, Moz, etc...
- Crawler that can validate hreflang annotations in XML sitemaps or within : The only tool on the market that can do this, and does it very well, is Deepcrawl.
Other nice-to-haves:
- Your own method of "gathering" international search results on scale. You should probably go with proxies.
- Your own method of parsing XML sitemaps and cross checking (even if you use something like Deepcrawl, you'll need to double check).
- Obvious, but worth a reminder: Google webmaster tools, Analytics, access to server logs so you can understand Google's crawl behaviour."
Please look for the report it in the AM,
Tom
-
Hi,
I did have a read through those two case studies previously. My understanding was that we could implement the code purely on the on the sitemap, and that nothing on the site was needed. Are you saying we also need some code on the site as well?
The verification tool doesn't say we have it implemented, but I wasn't sure that it would check the sitemaps or just the sites HTML.
-
Run the site through the tools below
it appears that your only running the site map and not the code itself a good look at the code need it is found in URL below
- http://moz.com/blog/using-the-correct-hreflang-tag-a-new-generator-tool
- http://www.seerinteractive.com/blog/case-study-the-impact-of-hreflang-tag
1 validator
- Validator http://flang.dejanseo.com.au/
- 2 site map tool
- site map tool http://www.themediaflow.com/tool_hreflang.php
- http://www.stateofdigital.com/hreflang-sitemap-tool/
- hreflang
- Generator http://www.internationalseomap.com/hreflang-tags-generator/
- From Google
- https://support.google.com/webmasters/answer/189077?hl=en
For language/country selectors or auto-redirecting homepages, you should add an annotation for the hreflang value "x-default" as well:
Thomas
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is having one blanket h1 tag across my site bad?
I've been working on figuring out if my site host (jigsy) is hindering my seo. The site editor doesn't allow me to assign individual h1 tags for each page, it just blankets one across the whole site as part of a theme.....is this normal?...would it be better to have individual h1s?....is that a thing? Or should I work on making a better one since all my pages should technically be related to counseling/therapy? Site: www.joeborders.com
Technical SEO | | joebordersmft1 -
Google is shortening many of my title tags although they are already quite concise
Hi, Title tags of our website are being truncated by Google even though they can be very short (sometimes < 40 characters) and with very few capital letters. We would like to understand why. Example: Principal component analysis (pca) in abcde - OurBrand shows up as: Principal component analysis (pca) in abcd... - OurBrand where abcde is the name of a very common software (5 characters), and OurBrand is a 6 characters long string (could be used in either lower case or upper case). Even when removing the brackets around pca, truncation still occurs... Any clue why?
Technical SEO | | trigaudias1 -
Duplicate Title Tags
Webmaster Tools is showing that I have 78 duplicate title tags. Here is an example of tags they deem to be duplicate: www.lifeinsure.com/blog/life-insurance-trusts/ www.lifeinsure.com/blog/life-insurance-trusts www.lifeinsure.com/life-insurance-trusts/ I have done 301 redirects from any posts beginning with "/blog/" to "/" Any idea why this is happening?
Technical SEO | | rdreich490 -
Negative Effect of Duplicate Title Tags
Hi guys, I'm trying to figure out if one my problems with search rankings might be a result of 2 of my main pages (home page and tier 2 page linked to main menu) have the same title and description tags. Let's say my keyword is 'building paper airplanes'. The url for my homepage is http://domain.com and the Title Tag is "Building Paper Airplanes I Domain". My tier 2 page url is http://domain.com/building-paper-airplanes and the Title tag is "Building Paper Airplanes I Domain". The home page is more of an introduction page before the user gets into the instructions on the tier 2 page. Can anybody advise if this might be having a negative affect on my search results for the search term 'building paper airplanes'? Thanks.
Technical SEO | | Rogs.SEO0 -
Duplicate pages in Google index despite canonical tag and URL Parameter in GWMT
Good morning Moz... This is a weird one. It seems to be a "bug" with Google, honest... We migrated our site www.three-clearance.co.uk to a Drupal platform over the new year. The old site used URL-based tracking for heat map purposes, so for instance www.three-clearance.co.uk/apple-phones.html ..could be reached via www.three-clearance.co.uk/apple-phones.html?ref=menu or www.three-clearance.co.uk/apple-phones.html?ref=sidebar and so on. GWMT was told of the ref parameter and the canonical meta tag used to indicate our preference. As expected we encountered no duplicate content issues and everything was good. This is the chain of events: Site migrated to new platform following best practice, as far as I can attest to. Only known issue was that the verification for both google analytics (meta tag) and GWMT (HTML file) didn't transfer as expected so between relaunch on the 22nd Dec and the fix on 2nd Jan we have no GA data, and presumably there was a period where GWMT became unverified. URL structure and URIs were maintained 100% (which may be a problem, now) Yesterday I discovered 200-ish 'duplicate meta titles' and 'duplicate meta descriptions' in GWMT. Uh oh, thought I. Expand the report out and the duplicates are in fact ?ref= versions of the same root URL. Double uh oh, thought I. Run, not walk, to google and do some Fu: http://is.gd/yJ3U24 (9 versions of the same page, in the index, the only variation being the ?ref= URI) Checked BING and it has indexed each root URL once, as it should. Situation now: Site no longer uses ?ref= parameter, although of course there still exists some external backlinks that use it. This was intentional and happened when we migrated. I 'reset' the URL parameter in GWMT yesterday, given that there's no "delete" option. The "URLs monitored" count went from 900 to 0, but today is at over 1,000 (another wtf moment) I also resubmitted the XML sitemap and fetched 5 'hub' pages as Google, including the homepage and HTML site-map page. The ?ref= URls in the index have the disadvantage of actually working, given that we transferred the URL structure and of course the webserver just ignores the nonsense arguments and serves the page. So I assume Google assumes the pages still exist, and won't drop them from the index but will instead apply a dupe content penalty. Or maybe call us a spam farm. Who knows. Options that occurred to me (other than maybe making our canonical tags bold or locating a Google bug submission form 😄 ) include A) robots.txt-ing .?ref=. but to me this says "you can't see these pages", not "these pages don't exist", so isn't correct B) Hand-removing the URLs from the index through a page removal request per indexed URL C) Apply 301 to each indexed URL (hello BING dirty sitemap penalty) D) Post on SEOMoz because I genuinely can't understand this. Even if the gap in verification caused GWMT to forget that we had set ?ref= as a URL parameter, the parameter was no longer in use because the verification only went missing when we relaunched the site without this tracking. Google is seemingly 100% ignoring our canonical tags as well as the GWMT URL setting - I have no idea why and can't think of the best way to correct the situation. Do you? 🙂 Edited To Add: As of this morning the "edit/reset" buttons have disappeared from GWMT URL Parameters page, along with the option to add a new one. There's no messages explaining why and of course the Google help page doesn't mention disappearing buttons (it doesn't even explain what 'reset' does, or why there's no 'remove' option).
Technical SEO | | Tinhat0 -
Should I include A Description Tag for an article page
Hi, I am wondering if i should include a description tag for an article page where the news on that page will change around five times a day. I am not sure if to fill a description in for local news or to leave it blank so the search engines pick up the latest local news and then the next day they show the latest local news again in the search engines instead of having a static description of the news page. any help would be great
Technical SEO | | ClaireH-1848860 -
H-tags and Page Name best practice
For the past few months I've been working on a new site launch, but have been left with a couple of annoyances from my predecessor.. I've got a couple of questions about best practise (and if it's worth changing now). For reference, a good example page is http://polestars.net/hen-party/life-drawing-hen-party/ H-tags The (external) web designer has insisted that wrapping the logo in an H1 tag (with the same branded H1 text on every page), and using an H2 for the actual title of the page is fine. I really don't believe him, but at the same time, feel like maybe google is smart enough to discern the theme of a page in this structure. Is it worth having this changed so that the actual title is the first H1? Page naming convention Another annoyance that I've been left with is the fact that every product page is named the same "burlesque hen party", "life drawing hen party", "whatever hen party"... It looks a little weird, but my real concern is that as we now have 60 "hen party" links in the navigation menu of a bunch of our pages, this may be seen as keyword stuffing - is this a real concern, or am I overthinking it?
Technical SEO | | AlecPR0 -
Backtracking from verification meta tag to the correct Google account is difficult
A Google verification meta tag was created and implemented on a site that I am now responsible for (I took over an SEO project after a long lapse), but no one seems to know what Google account was used to create the meta tag in the first place. I'm finding it very difficult to backtrack from verification meta tag to the Google account, and all the online help is for those having trouble moving forward with the verification. Any suggestions or advice?
Technical SEO | | MaryDoherty0