How do we handle sitemaps in robots.txt when multiple domains point to same physical location?
-
we have www.mysite.net, www.mysite.se, www.mysite.fi and so on. all of these domains point to the same physical location on our webserver, and we replace texts given back to client depending on which domain he/she requested.
My problem is this: How do i configure sitemaps in robots.txt when robots.txt is used by multiple domains? If I for instance put the rows
Sitemap: http://www.mysite.net/sitemapNet.xml
Sitemap: http://www.mysite.net/sitemapSe.xmlin robots.txt, would that result in some cross submission error?
-
Thanks for your help René!
-
yup
-
Yes, I mean GTW of course :).
A folder for each site would definitely make some things easier, but it would also mean more work every time we need to republish the site or make configurations.
Did I understand that googlelink correctly in that if we have verified ownership in GWT for all involved domains cross-site submission in robots.txt was okay? I guess google will think its okay anyway.
-
actually google has the answer, right here: http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=75712
I always try to do what google recommends even though something might work just as well.. just to be on the safe side
-
you can't submit a sitemap in GA so I'm guessing you mean GWT
Whether or not you put it in the robots.txt shouldn't be a problem. since in each sitemap, the urls would look something like this:
Sitemap 1:<url><loc>http:/yoursite.coim/somepage.html</loc></url>
Sitemap 2:<url><loc>http:/yoursite.dk/somepage.html</loc></url>
I see no need to filter what sitemap is shown to the crawler. If your .htaccess is set-up to redirect traffic from the TLD (top level domain eg .dk .com ex.) to the correct pages. Then the sitemaps shouldn't be a problem.
The best solution would be: to have a web in web. (a folder for each site on the server) and then have the htaccess redirect to the right folder. in this folder you have a robots.txt and a sitemap for that specific site. that way all your problems will be gone in a jiffy. It will be just like managing different 3 sites. even though it isn't.
I am no ninja with .htaccess files but I understand the technology behind it and know what you can do in them. for a how to do it guide, ask google thats what I allways do when I need to goof around in the htaccess. I hope it made sense.
-
Thanks for your response René!
Thing is we already submit the sitemaps in google analytics, but this SEO company we hired wants us to put the sitemaps in robots.txt as well.
The .htaccess idea sounds good, as long as google or someone else dont think we are doing some cross-site submission error (as described here http://www.sitemaps.org/protocol.php#submit_robots)
-
I see no need to use robots.txt for that. use Google and Bings webmaster tools. Here you have each domain registered and can submit sitemaps to them for each domain.
If you want to make sure that your sitemaps are not crawled by a bot for a wrong language. I would set it up in the .htaccess to test for the entrance domain and make sure to redirect to the right file. Any bot will enter a site just like a browser so it needs to obey the server. so if the server tells it to go somewhere it will.
the robots.txt can't by it self, do what you want. The server can however. But in my opinion using bing and google webmaster tools should do the trick.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best Way to Handle Multi-Language Sites
In the last year we've made a few significant changes to the structure of our site - namely adding translations for a few languages. We have historically been gaining in organic search by about 10% each month, but in the last two months we've leveled out and seen a slight dip. I am wondering if this has something to do with the addition of the second language, and namely if there's a chance we've been penalized due to duplicate content. We have almost all pages / content on the site translated by a translator, but the way the development works the site will grab the english version if a translation hasn't been added - potentially adding some duplicate content? The URL structure remains the same, other than the addiion of the language - site.com/our-tour vs site.com/de/our-tour We also haven't translated the tour name itself, so that remains the same. Just wondering if anyone has any feedback on best practices here or things I should be looking out for. Thanks in Advance.
On-Page Optimization | | mkgreyound1 -
How to handle outdated, unoptimized blog posts receiving little or zero traffic?
I'm doing some overdue spring cleaning on our WP blog. Some big visual updates are in the works, but currently I'm working on pruning and updating some poopy and outdated content. Many of the older posts weren't written with SEO in mind and were posted to the blog merely as an extension of our monthly enewsletter. Here's an example: http://pq.systems/2FkQyVG This post needs a lot of work to meet our new standards. The content is thin, readability is weak, kw targetting is non-existant, the visuals suck, zero links, and the charting software mentioned has since been replaced with another solution that we are currently promoting. There are quite a few other posts with similar issues... Any thoughts on the best way to handle these posts? From poking around similar Q&A threads, it seems my options are: Create new updated post, remove old post, 301 redirect from old to new Create new updated post, add blurb & link pointing to new post at the top of old post Edit/update old post, add "This post was updated....etc" blurb to top of old post Any other options or opinions on which solution I should go with would be much appreciated!
On-Page Optimization | | SamKlep0 -
Domain Authority
I made some changes to my site and many of my keywords improved dramatically. Its odd however, that the changes caused the domain authority to go down by 3 points, from 20 to 17. Is this a short term thing caused by recent changes and likely to recover?
On-Page Optimization | | KrisIrr0 -
What's the best SEO tactics when you have a dedicated web address pointing to a page on a different site?
Hope someone can help with a question I've got about sorting out some duplicate content issues. To simplify the question, imagine there is a website a.com which has a page a.com/newslettersignup. In addition to the a.com domain, there is also a different web address, ashortcut.com, which points to a.com/newslettersignup. ashortcut.com is the web address that is advertised in marketing material etc. So what is the best way then to tell Google etc. that ashortcut.com is the preferred URL for the page which sits at a.com/newslettersignup? The advice I've read about the canonical tag, for example, doesn't cover this exact scenario so although it can support cross-domain information, I'm not sure if that's the best route to follow. Thanks!
On-Page Optimization | | Nobody15755058948220 -
SVG image files causing multiple title tags on page - SEO issue?
Does anyone have any experience with SVG image files and on-page SEO? A client is using them and it seems they use the title tag in the same way a regular image (JPG/PNG) would use an image ALT tag. I'm concerned that search engines will see the multiple title tags on the page and that this will cause SEO issues. Regular crawlers like Moz flag it as a second title tag, however it's outside the header and in a SVG wrap so the crawlers really should understand that this is a SVG title rather than a second page title. But is this the case? If anyone has experience with this, I'd love to hear about it.
On-Page Optimization | | mrdavidingram2 -
Sitemaps Updating
Im using wordpress and I realise that my sitemaps doesnt update itself when i add an additional page on my website, like a blog post. I have to go to (1) setting > xml sitemap setting > click on build sitemap > save changes in wordpress, and then (2) Export the sitemal.xml file it to webmaster tools in google every single time i blog. Am i doing it wrong? i feel that all these should be automatic.
On-Page Optimization | | kevinbp0 -
URL Domain Used in Meta Description
Today I was asked if using a domain url in your meta description can have a negative impact on your website. This description includes a list of the homepage url, but directs visitors to a different internal page of the website. My concern fell with directing visitors to a different page of the site, but promoting the homepage in the description/snippet. With Penguin 2.1 release on the 4th, I'm very cautious of my links/urls. What are your thoughts behind this? What are the possible, if any negative impacts this could have on a site? This URL does have a brand name as so the Title.
On-Page Optimization | | flcity150 -
How to handle lots of outbound links
I decided to create a page on my website where I would list all of my favorite resources and 3rd party tools. There are now 35 links in the main content section of the page, all with anchor text, pointing to websites in my industry. My question is this: what is the best practice here? Should I add nofollow tags to the links? Should I do something else to indicate that these links shouldn't be crawled? Frankly, I don't mind passing some link juice to these tools (in this case, and from this particular website), but I might make a different decision with a client's website. Does anyone have any thoughts on this? Here is the page I'm referring to, in case anyone wants to look: http://willmarlow.com/resources-2/the-digital-marketing-toolbox/. Thanks!
On-Page Optimization | | williammarlow0