How do we handle sitemaps in robots.txt when multiple domains point to same physical location?
-
we have www.mysite.net, www.mysite.se, www.mysite.fi and so on. all of these domains point to the same physical location on our webserver, and we replace texts given back to client depending on which domain he/she requested.
My problem is this: How do i configure sitemaps in robots.txt when robots.txt is used by multiple domains? If I for instance put the rows
Sitemap: http://www.mysite.net/sitemapNet.xml
Sitemap: http://www.mysite.net/sitemapSe.xmlin robots.txt, would that result in some cross submission error?
-
Thanks for your help René!
-
yup
-
Yes, I mean GTW of course :).
A folder for each site would definitely make some things easier, but it would also mean more work every time we need to republish the site or make configurations.
Did I understand that googlelink correctly in that if we have verified ownership in GWT for all involved domains cross-site submission in robots.txt was okay? I guess google will think its okay anyway.
-
actually google has the answer, right here: http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=75712
I always try to do what google recommends even though something might work just as well.. just to be on the safe side
-
you can't submit a sitemap in GA so I'm guessing you mean GWT
Whether or not you put it in the robots.txt shouldn't be a problem. since in each sitemap, the urls would look something like this:
Sitemap 1:<url><loc>http:/yoursite.coim/somepage.html</loc></url>
Sitemap 2:<url><loc>http:/yoursite.dk/somepage.html</loc></url>
I see no need to filter what sitemap is shown to the crawler. If your .htaccess is set-up to redirect traffic from the TLD (top level domain eg .dk .com ex.) to the correct pages. Then the sitemaps shouldn't be a problem.
The best solution would be: to have a web in web. (a folder for each site on the server) and then have the htaccess redirect to the right folder. in this folder you have a robots.txt and a sitemap for that specific site. that way all your problems will be gone in a jiffy. It will be just like managing different 3 sites. even though it isn't.
I am no ninja with .htaccess files but I understand the technology behind it and know what you can do in them. for a how to do it guide, ask google thats what I allways do when I need to goof around in the htaccess. I hope it made sense.
-
Thanks for your response René!
Thing is we already submit the sitemaps in google analytics, but this SEO company we hired wants us to put the sitemaps in robots.txt as well.
The .htaccess idea sounds good, as long as google or someone else dont think we are doing some cross-site submission error (as described here http://www.sitemaps.org/protocol.php#submit_robots)
-
I see no need to use robots.txt for that. use Google and Bings webmaster tools. Here you have each domain registered and can submit sitemaps to them for each domain.
If you want to make sure that your sitemaps are not crawled by a bot for a wrong language. I would set it up in the .htaccess to test for the entrance domain and make sure to redirect to the right file. Any bot will enter a site just like a browser so it needs to obey the server. so if the server tells it to go somewhere it will.
the robots.txt can't by it self, do what you want. The server can however. But in my opinion using bing and google webmaster tools should do the trick.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How to best handle search landing pages - that don't exist
I have quite a bit of blog information that can be searched, which results in "pages" that don't actually live anywhere. These are scanned by Moz and appear as poor page quality for speed, etc. How do I get the service to either ignore all of these or is there a way to treat them as a real page with content? As there are quite a few generated over time, I'd like to be able to capture them somehow. Thanks.
On-Page Optimization | | amac70 -
Creating Tables with Multiple Links
Hello, Is Creating Tables with Multiple Links and using schema markup bad SEO? For example I have a real estate website which I have a hierarchy of home page -> County pages -> city pages-> blog posts.. However at the bottom of the county pages I list the city pages i serve under a table with links..
On-Page Optimization | | sqc0 -
Canonicalising a product with multiple variants
I am working with an ecommerce site and have encountered an issue I haven't come across before and would appreciate some advice on how to proceed. There are multiple variation products with one master product and then up to 20 or 30 variant products, the variation could be colour, size or both. The site has been set up to canonicalise all the variations to the master variant product, which I understand to be best practice. But, this is where the issue occurs, the master variant product URL 302 redirects to one of the variant product URLs. Example below. My question is, is this harmful to our SEO efforts? Would be be best to canonicalise to a preferred colour or size variation? EXAMPLE: Master variant product: www.example.co.uk/primary-category/product-123 Seeing this product on the page and clicking will 302 redirect to www.example/co.uk/primiary-category/product-123/colour-456 On page www.example/co.uk/primiary-category/product-123/colour-456 the canonical tag is www.example.co.uk/primary-category/product-123 Any help with this would be greatly appreciated.
On-Page Optimization | | SimonKenworthy0 -
Ranking over someone else who has your branded domain name
Hello! I have a client who has been in business for a long time, but was very late to the game online. As a result, the branded domain for his business (for explanation purposes I'll call it "Houston Tan"*) was already taken when he decided he needed a website, however it was not being used. He approached the company that owned "houstontan.com" and they refused to sell it to him. Not only that, they turned around and opened a company and developed the website using his trademarked company name as one word instead of two, "HoustonTan." It was brought to court and the judge determined that since they created a new name by combining the two words, there was nothing that he could do. Still having to create a website for his company, he chose the domain "HoustonSunTan.com." Not sure who was advising him on that one. So now he has a domain name with only a partial match to his company name. As you would imagine, when you search Houston Tan, HoustonTan.com is number 1, 2 & 3, and HoustonSunTan.com is #4. My question is, do you think it is even possible for his company to overtake the top spot of Google? Or have you ever come across a situation like this and if so what worked for you? FYI: Even though it says Houston, the company is a global company in over 500 cities (with one 800 number unfortunately), so local SEO strategies wouldn't necessarily apply. *Names are made up to protect both parties 🙂
On-Page Optimization | | American.Made0 -
Rich snippets for multiple breadcrumbs
Hey there,
On-Page Optimization | | Supertramp
I have been playing around with rich snippets for multiple breadcrumbs for quite a bit now - without any success. It would be great if someone could point me to an example where this has been implemented correctly. I followed the Google recommendation but it doesn't seem to work for me. See also: http://searchenginewatch.com/article/2335753/Matt-Cutts-If-You-Have-Multiple-Breadcrumbs-Google-Picks-the-First-One Thanks for your help. Cheers,
Jochen0 -
How to Structure URL's for Multiple Locations
We are currently undergoing a site redesign and are trying to figure out the best way to structure the URL's and breadcrumbs for our many locations. We currently have 60 locations nationwide and our URL structure is as follows: www.mydomain.com/locations/{location} Where {location} is the specific street the location is on or the neighborhood the location is in. (i.e. www.mydomain.com/locations/waterford-lakes) The issue is, {location} is usually too specific and is not a broad enough keyword. The location "Waterford-Lakes" is in Orlando and "Orlando" is the important keyword, not " Waterford Lakes". To address this, we want to introduce state and city pages. Each state and city page would link to each location within that state or city (i.e. an Orlando page with links to "Waterford Lakes", "Lake Nona", "South Orlando", etc.). The question is how to structure this. Option 1 Use the our existing URL and breadcrumb structure (www.mydomain.com/locations/{location}) and add state and city pages outside the URL path: www.mydomain.com/{area} www.mydomain.com/{state} Option 2 Build the city and state pages into the URL and breadcrumb path: www.mydomain.com/locations/{state}/{area}/{location} (i.e www.mydomain.com/locations/fl/orlando/waterford-lakes) Any insight is much appreciated. Thanks!
On-Page Optimization | | uBreakiFix0 -
Using Robots Meta Tag on Review Form Pages
I have gone over this so many times and I just can't seem to get it straight and hope someone can help me out with a couple of questions: Right now, on my dynamically created pages created by filters (located on the category pages) I am using rel""canonical" to point them to their respective category page. Should I also use the robots meta tag as well? Similarly, each product I have on my site has a review form on it and thus is getting indexed by Google. I have placed the same canonical tag on them as well pointing them to the page with the review form on it. In the past I used robots.txt to block google from the review pages but this didn't really do much. Should I be using the robots meta tag on these pages as well? If I used the robots meta tag should I noindex,nofollow? Thanks in advance, Jake
On-Page Optimization | | jake3720 -
Sitemap for webshop?
Had a client today who contacted me regarding having no sitemap on his webshop (with about 2000 products). He told me that his design company, who had made his webshop, adviced him against putting up a site since it would be giving google to much information at once, hence making it stop visiting his webshop as much? Well my first through was -okay this must be some kind of joke.
On-Page Optimization | | zarkas
But with seo, there are always something, now and then, that still surprises you, So are there anyone else here, who could share some info on this?0