Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Staging & Development areas should be not indexable (i.e. no followed/no index in meta robots etc)
-
Hi
I take it if theres a staging or development area on a subdomain for a site, who's content is hence usually duplicate then this should not be indexable i.e. (no-indexed & nofollowed in metarobots) ? In order to prevent dupe content probs as well as non project related people seeing work in progress or finding accidentally in search engine listings ?
Also if theres no such info in meta robots is there any other way it may have been made non-indexable, or at least dupe content prob removed by canonicalising the page to the equivalent page on the live site ?
In the case in question i am finding it listed in serps when i search for the staging/dev area url, so i presume this needs urgent attention ?
Cheers
Dan
-
- use robots.txt vs the meta tags - robots.txt is preferred.
-
I'm about to issue these instructions would appreciate it if you could quickly confirm covers your advice correctly and nothing missing:
1) Setup a completely different GWT account unrelated to the main site, so that there is a new GWT account specific to the staging subdomain
2) Add a robots.txt on the staging area subdomain site that disallows all pages and all crawlers OR use the noindex meta tag on all pages. Its obviously very important when you update the main site it DOES NOTinclude or push out these files too (since that would result in main site or pages being de-indexed)3) Request removal of all pages in GWT. Leave the form blank for the page to be removed since this will remove the entire site4) After about 1 month (or you see that the pages are all out of the serps), and google has spidered and seen the robots.txt, then put up a password on the entire staging site.Note:For brand new sites staging areas that don't yet exist or exist but are new and not yet showing up in the index then simply add a password for human access to prevent the above process being required in the future. -
Thanks for clarifying that CleverPHD & thanks again for all your help and great advice
Have a great weekend !!
All Best
Dan
-
That is a completely valid question. This is why setting up the separate GWT account for the dev.domain.ext vs www.domain.ext. When you submit the removal request it will only be in the dev.domain.ext account.
The only thing you want to watch is that if you setup robots.txt in your dev environment you want to make sure that it does not get pushed out to your production server. That is the only gotcha as I see it.
-
thanks !
as er my last question theres no risk of accidentally taking out the main site as part of this process ?
cheers
dan
-
Thanks so much for that great advice
just a bit worried about accidentally getting main site removed by accident, i take it so long as its a brand new GWT account for that specific subdomain then this cant happen ?
Cheers
Dan
-
Here is a Google documentation on how to use the GWT to remove a page/directory/site and then the interaction with robots.txt
http://googlewebmastercentral.blogspot.com/2010/03/url-removal-explained-part-i-urls.html
"In order for a directory or site-wide removal to be successful, the directory or site must be disallowed in the site's robots.txt file."
Side story. I once had a subdomain that I needed to take out, but I could not modify the robots.txt file properly (long story). So, we used the GWT tool and the meta noindex tag. It still worked, but I think that would only be a backup approach to the one suggested by the documentation.
-
Usually, this would be true that you would need to use the noindex tag to get things out of the SERPs and need to leave the robots.txt "open" to the crawlers. But when you are working with the remove URL tool in GWT,they rx that you then put the site in robots.txt to keep them out of it
The removal tool in GWT takes care of Google taking the URLs out and then the robots.txt keeps the bots from coming back. Just a different sequence than if you were to use the noindex meta.
-
If you create the GWT account for the dev site and you submit for removal, GWT requires that you either a) have the site blocked in robots.tx or have a noindex meta tag on the pages. Otherwise they will just crawl you again later and you are back in the index. See my post from earlier.
-
Short answer - no dev sites should be public to start with to anyone (let along Google et alia). The simplest way is to put an htacess password on all your dev sites. You can do a password per person in your company, or just one general one that everyone on the dev team shares.
If you do have a dev site in the Serps, the simplest way to get it out is to setup a GWT account for that subdomain and then e.g. dev.yourdomain.ext and then go into that account and request removal of all pages. You just leave the form blank for the page to be removed and it takes out the whole site. You then need a robots.txt on dev.yourdomain.ext (different from the www. version) that disallows all pages all crawlers - that or use the noindex meta tag on all page.
After about 1 month (or you see that the pages are all out of the serps), then I would put up a password on that entire site and be done with it. Key point, dont put the password up until you let google try to spider and it sees the robots etc.
Also, if you have any other staging sites that are out there like test.yourdomain.ext etc. If they are not indexed, go ahead and put the password up on them to limit your exposure.
Public dev sites are the fastest way to get duplicate content into the index and to jack with the ranking of your current site. It is key that all of them are locked down. If one of your developers say it is no big deal, call BS, it is a big deal and it can cause a big mess.
-
Hey Dan,
In this case, I would not exclude crawling via robots.txt. Perhaps later after you have verified the URLs are out of the index.
Just because Google can't crawl a page, doesn't mean they won't keep it in the index. Excluding crawling will not get a page out of the index.
Add the NOINDEX, FOLLOW tag you listed above and give it some time.
Use GWT if it's urgent or the information is sensitive.
-
Thanks Anthony,
The staging area already exists and is indexable as far as i can tell
So i need to tell developers to exclude crawling via robots.txt, add a no-index tag to head of each page but keep it followed so still crawlable i.e. within the Head section of every page on the dev area
OR alternatively just remove urls from GWT)
If excluding crawling via robots.txt file then why do you need to add a noindex tag to each page too, surely the robots.txt deals with this situation ?
cheers
dan
-
Ideally when creating a new staging area, you'd want to exclude crawling via robots.txt.
Add the NoIndex tag to the head of your pages to get them removed from the SERPs. Make sure the page is still crawlable though, as if you exclude it in robots.txt first and then NoIndex it, Google won't be able to see the new NoIndex tag.
If there are not a lot of pages to remove, you can request page removal within Google Webmaster Tools.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Removing a site from Google index with no index met tags
Hi there! I wanted to remove a duplicated site from the google index. I've read that you can do this by removing the URL from Google Search console and, although I can't find it in Google Search console, Google keeps on showing the site on SERPs. So I wanted to add a "no index" meta tag to the code of the site however I've only found out how to do this for individual pages, can you do the same for a entire site? How can I do it? Thank you for your help in advance! L
Technical SEO | | Chris_Wright1 -
My Homepage Won't Load if Javascript is Disabled. Is this an SEO/Indexation issue?
Hi everyone, I'm working with a client who recently had their site redesigned. I'm just going through to do an initial audit to make sure everything looks good. Part of my initial indexation audit goes through questions about how the site functions when you disable, javascript, cookies, and/or css. I use the Web Developer extension for Chrome to do this. I know, more recently, people have said that content loaded by Javascript will be indexed. I just want to make sure it's not hurting my clients SEO. http://americasinstantsigns.com/ Is it as simple as looking at Google's Cached URL? The URL is definitely being indexed and when looking at the text-only version everything appears to be in order. This may be an outdated question, but I just want to be sure! Thank you so much!
Technical SEO | | ccox10 -
No Index PDFs
Our products have about 4 PDFs a piece, which really inflates our indexed pages. I was wondering if I could add REL=No Index to the PDF's URL? All of the files are on a file server, so they are embedded with links on our product pages. I know I could add a No Follow attribute, but I was wondering if any one knew if the No Index would work the same or if that is even possible. Thanks!
Technical SEO | | MonicaOConnor0 -
Can you have a /sitemap.xml and /sitemap.html on the same site?
Thanks in advance for any responses; we really appreciate the expertise of the SEOmoz community! My question: Since the file extensions are different, can a site have both a /sitemap.xml and /sitemap.html both siting at the root domain? For example, we've already put the html sitemap in place here: https://www.pioneermilitaryloans.com/sitemap Now, we're considering adding an XML sitemap. I know standard practice is to load it at the root (www.example.com/sitemap.xml), but am wondering if this will cause conflicts. I've been unable to find this topic addressed anywhere, or any real-life examples of sites currently doing this. What do you think?
Technical SEO | | PioneerServices0 -
No indexing url including query string with Robots txt
Dear all, how can I block url/pages with query strings like page.html?dir=asc&order=name with robots txt? Thanks!
Technical SEO | | HMK-NL0 -
OK to block /js/ folder using robots.txt?
I know Matt Cutts suggestions we allow bots to crawl css and javascript folders (http://www.youtube.com/watch?v=PNEipHjsEPU) But what if you have lots and lots of JS and you dont want to waste precious crawl resources? Also, as we update and improve the javascript on our site, we iterate the version number ?v=1.1... 1.2... 1.3... etc. And the legacy versions show up in Google Webmaster Tools as 404s. For example: http://www.discoverafrica.com/js/global_functions.js?v=1.1
Technical SEO | | AndreVanKets
http://www.discoverafrica.com/js/jquery.cookie.js?v=1.1
http://www.discoverafrica.com/js/global.js?v=1.2
http://www.discoverafrica.com/js/jquery.validate.min.js?v=1.1
http://www.discoverafrica.com/js/json2.js?v=1.1 Wouldn't it just be easier to prevent Googlebot from crawling the js folder altogether? Isn't that what robots.txt was made for? Just to be clear - we are NOT doing any sneaky redirects or other dodgy javascript hacks. We're just trying to power our content and UX elegantly with javascript. What do you guys say: Obey Matt? Or run the javascript gauntlet?0 -
Invisible robots.txt?
So here's a weird one... Client comes to me for some simple changes, turns out there are some major issues with the site, one of which is that none of the correct content pages are showing up in Google, just ancillary (outdated) ones. Looks like an issue because even the main homepage isn't showing up with a "site:domain.com" So, I add to Webmaster Tools and, after an hour or so, I get the red bar of doom, "robots.txt is blocking important pages." I check it out in Webmasters and, sure enough, it's a "User agent: * Disallow /" ACK! But wait... there's no robots.txt to be found on the server. I can go to domain.com/robots.txt and see it but nothing via FTP. I upload a new one and, thankfully, that is now showing but I've never seen that before. Question is: can a robots.txt file be stored in a way that can't be seen? Thanks!
Technical SEO | | joshcanhelp0 -
/index.php in sitemap? take it out?
Hi Everyone, The following was automatically generated at xml-sitemaps.com Should I get rid of the index.php url from my sitemap? If so, how do I go about redirecting it in my htaccess ? <url><loc>http://www.mydomain.ca/</loc></url>
Technical SEO | | RogersSEO
<url><loc>http://www.mydomain.ca/index.php</loc></url> thank you in advance, Martin0