Blocking Subdomain from Google Crawl and Index
-
Hey everybody, how is it going?
I have a simple question, that i need answered.
I have a main domain, lets call it domain.com. Recently our company will launch a series of promotions for which we will use cname subdomains, i.e try.domain.com, or buy.domain.com. They will serve a commercial objective, nothing more.
What is the best way to block such domains from being indexed in Google, also from counting as a subdomain from the domain.com. Robots.txt, No-follow, etc?
Hope to hear from you,
Best Regards,
-
Hello George, Thank you for fast answer! I read that article and there is some issue with that. if you can see at it, i'd really appreciate it. So the problem is that if i do it directly from Tumblr, it will also block it from Tumblr users. Here is the note right below that option "Allow this blog to appear in search results":
"This applies to searches on Tumblr as well as external search engines, like Google or Yahoo."Also, if i do it from GWT, i'm very concerned to remove URLs with my subdomain because i afraid it will remove all my domain. For example, my domain is abc.com and the Tumblr blog is setup on tumblr.abc.com. So i afraid if i remove tumblr.abc.com from index, it will also remove my abc.com. Please let me know what you think.
Thank you!
-
Hi Marina,
If I understand your question correctly, you just don't want your Tumblr blog to be indexed by Google. In which case these steps will help: http://yourbusiness.azcentral.com/keep-tumblr-off-google-3061.html
Regards,
George
-
Hi guys, I read your conversation. I have similar issue but my situation is slightly different. I'll really appreciate if you can help with this. So i have also a subdomain that i don't want to be indexed by Google. However, that subdomain is not in my control. I mean, i created subdomain on my hosting but it is pointing to my Tumblr blog. So i don't have access to its robot txt. So can anybody advise what can i do in this situation to noindex that subdomain?
Thanks
-
Personally I wouldn't rely just on robots.txt, as one accidental, public link to any of the pages (easier than you may think!) will result in Google indexing that subdomain page (it just won't be followed). This means that the page can get "stuck" in Google's index and to resolve it you would need to remove it using WMT (instructions here). If there were a lot of pages accidentally indexed, you would need to remove the robots.txt restriction so Google can crawl it, and put a noindex/nofollow tags on the page so Google drops it from its index.
To cut a long story short, I would do both Steps 1 and 2 outlined by Federico if you want to sleep easy at night :).
George
-
It would also be smart to add the subdomains in Webmaster Tools in case one does get indexed and you need to remove it.
-
Robots.txt is easiest and quickest way. As a back up you can use the Noindex meta tag on the pages in the subdomain
-
2 ways to do it with different effects:
-
Robots.txt in each subdomain. This will entirely block any search engine to even access those pages, so they won't know what they have inside.
User-Agent:*
Disallow: /
-
noindex tags in those pages. This method allows crawlers to read the page and maybe index (if you set a "follow") the pages to which you link to.or "nofollow" if you don't want the linked pages to be indexed either.
Hope that helps!
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Pages Not Showing Up In Search After Being Indexed....
Hello, I'm trying optimize my google search for local business in Fort Myers, FL area. I've created specific keywords that people are searching and have made my page title and on-page experience reflect this relevency. However, the page isn't appearing anywhere on google search even though search console has stated that the page has been indexed. My question is, how do I trouble shoot this so that I can appear in local search for these search terms?
On-Page Optimization | | scottgray06200 -
Does the content in Joomla modules get indexed
Hello Moz, we are using Joomla to build websites. In Joomla you can put content in modules. What we would like to know if would search engines index content in modules, as it does with content attached to menu links? Thanks Ian
On-Page Optimization | | Substance-create0 -
Index SEO Performance
Hi guys, I was wondering if there is a difference in SEO performance between a page which ends with .html or just with a slash. For example: www.domain.com/test/ - www.domain.com/test.html Which one is better? And is there a difference between productpages and productcategory pages. Because we see mostly productpages ending with .html and category pages ending with a slash. Take a look at some big dutch companies like: http://www.coolblue.nl/
On-Page Optimization | | Happy-SEO
http://www.bol.com/
https://www.zalando.nl/ (product page) There are doing this with a reason.... i guess.... Thanks.1 -
Google crawler showing cache of another page
For the page http://www.thinkdigit.com/top-products/Laptops-and-PCs/top-10-laptops-124.php google is showing another page in cache (http://www.thinkdigit.com/top-products/Ultrabooks/top-10-ultrabooks-153.php). Please let me know how this happened and how to correct it.
On-Page Optimization | | 9dot90 -
Google rel hell
So apologies in advance for this question, but: Can someone explain whether as a site we should be using the "rel author" tag or the "rel publisher" tag? 1. We don't really need to distinguish between the people who write our content. 2. We definitely do need to establish ownership of our content, as unfortunately it has been widely copied. We are spending quite a bit of time filing DMCA notices. 3. Do we need to apply either tag to every page? Or does "del publisher" just need to be applied to the homepage to cover the rest of the site? 4. What looks better in the search results? - a person's face or a company logo? If prefer a face, but understand we need to promote our brand. Thanks P
On-Page Optimization | | dexm100 -
Which is Best Practice for creating URLs for subdomain?
My website is related to education. We have created sub domains for all major colleges, universities & Entrance exams like Gre, Toefl ETC. for eg: amityuniversity.abc.com (Amity is Name of University ) Now if have to mention city name in URL as well (college is located in multiple locations) amityuniversity-delhi.abc.com
On-Page Optimization | | rohanarora536
amityuniversitydelhi.abc.com Now my Q is can we use hyphens in sub domains if we have to add city name or shall we create without using any hyphens. In Directory structure we can always separate words with hyphens, can we follow same practice in subdomain as well Which is a best URL for subdomain amity-university-delhi.abc.com
amityuniversity-delhi.abc.com
or amityuniversitydelhi.abc.com0 -
Should I Remove This Subdirectory From Google?
On my site, I have a subdirectory. It posts articles from a bunch of websites that my readers are interested in & links back to all of those sites. There is no original content in it. There are over 1700 indexed pages in this subdirectory. The rest of my site has about 500 (all original content). The search engine traffic for this subdirectory only accounts for 3.9% of my sites overall visits. Should I consider removing this subdirectory? Could all the duplicate content be hurting the rankings of my legit pages? What do you all think?
On-Page Optimization | | PedroAndJobu0 -
Does Google respect User-agent rules in robots.txt?
We want to use an inline linking tool (LinkSmart) to cross link between a few key content types on our online news site. LinkSmart uses a bot to establish the linking. The issue: There are millions of pages on our site that we don't want LinkSmart to spider and process for cross linking. LinkSmart suggested setting a noindex tag on the pages we don't want them to process, and that we target the rule to their specific user agent. I have concerns. We don't want to inadvertently block search engine access to those millions of pages. I've seen googlebot ignore nofollow rules set at the page level. Does it ever arbitrarily obey rules that it's been directed to ignore? Can you quantify the level of risk in setting user-agent-specific nofollow tags on pages we want search engines to crawl, but that we want LinkSmart to ignore?
On-Page Optimization | | lzhao0