Subdomains seen as one site
-
Earlier this month Google announced that sub-domains are now
be treated as one site. At first I thought this was good news as I like to use
sub-domains for separation of categories and the like. But what about links
from one sub-domain to the other, they uised to be external links now they are
internal links. If you don’t have many external links, I would say that the
cross sub-domain links would have been important, if you have a lot of external
links then the flow of link juice would be of more benefit. I think overall its
is a good thing.Does anyone have any opinions about this or know of any writings
on the subject since this announcement?http://googlewebmastercentral.blogspot.com/2011/08/reorganizing-internal-vs-external.html
-
Well you want to see them as they are considered for ranking. If for ranking a subdomain link is considered an internal link then yes, thats how WMT should report it.
-
I am not a legal expert but it is my understanding if you own the root domain, then you own the subdomains. Even if you use sites like wordpress.com or blogspot the legal owner of the subdomains is the root domain owner. If you violate their terms of service, the main company will kick the subdomain user off and take control over the account.
With respect to Google WMT, I think it's a convenience thing. In most cases, you want to know what other sites are linking to your site. It is usually not helpful to know your own subdomain is linking to your root domain. You expect that as you do other internal linking. It makes the analysis of external linking a bit easier to remove all internal links.
-
Exactly so I am thinking they will basically just aggregate any sites that use the same domain under the same webmaster account.
-
Yes i did not read too clearly, the comments explaned much more.
but it makes one think,about the old question about how google handles subdomains(for rankings). when does it see them as the same site and when does it see them as seperate.
I am guesisng they already do rank sites using the rules in article. I look after a site that has nothing more then a home, about, search and contact page on root domain, with several sub-domains. yet the home page ranks well for all the subdomains catgories. -
The announcement is informative but in my opinion highly misunderstood.
Dr. Pete asked the key question which was answered in a reply from the original author:
"This update only changes how links are displayed in Webmaster Tools. It doesn't affect how links are valued in relation to the search algorithm or ranking. It has nothing to do with Panda, nothing to do with keywords, nothing to do with PageRank."
In short, nothing has changed. The only difference iis how links are displayed when you are viewing the root domain in Google WMT.
-
I get your point, what they dont make clear is if you only list the root domain then how do they know you own the sub domains also, aer you to put the same verification code in each site?
And if they are going to see then as the same site are they going to sdee them that way for ranking? or do they already for that reason?
-
From reading that article it looks like they are doing this for their webmaster tools, but don't mention if they are combining subdomain links and primary domain links into their SERP calculations as if they were the same site.
Knowing if it affected SERP calculations is more important. Have you seen any articles on that (going to search shortly for that ;))
That would be more interesting to me as our site currently has at least 6 subdomains for publicly accessible sites, and while I would love to move them all under sub folders it requires a decent bit of effort on our side. If Google just counted them all under the same link profile for rankings then it would save time and money, and give Google a better view of our web presence as a whole. This would be better for the end user as the end user would be able to get to our blogs, forums, portal, and API knowledgebase easier from search results when just searching for our company name. So I can see it being useful for Google since it is useful for users.
For that article (webmaster tool purposes) I think it makes sense as it helps simplify reporting for the "www" subdomain. It makes it easier to aggregate your data, and see it how Google really sees it (especially if you have a rewrite from www.example.com to just example.com or vice versa). So it sounds good to me
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Subdomain initials vs full city name(s) for a multi city subdomain site?
Helping with a multi-city non-profit magazine/news blog. Subdomain options; sf.domain.com, ny.domain.com, la.domain.com sanfrancisco.domain.com, newyork.domain.com, ... Some cities added, will as an example seol.domain.com a city that doesnt have a recognizable initlals, like NYC for example. For brand, recognition, seo benefit, what have you used and why? Thanks
Industry News | | vmialik0 -
SEO Risks for redirecting sites
Hey Everyone, I've tried searching for this question, but am not exactly sure what keywords to search for so I'm probably missing the resources if they already exist... My client has had duplicated sites for years, and after multiple penalizations of those sites I was finally able to convince him to consolidate them into a "mega-site". Currently, he has a main domain, a geo-subdomain for each office location under the main domain, and a geo-domain for each office location. We plan on redirecting each geo-domain to the corresponding geo-subdomain. So, the final result will be one main domain, and a sub-domain for each office location. I'm looking for any information regarding tracking SEO data after the redirects are in place, how to guard against potential drops in SERPs, what's the smartest strategy to implement, etc... My client is very sensitive to his sites' SEO data, so if anyone has any SEO-related advice regarding redirecting sites it would be greatly appreciated! Thank you!
Industry News | | Level2Designs0 -
Any sites out there that are better than freelancer and more local based and not international
Hey guys i am looking for someone local or in the us to handle my social media clients. ie: posts, blogging, tweets . Any sites where i can posts jobs better than freelancer. Also anyone know of sites that you can find bloggers to blog for your clients sites. ( more content writers , to take directions ) Thanks guys BizDetox
Industry News | | BizDetox0 -
Chrome blocked sites used by Googles Panda update
Google's Panda update said it used Chrome users blocked sites lists as a benchmark for what they now term poor quality content. They said the Panda update effectively took about 85% of them out of the search results. This got me thinking, it would be very nice to discover what are the exact sites they don't like. Does anyone know if there is an archive of what these sites might be? Or if none exists, maybe if people could share their Chrome blocked sites on here we might get an idea?
Industry News | | SpecialCase0 -
Anyone seen any dramatic changes in SERPs 18th Feb?
Hi Guys, Wondering if there has been an algorithm update or new Panda released today? Not seen any news in Google yet however noticed fair shuffle of all results in my niche, particularly in the top 3. The main take I've had from the brief research is that the brands in the niche have popped up a few spots and blogs / affiliate sites etc have dropped. Results noticed using Google.co.uk
Industry News | | shloy23-2945840 -
Alexa Site Audit
Hello All, I run an Alexa Site Audit report for my client. Is it normal that my Alexa site audit report will be public i.e. S3 URL for downloading the PDF file was open to anyone knowing the URL. No re-direct to login etc was there? Yes the S3 URL is funny but Google bots have ways to find this funny URLs so is this normal? If you have tried Alexa can you repeat the problem i.e. typing your S3 report URL? can you get it to? or is it just me. I have asked Alexa about it. I think there should be some security. Thanks.
Industry News | | SEO_Z0 -
SEO sites were blasted by Panda
Just noticed that lots of SEO sites were blasted by Panda... http://trends.google.com/websites?q=webmasterworld.com&geo=all&date=2011&sort=0 http://trends.google.com/websites?q=seomoz.org&geo=all&date=2011&sort=0 http://trends.google.com/websites?q=digitalpoint.com&geo=all&date=2011&sort=0 http://trends.google.com/websites?q=forums.seochat.com&geo=all&date=2011&sort=0
Industry News | | EGOL3 -
Schedule for Panda / Farmer Update rollout for non English sites?
Hey Mozzers, I just read about the global rollout of googles Panda/Farmer Update for English queries, e.g. at searchengineland. Has anyone read something about when google is planning to roll out the update for non-English speaking countries, i.e. queries in other languages? I googled a bit but couldn't find anything, not even speculations. Cheers, Frank
Industry News | | FranktheTank-474970