Do you get credit for an external link that points to a page that's being blocked by robots.txt
-
Hi folks,
No one, including me seems to actually know what happens!?
To repeat:
If site A links to /home.html on site B and site B blocks /home.html in Robots.txt, does site B get credit for that link?
Does the link pass PageRank? Will Google still crawl through it? Does the domain get some juice, but not the page?
I know there's other ways of doing this properly, but it is interesting no?
-
Hi Dave,
I believe there would be two answers to your question
1. If Googlebot finds the page via the external link, then YES:
- the link will pass PageRank
- Googlebot will crawl
- both the page and the domain will get juice, because Googlebot hasn't seen the robots.txt
2. If Googlebot comes to the site via the root (assuming that it obeys the command to block), then NO:
- None of the above would happen because the page would never be seen by Googlebot, so the incoming link would never be seen.
If, on the other hand, Googlebot comes to the page via the root and ignores the command to block, then it should be reasonable to assume that means the page would be crawled & links attributed as though there were no robots.txt, but that is only an assumption, so I guess your question would remain open.
Don't suppose that helped much
Sha
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Site Migration between CMS's
Hi There, I have a technical question about migrating CMS's but not servers. My client has site A on Joomla install, He want's ot migrate to Wordpress and we will call this site B. As he has a lot of old content on site A he doesn't want to lose, he has put site B (wordpress install) on a subdirectory site.com/siteb (for example). and will use a htaccess to forward the root domain to this wordpress site. Therefore anyone going to www.site.com will see the new wordpress site and the old content and joomla install will sit on the root of the server. Will Google have an issue with this? Will it even find the old content? what are the issues for the new site and new content? Look forward getting your guys input
Technical SEO | | nezona1 -
Solutions for too many on-page links?
We have just begun using SEO Moz a few months ago and have been busy cleaning up some of our warnings and errors. One of the errors that has been an issue is ... too many on-page links. I am trying to correct this issue and I am wondering how seo moz counts these links. For instance... we have links to many of our product categories in a drop down from our main menu, those same links are listed in our footer. Does this get counted as two or only one link. If two, should we make one of the link no follow or how would you best suggest correcting this. Our website is www.unikeyhealth.com Since the menu and the footer appear on virtually every page on our site correcting this issue will quickly sort out this problem. Thanks for any advice.
Technical SEO | | unikey0 -
Unfindable 404's
So I have noticed that my site has some really strange 404's that are only being linked to from internal links from the site.
Technical SEO | | Adamshowbiz
When I go to the pages that Web master tools suggests I can't actaully find the link which is pointing to the 404. In that instance what do you do? Any help would be much appreciated 🙂0 -
Allow or Disallow First in Robots.txt
If I want to override a Disallow directive in robots.txt with an Allow command, do I have the Allow command before or after the Disallow command? example: Allow: /models/ford///page* Disallow: /models////page
Technical SEO | | irvingw0 -
Too many on page links
Hi All, As we all know, having to much links on a page is an obstacle for search engine crawlers in terms of the crawl allowance. My category pages are labeled as pages with to many "one page" links by the SEOmoz crawler. This probably comes from the fact that each product on the category page has multiple links (on the image and model number). Now my question is, would it help to setup a text-link with a clickable area as big as the product area? This means every product gets just one link. Would this help get the crawlers deeper in these pages and distribute the link-juice better? Or is Google smart enough already to figure out that two links to the same product page shouldn't be counted as two? Thanks for your replies guys. Rich
Technical SEO | | Horlogeboetiek0 -
Internal links to low value pages
Hi, We're doing a big content update on our product pages and I'm looking for some advice about our internal linking. In a nutshell, the current design we're using links out from every product page (i.e. plants) to a set of accessory pages (i.e. things to help you plant the plants). The screenshot shows how this works. The accessories we sell are a very small part of our business and don't attract significant or valuable search traffic. It's the plant pages that pull in the visits and make the money.
Technical SEO | | jdeb
The reason for all these links to accessory pages is for usabilty & to reduce the volume of support calls about accessories (we get a lot of those). So my concern is that by linking out to these relatively low value accessory pages from all of our plant product pages, we will be spilling link juice from all our important pages to a small set of unimportant ones. Should I be concerned about this and if so, what should I do differently? I have considered: Making an intermediary page, listing the relevant accessories, so that each product page links to one intermediary page, which then links to all the accessories. Using nofollow on the accessory page links - there is so much info out there about this, much of it conflicting, that I just don't know if that's a good or bad idea. Using some kind of java-based pop-up box to list the accessory links that will hide the links from spiders. Linking back from the accessory pages to the relevant product sub-category pages to loop the flow of link juice. All ideas welcome zoBgC0 -
What to do about "blocked by meta-robots"?
The crawl report tells me "Notices are interesting facts about your pages we found while crawling". One of these interesting facts is that my blog archives are "blocked by meta robots". Articles are not blocked, just the archives. What is a "meta" robot? I think its just normal (since the article need only be crawled once) but want a second opinion. Should I care about this?
Technical SEO | | GPN0