Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Can using nofollow on magento layered navigation hurt?
-
Howdy Mozzers!
We would like to use no follow, no index on our magento layered navigation pages after any two filters are selected. (We are using single filter pages as landing page, so we would liked them indexed)
Is it ok to use nofollow, noindex on these filter pages? Are there disadvantages of using nofollow on internal pages?
Matt mentioned refraining from using nofollow internally https://www.youtube.com/watch?v=4SAPUx4Beh8
But we would like to conserve crawling bandwidth and PR flow on potentially 100's of thousands of irrelevant/duplicate filter pages.
-
I understand I might be a little late, but I had experienced this issue first hand with a Magento site. Once I added a wildcard exclusion in the robots.txt file my impressions and click improved noticeably.
-
HI,
That is quite a few pages!
If the main issue is crawling related then robots.txt is probably the best way to go, I think the meta tags will still allow the pages to be crawled (they have to be for the tag to be read). Check out the comments in this and this post for wildcard matching in robots.txt which should do what you need. If the pages are indexed then it might be wise to leave a bit of time so that the noindex tags are picked up and then implement the crawl blocking in the robots.txt (and test in GWT to make sure you are not accidentally blocking more then you think). In this case I think you could still leave out the nofollow meta tag but this might just be personal opinion now - I'm not sure if in practice it would make much difference once you have no indexed and blocked crawling!
-
Hi Lynn,
Thank you for your valuable input on the matter. Yes, using meta tags in the header. We are currently submitting filter pages that we want indexed through the site map, so google bot should be able to reach these pages. Also, we are displaying noindex, nofollow tags only on filter pages which have a combination of more than two filters selected as we do not need to go any deeper than that.
I understand your point of using noindex, follow instead of noindex, nofollow to prevent unexpected crawl issues. But on the contrary, don't you think we could conserve crawling bandwidth using noindex, nofollow tags on filter pages that serve no purpose being crawled and probably wont be externally linked to either?
We currently have around 7 filters, some with many values. This can create combinations of more than 500,000 filter pages...
Thanks
-
Hi,
I assume you mean in a meta header tag for these pages? As a general rule I would avoid using nofollow and simply noindex the pages in question. If you are implementing this with a meta tag then the pages will be reached from the layered navigation links anyway so they would then be a dead end for both PR and the crawler - with the potential to cause unexpected crawl problems rather than optimising it.
As long as you are addressing as best you can any duplicate content issues caused by the layered navigation (check out this post for a good rundown on the various solutions) then I would leave the noindex in place and let the crawler follow the links as normal.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can subdomains hurt your primary domain's SEO?
Our primary website https://domain.com has a subdomain https://subDomain.domain.com and on that subdomain we have a jive-hosted community, with a few links to and fro. In GA they are set up as different properties but there are many SEO issues in the jive-hosted site, in which many different people can create content, delete content, comment, etc. There are issues related to how jive structures content, broken links, etc. My question is this: Aside from the SEO issues with the subdomain, can the performance of that subdomain negatively impact the SEO performance and rank of the primary domain? I've heard and read conflicting reports about this and it would be nice to hear from the MOZ community about options to resolve such issues if they exist. Thanks.
Intermediate & Advanced SEO | | BHeffernan1 -
Can you index a Google doc?
We have updated and added completely new content to our state pages. Our old state content is sitting in a our Google drive. Can I make these public to get them indexed and provide a link back to our state pages? In theory it sounds like a great link building strategy... TIA!
Intermediate & Advanced SEO | | LindsayE1 -
Does cache-control : private hurt SEO?
Hi, I recently found I can no longer view our web pages in Google's cache. I get 404 errors. http://webcache.googleusercontent.com/search?q=cache%3Ashop.nordstrom.com%2Fc%2Fwomens-shoes&rlz=1C1GGRV_enUS751US751&oq=cache%3Ashop.nordstrom.com%2Fc%2Fwomens-shoes&aqs=chrome..69i57j69i58.3575j0j4&sourceid=chrome&ie=UTF-8 I did a fetch and render in Search Console and found our header includes a "cache-control: private" entry. The 404's started happening recently. Would this response be the culprit? If Google cannot cache the website, is this bad for SEO? On the surface of it, it sounds bad.
Intermediate & Advanced SEO | | shop.nordstrom0 -
Can I use duplicate content in different US cities without hurting SEO?
So, I have major concerns with this plan. My company has hundreds of facilities located all over the country. Each facility has it's own website. We have a third party company working to build a content strategy for us. What they came up with is to create a bank of content specific to each service line. If/when any facility offers that service, they then upload the content for that service line to that facility website. So in theory, you might have 10-12 websites all in different cities, with the same content for a service. They claim "Google is smart, it knows its content all from the same company, and because it's in different local markets, it will still rank." My contention is that duplicate content is duplicate content, and unless it is "localize" it, Google is going to prioritize one page of it and the rest will get very little exposure in the rankings no matter where you are. I could be wrong, but I want to be sure we aren't shooting ourselves in the foot with this strategy, because it is a major major undertaking and too important to go off in the wrong direction. SEO Experts, your help is genuinely appreciated!
Intermediate & Advanced SEO | | MJTrevens1 -
Using "nofollow" internally can help with crawl budget?
Hello everyone. I was reading this article on semrush.com, published the last year, and I'd like to know your thoughts about it: https://www.semrush.com/blog/does-google-crawl-relnofollow-at-all/ Is that really the case? I thought that Google crawls and "follows" nofollowed tagged links even though doesn't pass any PR to the destination link. If instead Google really doesn't crawl internal links tagged as "nofollow", can that really help with crawl budget?
Intermediate & Advanced SEO | | fablau0 -
How to track mentions with links using Mentions App
We have a large client that we've just taken on board for organic search marketing. A great client that continually gets links from the BBC, NY Times etc which actually takes care of having to do any marketing for them. However, they get 10-15 unlinked mentions per day. Just recently, they peaked at 32 mentions in one day from JUST websites. But I need a quick way to filter these out and check if there is a link pointing to their website. I want to be able to build up a list of opportunities without having to manually check each website. How do, Mozzlars!?
Intermediate & Advanced SEO | | jasondexter0 -
Noindex, Nofollow to previous domain
Hi, My programmer recently did a horrible mistkae by adding noindex, nofollow to our website without me noticing for two days. At the same time he did it we bought a new domain and redirected the old domain to the new domain: The Old domain is: http://www.websitebuildersworld.com and the new one is: http://www.websiteplanet.com Now unfortunatly I didn't notice the noindex,nofollow when it was on the old domain and I redirected it to websiteplanet.com before I fixed the noindex, nofollow. I fixed the problem around 10 hours ago on the new domain (www.websiteplanet.com) but the old domain didn't get indexed back (yet), so for example if you search for WebsiteBuildersWorld in google you will not reach the homepage as google deleted it because of the noindex,nofollow. My question is:
Intermediate & Advanced SEO | | Ouzan
Do you think that it will be fixed and google will retrieve websitebuildersworld homepage to his search results and then redirect it to websiteplanet? Or because I redirected websitebuildersworld.com to websiteplanet.com before letting google crawling websitebuildersworld.com without the noindex,no follow it wouldn't get indexed again? I hope I explained the problem good enough. Looking forward for your valuable replies. Thanks.0 -
How do you implement dynamic SEO-friendly URLs using Ajax without using hashbangs?
We're building a new website platform and are using Ajax as the method for allowing users to select from filters. We want to dynamically insert elements into the URL as the filters are selected so that search engines will index multiple combinations of filters. We're struggling to see how this is possible using symfony framework. We've used www.gizmodo.com as an example of how to achieve SEO and user-friendly URLs but this is only an example of achieving this for static content. We would prefer to go down a route that didn't involve hashbangs if possible. Does anyone have any experience using hashbangs and how it affected their site? Any advice on the above would be gratefully received.
Intermediate & Advanced SEO | | Sayers1