Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
When removing a product page from an ecommerce site?
-
What is the best practice for removing a product page from an Ecommerce site?
If a 301 is not available and the page is already crawled by the search engine
A. block it out in the robot.txt
B. let it 404
-
Bryan,
If I were removing 100 product pages from an eCommerce site because they barely convert I would approach it this way:
#1. Run the URls through a tool to see which ones have external backlinks.
Often times none of the pages will have any external backlinks, and those that do are usually not very good. If this is the case - or if you "really" aren't able to do any 301 redirects (and if so that's something that needs to be fixed) - skip to step #3. Otherwise...#2. 301 redirect those with external backlinks to the most relevant page, be that a similar product or the category page directly above the product to be removed. Try to avoid redirecting them all to the homepage or some other "catch all" page, as these may be treated like a 404 by Google.
#3. Simply remove the pages and show the custom 404 page that suggests other products, or whatever messaging you want to show there (e.g. "This product has been removed from our catalog. Please see these other products...") and be sure to check the http header response code (lots of free tools for that) to ensure these URLs actually serve a 404 response (note: This should show up on the removed URL, as opposed to redirecting the visitor to another page like .../404.html).
#4. Since the now-removed URLs are not linked to from anywhere, either internally or externally, it could take awhile for Google to recrawl them and see the 404 error. If you need this to happen more quickly, such as when dealing with duplicate manufacturer descriptions and removal of page to recover from Panda, it may be wise to provide some type of html sitemap file listing out the URLs long enough for Google to recrawl them.
I would not block them in the robots.txt file, as that could result in Google not seeing the 404 and not removing it from the index (though they will cease to show the meta description).
-
Okay the question is regarding indexing, I should of been more specific.
If we are removing 100 product pages from an ecommerce site because they barely convert (regardless of a nice 404 page) and we cannot transfer the user to a relevant page. Is it a best to leave the pages live? or remove them (404) and block them in the robots.txt file?
-
Hi Bryan,
There are various reasons to remove a product page from an eCommerce store. Before deciding to remove a product page, you should consider if removing it will in fact help your SEO. If not, you need to look into alternatives such as 301 redirects or informing visitors in the old product pages that the product is no longer available. I'm not sure why performing 301 redirects is not an option for you - you may want to consider trying to get access to do this.
We have written an article some time ago about the different scenarios an eCommerce store will face when deciding to remove old product pages, and how to deal with each scenario: http://blog.referralcandy.com/2011/12/14/how-to-remove-old-products/
Hope that helps!
-
Right the questions is regarding crawlability, link juice etc..
-
I would create a custom 404 page that gives users options of similar products or product categories.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Rel canonical tag from shopify page to wordpress site page
We have pages on our shopify site example - https://shop.example.com/collections/cast-aluminum-plaques/products/cast-aluminum-address-plaque That we want to put a rel canonical tag on to direct to our wordpress site page - https://www.example.com/aluminum-plaques/ We have links form the wordpress page to the shop page, and over time ahve found that google has ranked the shop pages over the wp pages, which we do not want. So we want to put rel canonical tags on the shop pages to say the wp page is the authority. I hope that makes sense, and I would appreciate your feeback and best solution. Thanks! Is that possible?
Intermediate & Advanced SEO | | shabbirmoosa0 -
Schema markup concerning category pages on an ecommerce site
We are adding json+ld data to an ecommerce site and myself and one of the other people working on the site are having a minor disagreement on things. What it comes down to is how to mark up the category page. One of us says it needs to be marked up with as an Itempage, https://schema.org/ItemPage The other says it needs to be marked up as products, with multiple product instances in the schema, https://schema.org/Product The main sticking point on the Itemlist is that Itemlist is a child of intangible, so there is a feeling that should be used for things like track listings or other arbitrary data.
Intermediate & Advanced SEO | | LesleyPaone2 -
Should I use noindex or robots to remove pages from the Google index?
I have a Magento site and just realized we have about 800 review pages indexed. The /review directory is disallowed in robots.txt but the pages are still indexed. From my understanding robots means it will not crawl the pages BUT if the pages are still indexed if they are linked from somewhere else. I can add the noindex tag to the review pages but they wont be crawled. https://www.seroundtable.com/google-do-not-use-noindex-in-robots-txt-20873.html Should I remove the robots.txt and add the noindex? Or just add the noindex to what I already have?
Intermediate & Advanced SEO | | Tylerj0 -
Handling of product variations and colours in ecommerce
Hi, our site prams.net has 72.000 crawled and only 2500 indexed urls according to deep crawl mainly due to colour variations (each colour has its own urls now). We now created 1 page per product, eg http://www.prams.net/easywalker-mini and noindexed all the other ones, which had a positive effect on our seo. http://www.prams.net/catalogsearch/result/?q=002.030.059.0 I might still hurt our crawl budget a lot that we have so many noindexed pages. The idea is now to redirect 301 all the colour pages to this main page and make them invisible. So google do not have to crawl them anymore, we included the variations in the product pages, so they should still be searchable for google and the user. Does this make sense or is there a better solution out there? Does anyone have an idea if this will likely have a big or a small impact? Thanks in advance. Dieter
Intermediate & Advanced SEO | | Storesco0 -
Best practice for removing indexed internal search pages from Google?
Hi Mozzers I know that it’s best practice to block Google from indexing internal search pages, but what’s best practice when “the damage is done”? I have a project where a substantial part of our visitors and income lands on an internal search page, because Google has indexed them (about 3 %). I would like to block Google from indexing the search pages via the meta noindex,follow tag because: Google Guidelines: “Use robots.txt to prevent crawling of search results pages or other auto-generated pages that don't add much value for users coming from search engines.” http://support.google.com/webmasters/bin/answer.py?hl=en&answer=35769 Bad user experience The search pages are (probably) stealing rankings from our real landing pages Webmaster Notification: “Googlebot found an extremely high number of URLs on your site” with links to our internal search results I want to use the meta tag to keep the link juice flowing. Do you recommend using the robots.txt instead? If yes, why? Should we just go dark on the internal search pages, or how shall we proceed with blocking them? I’m looking forward to your answer! Edit: Google have currently indexed several million of our internal search pages.
Intermediate & Advanced SEO | | HrThomsen0 -
How To Best Close An eCommerce Site?
We're closing down one of our eCommerce sites. What is the best approach to do this? The site has a modest link profile (a young site). It does have a run of site link to the parent site. It also has a couple hundred email subscribers and established accounts. Is there a gradual way to do this? How do I treat the subscribers and account holders? The impact won't be great, but I want to minimize collateral damage as much as possible. Thanks.
Intermediate & Advanced SEO | | AWCthreads0 -
Should the sitemap include just menu pages or all pages site wide?
I have a Drupal site that utilizes Solr, with 10 menu pages and about 4,000 pages of content. Redoing a few things and we'll need to revamp the sitemap. Typically I'd jam all pages into a single sitemap and that's it, but post-Panda, should I do anything different?
Intermediate & Advanced SEO | | EricPacifico0