Why are Pages returning 404 errors not being dropped?
-
Our webmaster tools continues to return anywhere upwards of 750 pages that have 404 errors. These are from pages of a previous site no longer used.
However this was over 1 year ago these pages were dropped along with the 301 re-directs. Why is Google not clearing these from webmaster tools but re-listing them again after 3 month cycle? Is it because external sites have links to these pages?
If so should I put a 301 in place (most of these site are forums and potentially dodgy directories etc from previous poor link building programs) or ask for a manual removal?
-
Thanks tom for all your help.
Regards
Craig
-
Very good point you've raised - 301ing those URLs effectively makes the links to your site "live" again. If the links sit on a dodgy/spammy/poor quality page, then it could harm your site and I wouldn't put the redirect in place.
By in large, if you're beginning to doubt whether the link is worthwhile or not, chances are its not. So if you have a bit of doubt about the link, then don't put the 301 in place.
-
Hi Tom,
That more than explains it and gives me the answers. If I put 301 redirects in place what will happen if any of these external links are bad, will it harm our site? Its taken me many months to deal with duplicate content issues, canonicalisation of the site and much more. It was a complete mess and I don;t want to harm any good that come of all this.
-
Hi Craig
You touched on one of the reasons this is happening in your post - you could external links to these pages. Also, they could still be appearing in the sitemap.
If you go into Webmaster tools > Health > Crawl Errors > Not Found and then click on one of the URLs, you can check whether or not the page is in the sitemap or whether it is being linked to from somewhere.
If you have external links, you have four options. First, you could attempt to change the URLs on the pages they're being linked from. This could be difficult and/or long. Second, as you say, you could 301 redirect. This would be useful if people are coming through those sites still, as you'll be fixing their user journey. It would also pass on any link "juice" that page has to another. Third would be to start returning a 410 error. This explains 410 response codes - it basically tells the Googlebot to treat the URL as gone permanently. This can be a bit tricky to setup and you have to be sure you want use the URL again in the future.
Finally, you could leave the 404s in place. If none of the pages have any strength, no referral traffic is coming from them and they aren't interrupting a user journey in any way, I would simply leave them. Google knows that 404s are just a matter of process and so recognises that 404 errors are simply a natural occurrence. It would only ever be a problem if you returned tens of thousands of them, so you may just want to leave them be.
I would probably 301 redirect any old pages carrying strength to relevant equivalents (if not, the root domain) and leave the other 404s in place. I would rewrite ASAP any URL that is interrupting a user journey.
Hope this helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
301 Redirect to Home Page or Sub-Page?
What do you think about 301 redirect of good expired domain to a sub-page instead of the home page? I'm doing this so I don't hurt my brand name. Let me know your thoughts please. Thank you
Intermediate & Advanced SEO | | JuanWork0 -
Why have my rankings dropped?
I have a client who has just seen his average page rank creep up from around 39 to 34 over about two months, then it appears to have dropped back to position 40+ in the space of a week. I believe he's made a lot of changes to targeted keywords, so I'd like to think it's simply because his old targeted keywords are dropping and new keywords still have to build their rankings. But I'm also worried in case he has over-optimised and might get getting penalised. Any advice on where to start digging?
Intermediate & Advanced SEO | | muzzmoz0 -
Category Pages
I'm debating on what the best category structure is for a recipe website and was looking to get some advice. It's a recipe/travel/health fitness blog but recipes reign on the site. Should it be: Option A website name\recipe\type of recipe\URL of specific recipe or Option B website name\type of recipe\url of specific recipe (and just cut out the 'recipe' category name) Any advise would be appreciated! Thanks!
Intermediate & Advanced SEO | | Rich-DC0 -
Using two 404 NOT FOUND pages
Hi all, I was wondering if any of you can advise whether it's no issue to use two separate custom 404 pages. The 404 pages would be different for different parts of the site. For instance, if you're on /community/ and you enter a non-existing page on: www.sample.com/community/example/ it would give you a different 404 page than someone who runs into a non existing page at: www.sample.com/definition/example/ Does anybody have experience with this and would this be fine?
Intermediate & Advanced SEO | | RonFav0 -
Urgent Site Migration Help: 301 redirect from legacy to new if legacy pages are NOT indexed but have links and domain/page authority of 50+?
Sorry for the long title, but that's the whole question. Notes: New site is on same domain but URLs will change because URL structure was horrible Old site has awful SEO. Like real bad. Canonical tags point to dev. subdomain (which is still accessible and has robots.txt, so the end result is old site IS NOT INDEXED by Google) Old site has links and domain/page authority north of 50. I suspect some shady links but there have to be good links as well My guess is that since that are likely incoming links that are legitimate, I should still attempt to use 301s to the versions of the pages on the new site (note: the content on the new site will be different, but in general it'll be about the same thing as the old page, just much improved and more relevant). So yeah, I guess that's it. Even thought the old site's pages are not indexed, if the new site is set up properly, the 301s won't pass along the 'non-indexed' status, correct? Thanks in advance for any quick answers!
Intermediate & Advanced SEO | | JDMcNamara0 -
Wrong page being ranked
Hi there, This seems a bit of a strange one, I have a particular keyword which I am trying to rank for, all internal links with the appropriate anchor text are pointing to the page I want to rank for, for this particular keyword, all external links are pointing to the page I want to rank for, for this particular keyword, however Google is ranking another page on my website for this keyword and the bizarre things is the page which is being ranked is a .PDF I am really not sure what else to do to give Google the hint that they are ranking the wrong page, any ideas? Kind Regards
Intermediate & Advanced SEO | | Paul780 -
Should we deindex duplicate pages?
I work on an education website. We offer programs that are offered up to 6 times per year. At the moment, we have a webpage for each instance of the program, but that's causing duplicate content issues. We're reworking the pages so the majority of the content will be on one page, but we'll still have to keep the application details as separate pages. 90% of the time, application details are going to be nearly identical, so I'm worried that these pages will still be seen as duplicate content. My question is, should we deindex these pages? We don't particularly want people landing on our application page without seeing the other details of the program anyway. But, is there problem with deindexing such a large chunk of your site that I'm not thinking of? Thanks, everyone!
Intermediate & Advanced SEO | | UWPCE0 -
Why duplicate content for same page?
Hi, My SEOMOZ crawl diagnostic warn me about duplicate content. However, to me the content is not duplicated. For instance it would give me something like: (URLs/Internal Links/External Links/Page Authority/Linking Root Domains) http://www.nuxeo.com/en/about/contact?utm_source=enews&utm_medium=email&utm_campaign=enews20110516 /1/1/31/2 http://www.nuxeo.com/en/about/contact?utm_source=enews&utm_medium=email&utm_campaign=enews20110711 0/0/1/0 http://www.nuxeo.com/en/about/contact?utm_source=enews&utm_medium=email&utm_campaign=enews20110811 0/0/1/0 http://www.nuxeo.com/en/about/contact?utm_source=enews&utm_medium=email&utm_campaign=enews20110911 0/0/1/0 Why is this seen as duplicate content when it is only URL with campaign tracking codes to the same content? Do I need to clean this?Thanks for answer
Intermediate & Advanced SEO | | nuxeo0