Webmaster tools crawl errors
-
Hi there,
iv been tracking my webmaster tools crawl errors for a while now(6 months) and im noticing some pages that are far gone 404 are still poping out on the crawl errors. - that pages have no data for xml linking, and remote linking are from pages that are far gone 404 also.
that pages have 404 error page + redirect to homepage, and google still notice them with old cache content.
does someone have a clue why is this happening?
-
Thank you very much Dana for the superb answer!
any clue for how much this errors are critical for my website seo? (is this problem worth fixing)
-
Hi!
Have you verified that there is a proper 301 redirect from the old URL's? If, for example, there is a 302 instead, I wouldn't be surprised if Google kept the old info in the index, as you have sort of said "I'll be back soon at this old address, just wait a minute (or year)".
Do you have an example URL for us that we could take a look at?
-
Hi Dana,
If your looking for someone to confirm what Dana said she is 100% right.
If your using a data base like WordPress it is posable for it to hold old pages and serve them. Make sure your host knows of your problem if you are using a CMS.
Hope this helps,
Thomas
-
Yes, I understand this issue very well and have seen it many times. Most often, it is happening because another of your pages that is being indexed is still referencing those pages. Please forgive Google, it is but a humble, not-so-intelligent bot (despite what the world would have you think). If you keep referencing a page that doesn't exist, Googlebot will say to itself "but it does exist! it does!" and keep indexing it, despite the 404 error. I mean, for all Google knows, it is your intention to bring that page back and maybe you just screwed up and it is 404-ing, you know?
Here's the remedy:
Do a content audit. Here's a great post on how to do that: http://www.distilled.net/blog/seo/how-to-perform-a-content-audit/
You will discover many things, without a doubt, including pages that are linking to these 404 pages. Decide what to do with those pages, i.e. nix them, re-build them, whatever. If you have pages in Google that you really do want out of the index and those same pages are 404-ing....simply go to your Google Webmaster Tools Account and in the left nav select "Google Index" and then "Remove URLs." Then simply click the box that says "create a new removal request" and enter the desired URL into that Box. Provided your page meets the requirements (and if it's producing a 404 error it does) Google will prioritize its removal. Under no circumstances should this be confused with the disavow tool. Just to make that clear.
I hope this helps. Any questions and I'm happy to help further if I can.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How can I remove parameters from the GSC URL blocking tool?
Hello Mozzers My client's previous SEO company went ahead and blindly blocked a number of parameters using the GSC URL blocking tool. This has now caused Google to stop crawling many pages on my client's website and I am not sure how to remove these blocked parameters so that they can be crawled and reindexed by Google. The crawl setting is set to "Let Google bot decide" but still there has been a drop in the number of pages being crawled. Can someone please share their experience and help me delete these blocked parameters from GSC's URL blocking tool. Thank you Mozzers!
Reporting & Analytics | | Vsood0 -
Moving to https - Webmaster Tools
Okay so I've moved a site I work on to https finally... wasn't easy, but I won't go into the whys and hows of that. Everything is working fine on the site and my rankings are holding steady but when I went into the GWT today I noticed that my main site (www.domain.com) has lost inquiry and traffic data almost entirely. So I quickly added and verified the https://www.domain.com URL (even though it doesn't ask for the protocol when adding a site) and saw the traffic.. kind of. Only 2 inquiries and a few pages indexed. Now I realize I need to submit a new sitemap and am currently working on that but don't know which account to upload it to. Do I really need two GWT sites separately and why is the other one still getting the impressions and clicks? Basically I'm looking for somebody to shed some light on how to handle this migration from a GWT standpoint. I've attached 3 screenshots to illustrate what I'm talking about. Should be obvious what's what to those who read this post. The Analytics shot is to show that I haven't actually lost traffic in this time and eliminate the fear of penalty or anything of the sort. Thanks! wbPpwVw.png YJifLiR.png cRVgqRN.png
Reporting & Analytics | | jesse-landry0 -
How can I see what Google sees when it crawls my page?
In other words, how can see the text and what not it sees from start to finish on each page. I know there was a site, but I can't remember it.
Reporting & Analytics | | tiffany11030 -
Webmaster Tools, why does it show 486 pages submitted to web, and only 40 indexed?
I am confused on what a client account shows in WMTs, client account is http://multiview.com. They have a graph showing 486 pages submitted to web, but only 40 are indexed. Also, they recently re-launched, i.e in April 2014, and the new site has about 40 pages indexed.... so I am guessing that the 486 number relates to all the pages that are showing errors in retrieving...i.e. 28 soft 404 errors, 10 access denied errors, 808 not found errors. Does this make sense to explain why there is such a gap between 486 and 40?
Reporting & Analytics | | DianeDP0 -
Webmaster tools help
Looking at some Google Webmaster tools data and im a bit stumped. Can anyone tell me if the "not selected" pages should be taken from the "total indexed pages" or are they a seperate entity? The way the data is displayed on the graph leads you to believe they should be taken from the total indexed pages, yet Googles own description of "not selected" pages says that these pages are not indexed. Confused.
Reporting & Analytics | | Silkstream0 -
My Google Webmaster Search Queries are 0??!!
For the past 2 days Google Webmaster has recorded 0 search queries to my site, but traffic seems stable. Any ideas about why this is happening and what I should do about it? Or is it a Google glitch I should not be concerned about (hoping for this option)
Reporting & Analytics | | theLotter0 -
Fb, twitter, etc stats tool?
I once saw a tool (from Distilled, I think?) where you would input your blog post or page URL, and it would give you a table showing how many Facebook likes, Tweets, etc that page has gotten. Does anyone have the URL for this or a similar tool?
Reporting & Analytics | | AdamThompson0 -
How long does it take to get results from the Term Extractor Tool?
Hi! I entered a web page into the Term Extractor Tool, and it's been running for about 3 minutes now. How long does it usually take? Thanks 🙂 Kelley
Reporting & Analytics | | kinsana0