The crawl report shows a lot of 404 errors
-
They are inactive products, and I can't find any active links to these product pages. How can I tell where the crawler found the links?
-
That's too easy Keri.
-
If you download the CSV of the report, there is a column that will list the referring URL for the 404.
-
Mike's suggestion is a great one. SF gives some great data that's easy to play with in Excel. The free version only crawls 500 pages, so if you have a small site, you'll probably get what you need. (I got 30k+ pages, so I use the paid version that's only about $160/yr.)
-
There is nothing wrong with having a "404 error" if the page is an expired product. Obviously, you don't wont to be linking to a 404 page on your site, so I'd suggest using a tool like Screaming Frog, potentially even OSE, and monitoring your 404 pages in Google Webmaster Tools to see if it is still currently being linked to.
If the page has external links pointing to it, I'd recommend 301 redirecting it to whatever category/subcategory the product belongs to.
-
Do a scan with Screaming Frog. It can point out exact pages where you are having broken links.
They have a free trial and a paid version.
This program is great for diagnosing many different website related issues.
Hope this helps.
Mike
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do fix an 803 Error?
I got am 803 error this week on the Moz crawl for one of my pages. The page loads normally in the browser. We use cloudflare. Is there anything that I should do or do I wait a week and hope it disappears? 803 Incomplete HTTP response received Your site closed its TCP connection to our crawler before our crawler could read a complete HTTP response. This typically occurs when misconfigured back-end software responds with a status line and headers but immediately closes the connection without sending any response data.
Moz Pro | | Zippy-Bungle1 -
Duplicate Errors found in my search
I have run my 1st site check with SEOMOZ and have 4000+ errors. The "duplicate Page Content" culprit appears to be a extended url that keeps showing as duplicating. This is only a customer log-in and can be redirected back to the main cust log in page, but is there a short way of doing it (rather than 4000x 301's)? The format of the url is: http://www.????.com.au/default/customer/account/login/referer/aSR0cDovL3d3dy1234YWNiYW Thanks
Moz Pro | | Paul_MC0 -
Crawl Diagnostics Shows thousands of 302's from a single url. I'm confused
Hi guys I just ran my first campaign and the crawl diagnostics are showing some results I'm unfamiliar with.. In the warnings section it shows 2,838 redirects.. this is where I want to focus. When I click here it shows 5 redirects per page. When I go to click on page 2, or next page, or any other page than page 1 for that matter... this is where things get confusing. Nothing shows. Downloading the csv reveals that 2,834 of these are all showing: URL: http://www.mydomain.com/401/login.php url: http://www.mydomain.com/401/login.php referrer: http://www.mydomain.com/401/login.php location_header: http://www.mydomain.com/401/login.php I guess I'm just looking for an explanation as to why it's showing so many to the same page and what possible actions can be taken on my part to correct it (if needed). Thanks in advance
Moz Pro | | sethwb0 -
Best Automated Report?
I would like to implement a reporting function to my website to offer a bit of value and information to potential clients. I am thinking along the lines of a simple input form to include a business name and url. The output would be a clean, branded (my business, url, phone) report that shows opportunities, lowest lying fruit, keywords most prominent, and any errors. I found this site - http://www.analyticsseo.com/ but, it seems pretty expensive. Does anyone have any suggestions on another suite that might work? Many thanks!
Moz Pro | | adell500 -
Crawl Errors Confusing Me
The SEOMoz crawl tool is telling me that I have a slew of crawl errors on the blog of one domain. All are related to the MSNbot. And related to trackbacks (which we do want to block, right?) and attachments (makes sense to block those, too) ... any idea why these are crawl issues with MSNbot and not Google? My robots.txt is here: http://www.wevegotthekeys.com/robots.txt. Thanks, MJ
Moz Pro | | mjtaylor0 -
Exporting ranking stats; errors
previously, when I went to export the weekly rank stats, the csv opened into excel and everything was fine. This week, one campaign data will not export other than purely in comma deliniated and will not translate to colums in ecel. A second campaign does export directly to excel but the data showing (on 27th June) is for the 22 february (report created march 1). so all my spread sheets for position analysis are up the creek and I cant even get last weeks positional report on my key domain. any thoughts??
Moz Pro | | Citybase0 -
Crawl Diagnostics bringing 20k+ errors as duplicate content due to session ids
Signed up to the trial version of Seomoz today just to check it out as I have decided I'm going to do my own SEO rather than outsource it (been let down a few times!). So far I like the look of things and have a feeling I am going to learn a lot and get results. However I have just stumbled on something. After Seomoz dones it's crawl diagnostics run on the site (www.deviltronics.com) it is showing 20,000+ plus errors. From what I can see almost 99% of this is being picked up as erros for duplicate content due to session id's, so i am not sure what to do! I have done a "site:www.deviltronics.com" on google and this certainly doesn't pick up the session id's/duplicate content. So could this just be an issue with the Seomoz bot. If so how can I get Seomoz to ignore these on the crawl? Can I get my developer to add some code somewhere. Help will be much appreciated. Asif
Moz Pro | | blagger0