Find pages containing broken links.
-
hi everyone,
for each internal broken links I need to find all the pages that contain it.
In the Seomoz report there is only a refferer link for each broken link, but google webmaster tools indicates that the dead link is present in many pages of the site.
there is a way to have these data with SEOmoz or other software, in a csv report ?
thanks
-
Hi wwmind,
I was going suggest Xenu, but also Screaming Frog. There's a both a free and paid version. To crawl more than 500 URLs, you'd need to get a license for the paid version.
http://www.screamingfrog.co.uk/seo-spider/
The other option, if you can't find all of your broken links, is to redirect them via a 301. Installing a bunch of 301's via .htaccess or another means isn't as good of a solution as fixing your links, but it might offer a more scalable solution.
Regardless, best of luck with your SEO!
-
I've used Xenu's Link Sleuth on sites with +5.000 pages. Worked just fine for me. But two setups are never equally alike.
But if you re not satisfied with Xenu try looking for some of the pro sitemap apps - a lot of them come with a "broken links checker" tool ...
-
I had problems with Xenu on large sites.
-
Hi
I would use Xenu's Link Sleuth (freeware)
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page authority
Hello, How can my page authority be different across various page built exactly on the same model and none of them having links ? Thank you,
Moz Pro | | seoanalytics0 -
Filter Pages
Howdy Moz Forum!! I have a headache of a job over here in the UK and I'd welcome any advice! - It's sunny today, only 1 of 5 days in a year and i'm stuck on this! I have a client that currently has 22,000 pages indexed to Google with almost 4000 showing as duplicate content. The site has a "jobs" and "candidates" list. This can cause all sorts of variations such as job title, language, location etc. The filter pages all seem to be indexed. Plus the static pages are indexed. For example if there were 100 jobs at Moz being advertised, it is displaying the jobs on the following URL structure - /moz
Moz Pro | | Slumberjac
/moz/moz-jobs
/moz/moz-jobs/page/2
/moz/moz-jobs/page/3
/moz/moz-jobs/page/4
/moz/moz-jobs/page/5 ETC ETC Imagine this with some going up to page/250 I have checked GA data and can see that although there are tons of pages indexed this way, non of them past the "/moz/moz-jobs" URL get any sort of organic traffic. So, my first question! - Should I use rel-canonical tags on all the /page/2 & /page/3 etc results and point them all at the /moz/moz-jobs parent page?? The reason for this is these pages have the same title and content and fall very close to "duplicate" content even though it does pull in different jobs... I hope i'm making sense? There is also a lot of pages indexed in a way such as- https://www.examplesite.co.uk/moz-jobs/search/page/9/?candidate_search_type=seo-consulant&candidate_search_language=blank-language These are filter pages... and as far as I'm concerned shouldn't really be indexed? Second question! - Should I "no follow" everything after /page in this instance? To keep things tidy? I don't want all the variations indexed! Any help or general thoughts would be much appreciated! Thanks.0 -
Clearing our on-page ranking reports?
Is there a way to "bulk delete" on-page ranking reports which are no longer relevant? I know we can delete them one at a time, but the reason I ask is that I've done a fair bit of work changing URL's, so the reports are often for old URL's which no longer exist. (yes, I made sure to do 301 redirects to the new ones!) Thanks in advance for any help!
Moz Pro | | koalatm0 -
On Link Analysis tab I my best pages are 301 and 404 pages.
I looked on my redirrect file and found that /* redirects to /v/404.asp.
Moz Pro | | sbetzen
However if you look below at the link analysis the 404 page is getting a 404 error.
The homepage ecowindchimes.com/ is getting a 301 (but I don't know where it is going to).
The third one is also redirected. 1. [No Data] ecowindchimes.com/ ||| 301 ||| 2 ||| 36 2. 2. [No Data] ecowindchimes.com/v/404.asp ||| 404 ||| 2 ||| 34 3. [No Data] 3. ecowindchimes.com/index.html?lang=en-us&target=d2.html ||| 301 ||| 1 ||| 33 So I have 2 questions: 1) should this be fixed? and 2) how? This is a volusion site and I believe the "catchall" redirect was done by them0 -
Duplicate content pages
Crawl Diagnostics Summary shows around 15,000 duplicate content errors for one of my projects, It shows the list of pages with how many duplicate pages are there for each page. But i dont have a way of seeing what are the duplicate page URLs for a specific page without clicking on each page link and checking them manually which is gonna take forever to sort. When i export the list as CSV, duplicate_page_content column doest show any data. Can anyone please advice on this please. Thanks <colgroup><col width="1096"></colgroup>
Moz Pro | | nam2
| duplicate_page_content |1 -
Sorting Dupe Content Pages
Hi, I'm no excel pro, and I'm having a bit of a challenge interpreting the Crawl Diagnostics export .csv file. I'd like to see at a glance which of my pages (and I have many) are the worst offenders for dupe content – ie. which have the most "Other URLs" associated with them. Thanks, would appreciate any advice on how other people are using this data, and/or how 'Moz recommends to do it. 🙂
Moz Pro | | ntcma0 -
Old Incoming Links Redirected to new pages are not Being Factored on the Open Site Explorer
Hi, My website has been online since 1994. We have old links pointing to pages that no longer exist so what we have done is to create redirects to the specific page where the content is being displayed now. However, when we use the Open Site Explorer, the linking root domains do not show those sites that contain old links pointing to our domain. One good example of this one is: dmoz.org. Am I improperly handling the redirects? Or, what do I have to do so that old links that are being redirected to the new pages where the information is are accounted when calculating domain authority and trust? Thanks, Alex
Moz Pro | | costarica.com0 -
Why is Followed Linking Root Domains higher than External Followed Links?
Surely there must be at least one external link for each linking root domain? Some results for smaller sites give a higher number of domains linking in than incoming links - e.g. www.forbesandsawyer.co.uk Under Subdomain metrics: External Followed Links - 1 Followed Linking Root Domains - 2 Surely 2 root domains would mean AT LEAST 2 external followed links? Thanks, Andrew
Moz Pro | | Silktide0