Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Best free tool to check internal broken links
-
Question says it all I guess. What would your recommend as the best free tool to check internal broken links?
-
[Spammy comment removed by forum moderator.]
-
How awesome is that Screaming Frog tool?
-
Hey Kevin,
If you download the CSV report, we should list the referring page in the CSV. I know it'd be better to have it on the on-screen report, and it is on our wish list, but you should be able to see the info in an export.
-
Hi Kevin
Thanks - yes, although I am a Moz Associate, out of full transparency I agree with that you are saying. It is not a tool I specifically use with Moz, and do prefer Screaming Frog for the exact functions you are referring to. Screaming Frog will tell you what pages the bad links are on.
You can definitely pass this on to the product team as a feature request on this page. Personally I love tools like their analytics, keyword difficulty, rank tracker, open site explorer, but I agree from this aspect the crawler is not the strongest, and I would suggest supplementing with Screaming Frog (as I do).
-Dan
-
Dan, although I love me some MOZ the crawl diagnostics kinda suck... all they do is report errors but they don't give any insight on where the bad link is originated... it merely shows the page which is 404ed, which is a BIG FAT WASTE OF TIME.
You guys should know that we need to know WHICH PAGE THE BAD LINK IS ON, and furthermore WHAT LINE OF CODE HAS THE BAD LINK. Who cares about the broken page!?!?!?!?!?
-
Yes when you set up a campaign with Moz Analytics, Moz will crawl your website and return a whole report of suggestions. It will start a sample crawl which gets returned to you pretty quickly, and a full crawl a little bit thereafter. It will report things to you like 400 errors etc.
You can also use the Moz Crawl Test (PRO only as well) which will return to you the HTTP status code of each URL crawled.
Optionally, you can also use Screaming Frog SEO Spider (free up to 500 pages to crawl, paid for bigger sites).
-Dan
-
Does anyone have any clue on how I can check my internal broken links using MOZ tools?
-
Hello,
If you have a full Sitemap, transform it to a .txt by putting only URL (Excel can do that easily) try Screaming Frog SEO on list mode, which is free regardless of the number of links.
Even if you search a free tool, I strongly recommand to invest 99£/year on this tool, it's worth it.
-
Ok, first, does the url with /errors/error_404 (without the query string) exist?
Also - are you using something to create your 404 pages for tracking purposes?
The query string on the url would appear to be stating the referring url anyway - hence ?q=-i3 basically equates to the page /i3
At a guess - it could be there is a script running to create logs on specific 404's, or create a new log each time one occurs - if this folder is visable to the crawlers, it would get spidered and subsequently the problem would arise.
-
Yeah I tried that but the URL in the URL coloumn shows for example:
/errors/error_404?q=-i3
and the Referrer shows:
/-i3
Neither page exists and so I do not know where the page which contains this broken link is...
-
It does, when you get the report, filter column D marked "4XX (Client Error)" and you will see your 404's there - further along in the report you will also be able to see referring url which will show you which page is linking to it.
-
Checked out the tool I just linked to on here and it doesn't include broken link data
-
Is the the custom crawl tool on here you are referring too? http://pro.seomoz.org/tools/crawl-test
-
Custom crawl right here on SEOmoz, or you could use Xenu
-
Hi Dan,
Thanks for that tool. It works great and especially like that it works in Chrome.
Do you however know of a tool which would do this site wide rather than per page?
-
This tool is a great extension for Chrome
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best redirect destination for 18k highly-linked pages
Technical SEO question regarding redirects; I appreciate any insights on best way to handle. Situation: We're decommissioning several major content sections on a website, comprising ~18k webpages. This is a well established site (10+ years) and many of the pages within these sections have high-quality inbound links from .orgs and .edus. Challenge: We're trying to determine the best place to redirect these 18k pages. For user experience, we believe best option is the homepage, which has a statement about the changes to the site and links to the most important remaining sections of the site. It's also the most important page on site, so the bolster of 301 redirected links doesn't seem bad. However, someone on our team is concerned that that many new redirected pages and links going to our homepage will trigger a negative SEO flag for the homepage, and recommends instead that they all go to our custom 404 page (which also includes links to important remaining sections). What's the right approach here to preserve remaining SEO value of these soon-to-be-redirected pages without triggering Google penalties?
Technical SEO | | davidvogel0 -
Should internal links in my table of contents be tagged as nofollow?
Hi All, I have the LuckyWP Table of Contents plugin installed. I recently noticed that you can tag your internal links with and nofollow. I understand that it's always a good idea to link internally and to pass link juice to my own content. But with detailed posts that have over 20 headings, I'm thinking that internal linking for headings may actually hurt me because it takes my links well above 100. Any ideas what the best practises are in this scenario? Thanks.
Technical SEO | | nomad_blogger0 -
How can I stop a tracking link from being indexed while still passing link equity?
I have a marketing campaign landing page and it uses a tracking URL to track clicks. The tracking links look something like this: http://this-is-the-origin-url.com/clkn/http/destination-url.com/ The problem is that Google is indexing these links as pages in the SERPs. Of course when they get indexed and then clicked, they show a 400 error because the /clkn/ link doesn't represent an actual page with content on it. The tracking link is set up to instantly 301 redirect to http://destination-url.com. Right now my dev team has blocked these links from crawlers by adding Disallow: /clkn/ in the robots.txt file, however, this blocks the flow of link equity to the destination page. How can I stop these links from being indexed without blocking the flow of link equity to the destination URL?
Technical SEO | | UnbounceVan0 -
Will Google Recrawl an Indexed URL Which is No Longer Internally Linked?
We accidentally introduced Google to our incomplete site. The end result: thousands of pages indexed which return nothing but a "Sorry, no results" page. I know there are many ways to go about this, but the sheer number of pages makes it frustrating. Ideally, in the interim, I'd love to 404 the offending pages and allow Google to recrawl them, realize they're dead, and begin removing them from the index. Unfortunately, we've removed the initial internal links that lead to this premature indexation from our site. So my question is, will Google revisit these pages based on their own records (as in, this page is indexed, let's go check it out again!), or will they only revisit them by following along a current site structure? We are signed up with WMT if that helps.
Technical SEO | | kirmeliux0 -
Best practices for controlling link juice with site structure
I'm trying to do my best to control the link juice from my home page to the most important category landing pages on my client's e-commerce site. I have a couple questions regarding how to NOT pass link juice to insignificant pages and how best to pass juice to my most important pages. INSIGNIFICANT PAGES: How do you tag links to not pass juice to unimportant pages. For example, my client has a "Contact" page off of there home page. Now we aren't trying to drive traffic to the contact page, so I'm worried about the link juice from the home page being passed to it. Would you tag the Contact link with a "no follow" tag, so it doesn't pass the juice, but then include it in a sitemap so it gets indexed? Are there best practices for this sort of stuff?
Technical SEO | | Santaur0 -
International Site Links In Footer
We have several international sites and we have them linked in the footer of our main .com site . Should we add "nofollow" to these links? Our concern is that Google could see these sites as a network?
Technical SEO | | EwanFisher0 -
Does Google pass link juice a page receives if the URL parameter specifies content and has the Crawl setting in Webmaster Tools set to NO?
The page in question receives a lot of quality traffic but is only relevant to a small percent of my users. I want to keep the link juice received from this page but I do not want it to appear in the SERPs.
Technical SEO | | surveygizmo0 -
Internal vs external blog and best way to set up
I have a client that has two domians registered - one uses www.keywordaustralia.com the other uses www.keywordaelaide.com He had already bought and used the first domain when he came to me I suggested the second as being worth buying as going for a more local keyword would be more appropriate. Now I have suggested to him that a blog would be a worthy use of the second domain and a way to build links to his site - however I am reading that as all links will be from the same site it wont be worth much in the long run and an internal blog is better as it means updated content on his site. should i use the second domain for blog, or just 301 the second domain to his first domain. Or is it viable to use the second domain as the blog and just set up an rss feed on his page ? Is there a way to have the second domain somehow 'linked' to his first domain with the blog so that google sees them as connected ? NOOBIE o_0
Technical SEO | | mamacassi0