Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Trying to find all internal links to a specific page (without index)
-
Hi guys --
Still waiting on Moz to index a page of mine. We launched a new site over two months ago.
In the meantime, I really just need a list of internal links to a specific page because I want to change its URL. Does anybody know how to find that list (of internal links to 1 of my pages) without the Moz index?
I appreciate the help!
-
@marchexmarketingmcc intelligible results from Google tools is never to be expected.
-
Hi! You can use Screaming Frog and Google Search Console, as it is mentioned in this article: https://rush-analytics.com/blog/find-all-pages-on-website
Also, they've mentioned their own tool, but I haven't tried it yet. -
You can make complete Crawl of your webpage with Screaming Frog. If the number of pages is not large. You can check the internal Links and export it. With Pivot tables you can identify all pages which link to a specific page.
-
If you have a new page that replaces an out of date one. Maybe you'd be better off putting a 301 redirect in place.?
-
Screaming Frog
-
My answer is outdated, now it works like this:
- Go to Google Search Console (former Webmaster Tools) and chose the site you want.
- On the left-hand menu, select LINKS, there you can also see your sites Internal links.
-
@sceorily Yes my answer is more than 4 years old...
now the link in the Google Search Console (former Webmaster Tools) is called LINKS. There you can also see your sites Internal links.
The asteriks were wrong formatting from my side I guess.
-
If you are wanting to find all internal links on a specific page on your website. You can follow these steps in Ahref webmaster tool:
- Enter your website URL in Site Explorer
- Then, Click on Internal Backlinks - Where you can see all internal backlinks of specific page of your website.
I hope this helps!
-
@cesare-marchetti Hi there. As Moz has reminded me just now, it's almost 2022. I'm afraid your instruction appears to be out of date. Google has moved the Search box, and I find no "Traffic" link. By the way, are the ** (asterisks) supposed to be included in the search string? I tried with and without, no intelligible results.
Then again, intelligible results from Google tools is never to be expected.
I'll keep trying. But this is like throwing darts blindfolded, and not even knowing which direction the dartboard is.
-
Download XENU - http://home.snafu.de/tilman/xenulink.html - run a crawl of your site and right click the properties of any page to see all internal links linking to that page.
-
Simple enough. Thanks Cesare.
-
Hi, you can use Search Console (Webmaster Tools) for that and see a list of internal links for your website as follows:
- On the Webmaster Tools Home page, click the site you want.
- On the left-hand menu, click **Search **Traffic, and then click Internal Links.
Cheers,
Cesare
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should search pages be indexed?
Hey guys, I've always believed that search pages should be no-indexed but now I'm wondering if there is an argument to index them? Appreciate any thoughts!
Technical SEO | | RebekahVP0 -
Investigating a huge spike in indexed pages
I've noticed an enormous spike in pages indexed through WMT in the last week. Now I know WMT can be a bit (OK, a lot) off base in its reporting but this was pretty hard to explain. See, we're in the middle of a huge campaign against dupe content and we've put a number of measures in place to fight it. For example: Implemented a strong canonicalization effort NOINDEX'd content we know to be duplicate programatically Are currently fixing true duplicate content issues through rewriting titles, desc etc. So I was pretty surprised to see the blow-up. Any ideas as to what else might cause such a counter intuitive trend? Has anyone else see Google do something that suddenly gloms onto a bunch of phantom pages?
Technical SEO | | farbeseo0 -
How to remove all sandbox test site link indexed by google?
When develop site, I have a test domain is sandbox.abc.com, this site contents are same as abc.com. But, now I search site:sandbox.abc.com and aware of content duplicate with main site abc.com My question is how to remove all this link from goolge. p/s: I have just add robots.txt to sandbox and disallow all pages. Thanks,
Technical SEO | | JohnHuynh0 -
Tool to search relative vs absolute internal links
I'm preparing for a site migration from a .co.uk to a .com and I want to ensure all internal links are updated to point to the new primary domain. What tool can I use to check internal links as some are relative and others are absolute so I need to update them all to relative.
Technical SEO | | Lindsay_D0 -
Footer Links with same anchor text on all pages
We have different websites targeted at the different services our company provides. (e.g. For our document storage services, we have www.ukdocumentstorage.com. For document management, we have www.document-management-solutions.co.uk). If we take the storage site for example, every single page has a link in the footer to our document management site, with the anchor text 'Cleardata Document Management' SEOMoz is telling me that these are seen as external links (as they are on a different URL's), and I'm just clarifying that would this be a major possible factor in the website not ranking highly? How should I rectify this issue?
Technical SEO | | janc0 -
De-indexing millions of pages - would this work?
Hi all, We run an e-commerce site with a catalogue of around 5 million products. Unfortunately, we have let Googlebot crawl and index tens of millions of search URLs, the majority of which are very thin of content or duplicates of other URLs. In short: we are in deep. Our bloated Google-index is hampering our real content to rank; Googlebot does not bother crawling our real content (product pages specifically) and hammers the life out of our servers. Since having Googlebot crawl and de-index tens of millions of old URLs would probably take years (?), my plan is this: 301 redirect all old SERP URLs to a new SERP URL. If new URL should not be indexed, add meta robots noindex tag on new URL. When it is evident that Google has indexed most "high quality" new URLs, robots.txt disallow crawling of old SERP URLs. Then directory style remove all old SERP URLs in GWT URL Removal Tool This would be an example of an old URL:
Technical SEO | | TalkInThePark
www.site.com/cgi-bin/weirdapplicationname.cgi?word=bmw&what=1.2&how=2 This would be an example of a new URL:
www.site.com/search?q=bmw&category=cars&color=blue I have to specific questions: Would Google both de-index the old URL and not index the new URL after 301 redirecting the old URL to the new URL (which is noindexed) as described in point 2 above? What risks are associated with removing tens of millions of URLs directory style in GWT URL Removal Tool? I have done this before but then I removed "only" some useless 50 000 "add to cart"-URLs.Google says themselves that you should not remove duplicate/thin content this way and that using this tool tools this way "may cause problems for your site". And yes, these tens of millions of SERP URLs is a result of a faceted navigation/search function let loose all to long.
And no, we cannot wait for Googlebot to crawl all these millions of URLs in order to discover the 301. By then we would be out of business. Best regards,
TalkInThePark0 -
What is the best way to find stranded pages?
I have a client that has a site that has had a number of people in charge of it. All of these people have very different opinions about what should be on the site itself. When I look at their website on the server I see pages that do not have any obvious navigation to them. What is the best way to find out the internal linking structure of a site and see if these pages truly are stranded?
Technical SEO | | anjonr0 -
Best free tool to check internal broken links
Question says it all I guess. What would your recommend as the best free tool to check internal broken links?
Technical SEO | | RikkiD225