Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How can I make a list of all URLs indexed by Google?
-
I have a large site with over 6000 pages indexed but only 600 actual pages and need to clean up with 301 redirects. Haven't had this need since Google stopped displaying the url's in the results.
-
very depressed of being not indexing URL - https://www.wrightleatherworks.com/collections/gun-holsters-for-women please someone help
-
Open the sitemap in a text editor or XML viewer. Look at the <loc> tags in the sitemap file. These tags contain the URL of each page on the website. You can copy the URLs from the <loc> tags into a spreadsheet or text document.
-
Open the sitemap in a text editor or XML viewer. Look at the <loc> tags in the sitemap file. These tags contain the URL of each page on the website. You can copy the URLs from the <loc> tags into a spreadsheet or text document.
-
@aplusnetsolutions create a seperate sitemap for not indexed pages and upload to google web master tools. i had a website had same issues. Also remember it will take some time also
-
@aplusnetsolutions
Make sure all these things are selected in screaming frog and -
Going to Google search console it will tell you how many inches are inside Google search index there are a few other ways to do this if you want
#1
By far the easiest is to going to search console and select coverage and all known pages
number 2
For the quick and dirty method, simply perform a simple site search in your Google search bar with “site:yourdomain.com”
#3
You can crawl your site using screaming frog (it will 500 pages for free but every SEO should have it) use it to index your site and connect it to Google search console API as well as Google analytics and you will have a easy-to-read CSV export of your index and non-index pages
more on how this is done![ you can see the non-indexable URLs on the the right click that]( image url)
video https://youtu.be/iYeXSdUt_hg
#4
or indexer site with any thing that can export a CSV download the CSV uploaded here and it will tell you which pages are in Google's index
https://www.rankwatch.com/free-tools/google-index-checkerhttps://searchengineland.com/check-urls-indexed-without-upsetting-google-follow-267472
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best redirect destination for 18k highly-linked pages
Technical SEO question regarding redirects; I appreciate any insights on best way to handle. Situation: We're decommissioning several major content sections on a website, comprising ~18k webpages. This is a well established site (10+ years) and many of the pages within these sections have high-quality inbound links from .orgs and .edus. Challenge: We're trying to determine the best place to redirect these 18k pages. For user experience, we believe best option is the homepage, which has a statement about the changes to the site and links to the most important remaining sections of the site. It's also the most important page on site, so the bolster of 301 redirected links doesn't seem bad. However, someone on our team is concerned that that many new redirected pages and links going to our homepage will trigger a negative SEO flag for the homepage, and recommends instead that they all go to our custom 404 page (which also includes links to important remaining sections). What's the right approach here to preserve remaining SEO value of these soon-to-be-redirected pages without triggering Google penalties?
Technical SEO | | davidvogel0 -
How can I check my website is not in spam?
I have a blogging website where I post about famous food, home remedies, and more. When I started my website's keywords were ranking on Google But Now a single keyword is not in the ranking list. That's why I have concerns about how I can fix it.
SEO Tactics | | worldviajar.com0 -
Google search console 380,000 spam backlinks
Hi guys, I recently suffered a major negative seo attack against my site, with more than 380K spam backlinks using more than 5k domains. Because of this, I'm having serious problems tracking my site's statistics in GSC due to the limit of only 1000 query lines. Please, I need help on how I can get access to all these 5,000 domains in the search console so I can create a disavow list. Any tips on how to clean this up?
SEO Tactics | | xurupita0 -
google webmaster tools Indexing request rejected
when i try to index my posts in google webmaster tools i see this eror : Indexing request rejected
SEO Tactics | | sasansasyino
During live testing, indexing issues were detected with the URL
Crawl
Time
Sep 23, 2023, 11:05:05 PM
Crawled as
Google Inspection Tool desktop
Crawl allowed?
Yes
Page fetch
error
Failed: Hostload exceeded
Indexing allowed?
N/A
Indexing
User-declared canonical
N/A
Google-selected canonical
Only determined after indexing my website : http://123select.ir/0 -
Unsolved Using NoIndex Tag instead of 410 Gone Code on Discontinued products?
Hello everyone, I am very new to SEO and I wanted to get some input & second opinions on a workaround I am planning to implement on our Shopify store. Any suggestions, thoughts, or insight you have are welcome & appreciated! For those who aren't aware, Shopify as a platform doesn't allow us to send a 410 Gone Code/Error under any circumstance. When you delete or archive a product/page, it becomes unavailable on the storefront. Unfortunately, the only thing Shopify natively allows me to do is set up a 301 redirect. So when we are forced to discontinue a product, customers currently get a 404 error when trying to go to that old URL. My planned workaround is to automatically detect when a product has been discontinued and add the NoIndex meta tag to the product page. The product page will stay up but be unavailable for purchase. I am also adjusting the LD+JSON to list the products availability as Discontinued instead of InStock/OutOfStock.
Technical SEO | | BakeryTech
Then I let the page sit for a few months so that crawlers have a chance to recrawl and remove the page from their indexes. I think that is how that works?
Once 3 or 6 months have passed, I plan on archiving the product followed by setting up a 301 redirect pointing to our internal search results page. The redirect will send the to search with a query aimed towards similar products. That should prevent people with open tabs, bookmarks and direct links to that page from receiving a 404 error. I do have Google Search Console setup and integrated with our site, but manually telling google to remove a page obviously only impacts their index. Will this work the way I think it will?
Will search engines remove the page from their indexes if I add the NoIndex meta tag after they have already been index?
Is there a better way I should implement this? P.S. For those wondering why I am not disallowing the page URL to the Robots.txt, Shopify won't allow me to call collection or product data from within the template that assembles the Robots.txt. So I can't automatically add product URLs to the list.0 -
Multiple URLs from One Domain on Page 1
Is it possible (and if it is possible, how difficult is it) to get multiple URLs from the same domain name to be indexed on the first page of Google for the same keyword or phrase? Assuming each URL has significantly different content of course... Context: A large real estate brand has 6 franchise offices in the same city. Each office has its own listing/landing page on the parent brand website. Each franchise owner wants their page on the website to rank for the term 'Christchurch real estate'. The homepage of the parent brand website currently ranks on the first page of Google for this search term, but none of the franchisee pages do. So my question is: with the right unique content on each franchisee page, supported by quality backlinks to each of the different offices, is it possible to get multiple franchisees listed on the first page of Google given that ultimately the 6 URLs are all attached to the same domain name? (And, if so, do you have any hot tips you can share to assist me on this uphill battle?)
SEO Tactics | | BeKonstructive0 -
What is the difference between "document" and "object" moved redirect errors?
What is the difference between "document" and "object" moved redirect errors? I'm used to see "object moved" as a redirect chain issue that needs to be fixed, but this week my report contained a "document moved" redirect chain issue. And it's on our homepage. Looks like it might be a HTTP versus an HTTPS issue.
Reporting & Analytics | | Kate_Nadeau0 -
Solve Redirect Chains
I've gotten a few Redirect Issues that involve Redirect Chains, with the https:// version redirecting to the www. version and then redirecting to the right URL. Here is an example:
On-Page Optimization | | Billywig
Schermafbeelding 2021-12-07 om 11.04.32.png I've tried setting a direct redirect between the first and the last URL, but WordPress doesn't seem to allow that (it's overwritten). I've also tried checking the internal links to make sure that none of the links are the first one. They don't seem to be there. Does anyone have any tips on solving these Redirect Chains?0