Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Unsolved Crawling only the Home of my website
-
Hello,
I don't understand why MOZ crawl only the homepage of our webiste https://www.modelos-de-curriculum.comWe add the website correctly, and we asked for crawling all the pages. But the tool find only the homepage. Why?
We are testing the tool before to suscribe. But we need to be sure that the tool is working for our website. If you can please help us.
-
@Azurius
Certainly, I understand your concern about MOZ only crawling the homepage of your website despite adding it correctly and requesting a full crawl of all pages. It's frustrating when testing a tool before subscribing, and not getting the expected results can raise doubts about its functionality. To address this issue, I recommend double-checking your website's settings within the MOZ platform. Ensure that the website URL provided matches the actual structure of your site and that there are no typos or errors in the input. Additionally, review MOZ's documentation or contact their customer support to confirm if there are specific settings or configurations needed for a comprehensive crawl. Sometimes, minor adjustments in the tool's settings or the website's structure can make a significant difference in the crawling process. I hope this helps, and I wish you success in resolving this matter and making an informed decision about your subscription. -
Crawling only the home page of your website is a common practice in web indexing and search engine optimization. This approach allows search engine bots to focus on your site's main landing page, ensuring that it's properly indexed and ranked. Here's how you can specify that only the home page should be crawled:
Robots.txt File,
Canonical Tag
Sitemap.xml,
Noindex Tags,
Meta Robots TagBy implementing these methods, you can direct search engine crawlers to focus primarily on your website's home page, ensuring that it receives the most attention in terms of indexing and ranking. This can be particularly useful if your home page is the most important and relevant page for your website's SEO strategy.
https://www.clippingnext.com/ -
@Azurius i have the same issue,. I think the answer here is quite helpful
-
Hello,
There could be several reasons why MOZ is only crawling the homepage of your website, https://www.modelos-de-curriculum.com. Here are a few possibilities:
Robots.txt file: Check your website's robots.txt file to ensure that it's not blocking MOZ's web crawlers from accessing other pages. Make sure there are no disallow rules that could restrict access to certain areas of your site.
Nofollow tags: Ensure that your website doesn't have "nofollow" tags on internal links that may be preventing MOZ from following and crawling those links.
JavaScript: If your website heavily relies on JavaScript for content rendering, MOZ may face difficulty crawling the content. Ensure that important content is accessible without JavaScript.
Canonical tags: Check for canonical tags in your website's HTML. If you have specified the homepage as the canonical page for all other pages, this could limit MOZ's ability to crawl additional pages.
Site structure: MOZ may have trouble crawling pages with complex or unconventional site structures. Ensure that your website follows standard navigation and linking practices.
Crawl settings: Double-check the settings in your MOZ account to confirm that you have requested a full site crawl, not just the homepage.
If you've verified all these aspects and MOZ is still not crawling your website correctly, you may want to reach out to their support team for further assistance. They can provide more specific guidance based on your account and settings.
Testing the tool before subscribing is a wise approach to ensure it meets your needs. I recommend contacting MOZ's support for personalized assistance in resolving the crawling issue with your website. (PMP Exam Prep) (Canada PR) (Study abroad) (Project Management Life Cycle)
-
Hello,
The reason Moz is only crawling the homepage of your website could be due to various factors. Here are a few possibilities:
Robots.txt File: Check your website's robots.txt file to ensure that it doesn't block search engine crawlers from accessing specific pages.
Meta Robots Tags: Make sure there are no "noindex" meta tags on your internal pages that might prevent them from being indexed.
Crawl Restrictions: Moz may not have had enough time to crawl all the pages on your website. It might take some time for the tool to explore and index your entire site.
Sitemap: Ensure that your website's sitemap is correctly submitted to Moz. A sitemap helps search engines find and index all the pages on your site.
Internal Linking: Check if there are internal links from your homepage to other pages. A lack of internal links can make it harder for search engines to discover and crawl your site.
Access Permissions: Make sure there are no access restrictions or password protection on certain pages.
To resolve this issue and ensure Moz crawls all the pages of your website, consider checking and addressing these factors. If the problem persists, it might be helpful to reach out to Moz's customer support for specific assistance with your subscription trial.
(Canada PR) (Study abroad) (PMP Exam Prep) (Reference Letter for Canada PR) -
check your .htaccess if there is something that is blocking crawlers.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Crawler was not able to access the robots.txt
I'm trying to setup a campaign for jessicamoraninteriors.com and I keep getting messages that Moz can't crawl the site because it can't access the robots.txt. Not sure why, other crawlers don't seem to have a problem and I can access the robots.txt file from my browser. For some additional info, it's a SquareSpace site and my DNS is handled through Cloudflare. Here's the contents of my robots.txt file: # Squarespace Robots Txt User-agent: GPTBot User-agent: ChatGPT-User User-agent: CCBot User-agent: anthropic-ai User-agent: Google-Extended User-agent: FacebookBot User-agent: Claude-Web User-agent: cohere-ai User-agent: PerplexityBot User-agent: Applebot-Extended User-agent: AdsBot-Google User-agent: AdsBot-Google-Mobile User-agent: AdsBot-Google-Mobile-Apps User-agent: * Disallow: /config Disallow: /search Disallow: /account$ Disallow: /account/ Disallow: /commerce/digital-download/ Disallow: /api/ Allow: /api/ui-extensions/ Disallow: /static/ Disallow:/*?author=* Disallow:/*&author=* Disallow:/*?tag=* Disallow:/*&tag=* Disallow:/*?month=* Disallow:/*&month=* Disallow:/*?view=* Disallow:/*&view=* Disallow:/*?format=json Disallow:/*&format=json Disallow:/*?format=page-context Disallow:/*&format=page-context Disallow:/*?format=main-content Disallow:/*&format=main-content Disallow:/*?format=json-pretty Disallow:/*&format=json-pretty Disallow:/*?format=ical Disallow:/*&format=ical Disallow:/*?reversePaginate=* Disallow:/*&reversePaginate=* Any ideas?
Getting Started | | andrewrench0 -
Why MOZ just index some of the links?
hello everyone i've been using moz pro for a while and found a lot of backlink oppertunites as checking my competitor's backlink profile.
Link Building | | seogod123234
i'm doing the same way as my competitors but moz does not see and index lots of them, maybe just index 10% of them. though my backlinks are commenly from sites with +80 and +90 DA like Github, Pinterest, Tripadvisor and .... and the strange point is that 10% are almost from EDU sites with high DA. i go to EDU sites and place a comment and in lots of case, MOZ index them in just 2-3 days!! with maybe just 10 links like this, my DA is incresead from 15 to 19 in less than one month! so, how does this "SEO TOOL" work?? is there anyway to force it to crawl a page?0 -
Advise on the right way to block country specific users but not block Googlebot - and not be seen to be cloaking. Help please!
Hi, I am working on the SEO of an online gaming platform - a platform that can only be accessed by people in certain countries, where the games and content are legally allowed.
International SEO | | MarkCanning
Example: The games are not allowed in the USA, but they are allowed in Canada. Present Situation:
Presently when a user from the USA visits the site they get directed to a restricted location page with the following message: RESTRICTED LOCATION
Due to licensing restrictions, we can't currently offer our services in your location. We're working hard to expand our reach, so stay tuned for updates! Because USA visitors are blocked Google which primarily (but not always) crawls from the USA is also blocked, so the company webpages are not being crawled and indexed. Objective / What we want to achieve: The website will have multiple region and language locations. Some of these will exist as standalone websites and others will exist as folders on the domain. Examples below:
domain.com/en-ca [English Canada]
domain.com/fr-ca [french Canada]
domain.com/es-mx [spanish mexico]
domain.com/pt-br [portugese brazil]
domain.co.in/hi [hindi India] If a user from USA or another restricted location tries to access our site they should not have access but should get a restricted access message.
However we still want google to be able to access, crawl and index our pages. Can i suggest how do we do this without getting done for cloaking etc? Would this approach be ok? (please see below) We continue to work as the present situation is presently doing, showing visitors from the USA a restricted message.
However rather than redirecting these visitors to a restricted location page, we just black out the page and show them a floating message as if it were a model window.
While Googlebot would be allowed to visit and crawl the website. I have also read that it would be good to put paywall schema on each webpage to let Google know that we are not cloaking and its a restricted paid page. All public pages are accessible but only if the visitor is from a location that is not restricted Any feedback and direction that can be given would be greatly appreciated as i am new to this angle of SEO. Sincere thanks,0 -
GoogleBot still crawling HTTP/1.1 years after website moved to HTTP/2
Whole website moved to https://www. HTTP/2 version 3 years ago. When we review log files, it is clear that - for the home page - GoogleBot continues to only access via HTTP/1.1 protocol Robots file is correct (simply allowing all and referring to https://www. sitemap Sitemap is referencing https://www. pages including homepage Hosting provider has confirmed server is correctly configured to support HTTP/2 and provided evidence of accessing via HTTP/2 working 301 redirects set up for non-secure and non-www versions of website all to https://www. version Not using a CDN or proxy GSC reports home page as correctly indexed (with https://www. version canonicalised) but does still have the non-secure version of website as the referring page in the Discovery section. GSC also reports homepage as being crawled every day or so. Totally understand it can take time to update index, but we are at a complete loss to understand why GoogleBot continues to only go through HTTP/1.1 version not 2 Possibly related issue - and of course what is causing concern - is that new pages of site seem to index and perform well in SERP ... except home page. This never makes it to page 1 (other than for brand name) despite rating multiples higher in terms of content, speed etc than other pages which still get indexed in preference to home page. Any thoughts, further tests, ideas, direction or anything will be much appreciated!
Technical SEO | | AKCAC1 -
Unsolved Rogerbot blocked by cloudflare and not display full user agent string.
Hi, We're trying to get MOZ to crawl our site, but when we Create Your Campaign we get the error:
Moz Pro | | BB_NPG
Ooops. Our crawlers are unable to access that URL - please check to make sure it is correct. If the issue persists, check out this article for further help. robot.txt is fine and we actually see cloudflare is blocking it with block fight mode. We've added in some rules to allow rogerbot but these seem to be getting ignored. If we use a robot.txt test tool (https://technicalseo.com/tools/robots-txt/) with rogerbot as the user agent this get through fine and we can see our rule has allowed it. When viewing the cloudflare activity log (attached) it seems the Create Your Campaign is trying to crawl the site with the user agent as simply set as rogerbot 1.2 but the robot.txt testing tool uses the full user agent string rogerbot/1.0 (http://moz.com/help/pro/what-is-rogerbot-, rogerbot-crawler+shiny@moz.com) albeit it's version 1.0. So seems as if cloudflare doesn't like the simple user agent. So is it correct the when MOZ is trying to crawl the site it uses the simple string of just rogerbot 1.2 now ? Thanks
Ben Cloudflare activity log, showing differences in user agent strings
2022-07-01_13-05-59.png0 -
Unsolved Is Performance Metrics only available in a Campaign?
I'm looking to do a 1-off Performance Metrics analysis across dozens of pages on a single website - a prospective client. I thought it would be part of the On-Demand Crawl.
Moz Tools | | amandacash858960 -
How to block Rogerbot From Crawling UTM URLs
I am trying to block roger from crawling some UTM urls we have created, but having no luck. My robots.txt file looks like: User-agent: rogerbot Disallow: /?utm_source* This does not seem to be working. Any ideas?
Product Support | | Firestarter-SEO0 -
I have removed a subdomain from my main domain. We have stopped the subdomain completely. However the crawl still shows the error for that sub-domain. How to remove the same from crawl reports.
Earlier I had a forum as sub-domain and was mentioned in my main domain. However i have now discontinued the forum and have removed all the links and mention of the forum from my main domain. But the crawler still shows error for the sub-domain. How to make the crawler issues clean or delete the irrelevant crawl issues. I dont have the forum now and no links at the main site, bu still shows crawl errors for the forum which doesnt exist.
Product Support | | potterharry0