Why is the exact same URL being seen as duplicate and showing an error in my SEO reports
-
Well, I am still having duplicate page issues.
I have a question about one of the errors SEO is giving me when I download a crawl report. I am going to attach a screen shot of part of the report so you can see for yourself, along with explaining it here.
SEO shows the list of URL's that it crawled in the report. In this(see attachment) portion of the report it has 321 results for the exact same URL. It also says all of these exact same URL's have received a 404 error. What I want to know is how does it make 321 results for the same URL? And with this error that I don't see when I look at the page?
-
Hey Josh,
You may want to speak with your developer on this one PHP is a server language and could be generating the unique pages and causing Roger to crawl twice depending on which way he approaches the page. I apologize for the delay here as I was not receiving notifications on this post.
-
I think it does so by PHP. Is there any easy way I can make sure?
-
Hey Josh,
Rogerbot discovered this link through the /blog/ subfolder on your page which led him to the which battery post. By any chance does the auto forward do so by PHP or Javascript? Sometimes Roger can get a little hung up on these pages and think they don't exist.
Let me know either way so we can get this taken care of!
Thanks
-
Ok I get which page it is now but I do not understand the issue. There isnt a direct link to that blog post from any of the other pages crawled. And if you use either link they both work. The Column a link however does auto forward to the MaxAmps site equivalent page. Could it somehow have to do with this?
-
Hey Josh,
This is James from Moz Help. I'd like to see if I can assist in diagnosing this problem. I had a chance to pull your crawl .CSV and the 321 pages you have in column F are actually the referring pages the actual page we crawled will be in Column A. Essentially the What Battery Can we MaxAmps Build for You has 321 unique pages it links too. While this page does not 404 it seems Roger can not get to the pages after this.
Feel free to follow up here or send us a message to help@moz.com
Have a great day!
-
Okay I will be sure to try and implement this to the blog.
You seem to know how a php dynamic site builds itself from one page. May I ask you another question? We have been going through this duplicate page problem for a long time. I generally get the answer to redirect or canonical the pages. I understand how I can do this to our blog as it is in word press. However, the rest of our site is not. The category section of our page, for instance, generates depending on how the person gets there from one single page of coding. And that is for all of our categories, not just one. How can I implement a redirect or canonical to this type of site. I would not want all of the categories to lead to one particular category. So if I put the canonical tag in my category page with a single url than that won't work. Also if I use a redirct in this page it will still lead me to only one category (correct?) instead of the people having the option to go to several different categories.
-
Hi Josh,
By chance does this url have parameters. Those may not be reporting properly in this report. That would be my first thought. I have seen that frequently in blog / forum crawls, as those usually have many parameters for starting at a certain post number. The simple solution is to just rel canonical the page to its root.
As for the 404 error I would guess that some page is generated or linking to urls with certain parameters that the page itself doesn't know how to handle.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How To Stop Moz Crawl From Prepending /blog/ on all our site urls that it crawls
Hello, At some time in the past our WP site had urls like this: www.oursite.com/blog/post-title-pretty-link The site has not used that url structure for quite some time, but Moz crawl is still hitting every post with /blog/prepended and as a result is generating thousands of 404s. When the /blog/ is removed from the url, then the urls work fine. Where are those old urls being stored and how can we update them? How do we address this issue? Any assistance will be appreciated. Thanks!
Moz Bar | | dbcooper1 -
Keyword Ranking Report is Different than Real Result on Google
Hi Guys,
Moz Bar | | KanikaG
I am getting MOZ reports on a few keywords and that is good But when My client search for the same keywords he doesn't get the same rank, not even on that page. Let me explain in more details:
I have a keyword that shows on rank #5 in Google Mobile for a Local location for the current week. The client is from the same location as well. But when client search for that keyword from his mobile he doesn't get the keyword listed on that page, not even on other pages. I checked with other online keyword rank checker tool and all are showing the result similar to the Moz ranking report. It's really frustrating for my client. Google analytics shows that mobile users and organic reach are increasing so I am sure MOZ and other tools are showing us correct result. What could be the possible reason that he is not getting the same result as showing by MOZ? I am not from his location so how can I check the keyword result for the selected city using MOZ that is accurate? Any help would be highly appreciated. Please help. Thanks1 -
On-Page Grader Url is inaccessible
Hi everybody. I'm trying to use on -page grader for https://www.upscaledinnerclub.com and get "Sorry, but that URL is inaccessible." Robots.txt are empty, another thread on MOZ was talking about DNS check - it's all good. So, I can't figure out why this is happening. Also I am trying the same for another website https://www.regexseo.com - the same story. Common thing is that they both are on Google App Engine. And at first i thought that was the problem. Bu then i checked this one : https://www.logitinc.com/ and it's working, even though this website is on GAE as well. None of these website have robots.txt or any differences in setup or settings. Any thoughts?
Moz Bar | | DmitriiK0 -
Learn how to use MozBar to analyze competitors' schema markup. Get you Daily SEO Fix!
Schema markup helps Google and Bing identify what your (and your competitors') website pages are all about and as a result, helps search engines better index them. Better indexing can lead to better rankings for relevant keywords and phrases and herein lies the opportunity. In today's Daily SEO FiX, Jordan shows you how to use MozBar to analyze the schema markup of the competition and optimize your own site and pages for rich snippets. If you don't have MozBar, no worries! You may download it for free here. This video is part of The Moz Daily SEO Fix tutorial series--Moz tool tips and tricks in under 2 minutes. To watch all of our videos so far, and to subscribe to future ones, make sure to visit the Daily SEO Fix channel on YouTube.
Moz Bar | | kellyjcoop1 -
Dashboard module in Custom report
Hello Earth I'm putting a custom report together for higher management and have added the dashboard module as a good starting point. However, all I want from it is the overall visit & visit break down and the domain authority data, all the 'Mentions Found' 'On Page Wins' 'Ranking Highlights' info is just making the report over long and less to the point. Is there anyway of taking some of the dashboard into the report instead of the whole kaboodle? Or is there another module that will give me total visits not just organic? Many Thanks in advance Chris
Moz Bar | | ChrisParry0 -
How can I find the old ERRORS and WARNINGS report in the NEW Moz design?
I'm looking for a complete list of errors and bugs that need to be fixed within a website. I used to use the MAIN tool (at least it seemed it was the most popular) but now that its just MOZ.com I can't seem to find that great report. It had data such as: 1. List of pages with Title Tags too long 2. List of pages with Description Tags too long 3. List of RED errors and YELLOW warnings, BLUE somethings... etc... Ring a bell? I LOVED this report, where can I find this data? Thanks! Derek
Moz Bar | | DerekM42420 -
Duplicate content errors
Hi I am getting some errors for duplicate content errors in my crawl report for some of our products www.....com/brand/productname1.html www.....com/section/productname1.html www.....com/productname1.html we have canonical in the header for all three pages rel="canonical" href="www....com/productname1.html" />
Moz Bar | | phes0 -
Ajax #! URL support?
Hi Moz, My site is currently following the convention outlined here: https://support.google.com/webmasters/answer/174992?hl=en Basically since pages are generated via Ajax we are setup to direct bots that replace the #! in a url with ?escaped_fragment to cached versions of the ajax generated content. For example, if the bot sees this url: http://www.discoverymap.com/#!/California/Map-of-Carmel/73 it will replace it will instead access the page: http://www.discoverymap.com/?escaped_fragment=/California/Map-of-Carmel/73 In which case my server serves the cached html instead of the live page. This is all per Googles direction and is indexing fine. However the MOZ bot does not do this. It seems like a fairly straight-forward feature to support. Rather than ignoring the hash, you look to see if it is a #! and then try to spider the url replaced with ?escaped_fragment. Our server does the rest. If this is something MOZ plans on supporting in the future I would love to know. If there is other information that would be great. Also, pushstate is not practical for everyone due to limited browser support, etc. Thanks, Dustin Updates: I am editing my question because it won't let me respond to my own question. It says I need to sign up for MOZ analytics. I was signed up for Moz Analytics?! Now I am not? I responded to my invitation weeks ago? Anyway, you are misunderstanding how this process works. There is no site-map involved. The bot reads this URL on the page: http://www.discoverymap.com/#!/California/Map-of-Carmel/73 And when it is ready to spider the page for content it, it spider's this URL instead: http://www.discoverymap.com/?escaped_fragment=/California/Map-of-Carmel/73 The server does the rest, it is simply telling Roger to recognize the #! format and replace it with ?escaped_fragment Though I obviously do not know how Roger is coded but it is a simple string replacement. Thanks.
Moz Bar | | oneactlife0