Duplicate Page Content Errors on Moz Crawl Report
-
Hi All,
I seem to be losing a 'firefighting' battle with regards to various errors being reported on the Moz crawl report relating to;
- Duplicate Page Content
- Missing Page Title
- Missing Meta
- Duplicate Page Title
While I acknowledge that some of the errors are valid (and we are working through them), I find some of them difficult to understand...
Here is an example of a 'duplicate page content' error being reported;
http://www.bolsovercruiseclub.com (which is obviously our homepage)
Is reported to have 'duplicate page content' compared with the following pages;
http://www.bolsovercruiseclub.com/guides/gratuities
http://www.bolsovercruiseclub.com/about-us/meet-the-team/craig
All 3 of those pages are completely different hence my confusion...
This is just a solitary example, there are many more!
I would be most interested to hear what people's opinions are...
Many thanks
Andy
-
Hey Andy,
Thanks for reaching out! The Duplicate Content errors are actually aggregated using similar criteria as Google - that's a 95% similarity on the code level.
So even if something like the title is different, if the overall code is 95% the same they will be flagged and could potentially get penalized by the Search Engines. Using third party duplicate content checkers on the URLs you provided, it does look like all but one combo meet the 95% threshold. In that one case, we may have found them to be 95% similar at the time of the crawl, but they aren't any longer so that will be reflected in the next crawl.
To get more information on Duplicate Content, check out our Help Hub.
I hope that makes sense. Let me know if you have any other questions and have a great day!
-
here is a guess (jsut a guess though). I think this is happening due to your code and your side bars. If you CNTR+U on Chrome you will see that code wise everything is the same and your content on the 2 examples is very thin (100 words or so) thus it might see it as duplicate.
I wouldnt worry much for something like that though as "duplciate content" normally are pages that target the same product for a different location with people being lazy changing anything but a post code and a city name, or some errors in your database that automatically create similar pages with different url variations.
Hope this helps
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Mixing up languages on the same page + possible duplicate content
I have a site in English hosted under .com with English info, and then different versions of the site under subdirectories (/de/, /es/, etc.) Due to budget constraints we have only managed to translate the most important info of our product pages for the local domains. We feel however that displaying (on a clearly identified tab) the detailed product info in English may be of use for many users that can actually understand English, and may help us get more conversions to have that info. The problem is that this detailed product info is already used on the equivalent English page as well. This basically means 2 things: We are mixing languages on pages We have around 50% of duplicate content of these pages What do you think that the SEO implications of this are? By the way, proper Meta Titles and Meta Descriptions as well as implementation of href lang tag are in place.
Intermediate & Advanced SEO | | lauraseo0 -
Duplicate content issue
Hello! We have a lot of duplicate content issues on our website. Most of the pages with these issues are dictionary pages (about 1200 of them). They're not exactly duplicate, but they contain a different word with a translation, picture and audio pronunciation (example http://anglu24.lt/zodynas/a-suitcase-lagaminas). What's the better way of solving this? We probably shouldn't disallow dictionary pages in robots.txt, right? Thanks!
Intermediate & Advanced SEO | | jpuzakov0 -
After Server Migration - Crawling Gets slow and Dynamic Pages wherein Content changes are not getting Updated
Hello, I have just performed doing server migration 2 days back All's well with traffic moved to new servers But somehow - it seems that w.r.t previous host that on submitting a new article - it was getting indexed in minutes. Now even after submitting page for indexing - its taking bit of time in coming to Search Engines and some pages wherein content is daily updated - despite submitting for indexing - changes are not getting reflected Site name is - http://www.mycarhelpline.com Have checked in robots, meta tags, url structure - all remains well intact. No unknown errors reports through Google webmaster Could someone advise - is it normal - due to name server and ip address change and expect to correct it automatically or am i missing something Kindly advise in . Thanks
Intermediate & Advanced SEO | | Modi0 -
Category Pages For Distributing Authority But Not Creating Duplicate Content
I read this interesting moz guide: http://moz.com/learn/seo/robotstxt, which I think answered my question but I just want to make sure. I take it to mean that if I have category pages with nothing but duplicate content (lists of other pages (h1 title/on-page description and links to same) and that I still want the category pages to distribute their link authority to the individual pages, then I should leave the category pages in the site map and meta noindex them, rather than robots.txt them. Is that correct? Again, don't want the category pages to index or have a duplicate content issue, but do want the category pages to be crawled enough to distribute their link authority to individual pages. Given the scope of the site (thousands of pages and hundreds of categories), I just want to make sure I have that right. Up until my recent efforts on this, some of the category pages have been robot.txt'd out and still in the site map, while others (with different url structure) have been in the sitemap, but not robots.txt'd out. Thanks! Best.. Mike
Intermediate & Advanced SEO | | 945010 -
Canonical and On-Page Report Card
Hello, One quick question about rel canonical. If i use SeoMoz amazing on-page optimization tool i get a grade B if i use www.mydomain.com and my keyword. I get a grade A if i use https://www.mydomain.com and same keyword. I get the grade B coz i don't get the check mark to "Appropriate Use of Rel Canonical" box. Should i use this rel canonical stuff if i am 301 redirecting www. version to https://www. version already. Regards, OÜInigo
Intermediate & Advanced SEO | | InigoOU0 -
Duplicate content
I run about 10 sites and most of them seemed to fall foul of the penguin update and even though I have never sought inorganic links I have been frantically searching for a link based answer since April. However since asking a question here I have been pointed in another direction by one of your contributors. It seems At least 6 of my sites have duplicate content issues. If you search Google for "We have selected nearly 200 pictures of short haircuts and hair styles in 16 galleries" which is the first bit of text from the site short-hairstyles.com about 30000 results appear. I don't know where they're from nor why anyone would want to do this. I presume its automated since there is so much of it. I have decided to redo the content. So I guess (hope) at some point in the future the duplicate nature will be flushed from Google's index? But how do I prevent it happening again? It's impractical to redo the content every month or so. For example if you search for "This facility is written in Flash® to use it you need to have Flash® installed." from another of my sites that I coincidently uploaded a new page to a couple of days ago, only the duplicate content shows up not my original site. So whoever is doing this is finding new stuff on my site and getting it indexed on google before even google sees it on my site! Thanks, Ian
Intermediate & Advanced SEO | | jwdl0 -
Get Duplicate Page content for same page with different extension ?
I have added a campaign like "Bannerbuzz" in SEOMOZ Pro account and before 2 or 3 days i got errors related to duplicate page content . they are showing me same page with different extension. As i mentioned below http://www.bannerbuzz.com/outdoor-vinyl-banners.html
Intermediate & Advanced SEO | | CommercePundit
&
http://www.bannerbuzz.com/outdoor_vinyl_banner.php We checked our whole source files but we didn't define php related urls in our source code. we want to catch only our .html related urls. so, Can you please guide us to solve this issue ? Thanks <colgroup><col width="857"></colgroup>
| http://www.bannerbuzz.com/outdoor-vinyl-banners.html |0 -
How much (%) of the content of a page is considered too much duplication?
Google is not fond of duplication, I have been very kindly told. So how much would you suggest is too much?
Intermediate & Advanced SEO | | simonberenyi0