SEO Myth-Busters -- Isn't there a "duplicate content" penalty by another name here?
-
Where is that guy with the mustache in the funny hat and the geek when you truly need them?
So SEL (SearchEngineLand) said recently that there's no such thing as "duplicate content" penalties.
http://searchengineland.com/myth-duplicate-content-penalty-259657
by the way, I'd love to get Rand or Eric or others Mozzers aka TAGFEE'ers to weigh in here on this if possible.
The reason for this question is to double check a possible 'duplicate content" type penalty (possibly by another name?) that might accrue in the following situation.
1 - Assume a domain has a 30 Domain Authority (per OSE)
2 - The site on the current domain has about 100 pages - all hand coded. Things do very well in SEO because we designed it to do so.... The site is about 6 years in the current incarnation, with a very simple e-commerce cart (again basically hand coded). I will not name the site for obvious reasons.
3 - Business is good. We're upgrading to a new CMS. (hooray!) In doing so we are implementing categories and faceted search (with plans to try to keep the site to under 100 new "pages" using a combination of rel canonical and noindex. I will also not name the CMS for obvious reasons.
In simple terms, as the site is built out and launched in the next 60 - 90 days, and assume we have 500 products and 100 categories, that yields at least 50,000 pages - and with other aspects of the faceted search, it could create easily 10X that many pages.
4 - in ScreamingFrog tests of the DEV site, it is quite evident that there are many tens of thousands of unique urls that are basically the textbook illustration of a duplicate content nightmare. ScreamingFrog has also been known to crash while spidering, and we've discovered thousands of URLS of live sites using the same CMS.
There is no question that spiders are somehow triggering some sort of infinite page generation - and we can see that both on our DEV site as well as out in the wild (in Google's Supplemental Index).
5 - Since there is no "duplicate content penalty" and there never was - are there other risks here that are caused by infinite page generation?? Like burning up a theoretical "crawl budget" or having the bots miss pages or other negative consequences?
6 - Is it also possible that bumping a site that ranks well for 100 pages up to 10,000 pages or more might very well have a linkuice penalty as a result of all this (honest but inadvertent) duplicate content? In otherwords, is inbound linkjuice and ranking power essentially divided by the number of pages on a site? Sure, it may be some what mediated by internal page linkjuice, but what's are the actual big-dog issues here?
So has SEL's "duplicate content myth" truly been myth-busted in this particular situation?
???
Thanks a million!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Do the referring domains matter a lot in back-links? Google's stand?
Hi, It's a known fact about quality of back-links than quantity. Still domains are heavily different from links. Multiple domains are huge comparing to multiple links. Taking an average, how much does 'number of referring domains" boost website authority? I am not speaking about low quality domains, just number of domains including which are irrelevant to the topic or industry. Thanks
Algorithm Updates | | vtmoz0 -
I'm Pulling Hairs! - Duplicate Content Issue on 3 Sites
Hi, I'm an SEO intern trying to solve a duplicate content issue on three wine retailer sites. I have read up on the Moz Blog Posts and other helpful articles that were flooded with information on how to fix duplicate content. However, I have tried using canonical tags for duplicates and redirects for expiring pages on these sites and it hasn't fixed the duplicate content problem. My Moz report indicated that we have 1000s of duplicates content pages. I understand that it's a common problem among other e-commerce sites and the way we create landing pages and apply dynamic search results pages kind of conflicts with our SEO progress. Sometimes we'll create landing pages with the same URLs as an older landing page that expired. Unfortunately, I can't go around this problem since this is how customer marketing and recruitment manage their offers and landing pages. Would it be best to nofollow these expired pages or redirect them? Also I tried to use self-referencing canonical tags and canonical tags that point to the higher authority on search results pages and even though it worked for some pages on the site, it didn't work for a lot of the other search result pages. Is there something that we can do to these search result pages that will let google understand that these search results pages on our site are original pages? There are a lot of factors that I can't change and I'm kind of concerned that the three sites won't rank as well and also drive traffic that won't convert on the site. I understand that Google won't penalize your sites with duplicate content unless it's spammy. So If I can't fix these errors -- since the company I work conducts business where we won't ever run out of duplicate content -- Is it worth going on to other priorities in SEO like Keyword research, On/Off page optimization? Or should we really concentrate on fixing these technical issues before doing anything else? I'm curious to know what you think. Thanks!
Algorithm Updates | | drewstorys0 -
"Update" in Search Console is NOT an Algo Update
We've had a few questions about the line labeled "Update" in Google Search Console on the Search Analytics timeline graph (see attached image). Asking around the industry, there seems to be a fair amount of confusion about whether this indicates a Google algorithm update. This is not an algorithm update - it indicates an internal update in how Google is measuring search traffic. Your numbers before and after the update may look different, but this is because Google has essentially changed how they calculate your search traffic for reporting purposes. Your actual ranking and traffic have not changed due to these updates. The latest updated happened on April 27th and is described by Google on this page: Data anomalies in Search Console Given the historical connotations of "update" in reference to Google search, this is a poor choice of words and I've contacted the Webmaster Team about it. 2CsyN7Q
Algorithm Updates | | Dr-Pete12 -
Ecommerce - SEO Quick Wins?
Hi I wanted to find out if anyone had any quick wins for an ecommerce site & SEO. I am the only SEO and we have a small online team and an ecommerce site with thousands of product pages. It's impossible to optimise everything, and we have taken the top 100 products and optimised them - starting from scratch with keyword research. I'm now struggling to prioritize what we need next - I know we need better internal linking, content, social and lots more, but this isn't something I can get through alone. I need a starting point and perhaps something with a quick win initially? Thanks 🙂
Algorithm Updates | | BeckyKey0 -
Post penguin & panda update. what would be a good seo strategies for brand new sites
Hi there. I have the luxury of launching a few sites after the penguin and panda updates, so I can start from scratch and hopefully do it right. I will get SEO companies to help me with this so i just want to ask for advices on what would be a good strategies for a brand new site. my understand of the new updates is this content and user experience is important, like how long they spend, how many pages etc social media is important. we intent to engage FB and twitter alot. in New Zealand, not too many people use google+ so we will probbaly just concentrate on the first two hopefully we will try to get people to share our website via social media, apparent that is important should only concentrate on high quality backlinks with a good diverse set of alt tags, but concentrate on branding rather than keywords. Am i correct to say that so far? if that is the principle, what would be the strategy to implement these goals? Links to any articles would also be great please. Love learning. i just want to do this right and hopefully try to future proof the sites against updates as possible. i guess quality content and links will most likely to be safe. Thank you for your help.
Algorithm Updates | | btrinh0 -
Local SEO NAP - Two Different Cities....Same Zip Code
I've come across this recently and wanted to get your thoughts. I personally live in a city called Greenacres (yes, it's the place to be) but my zip code is also for Lake Worth. I'm a local SEO company so doing Local SEO stuff is pretty pointless (Google changed that in 2010) but I am sure other people have this issue for their business. Question, What do you do when your zip code is for two different cities. Do you try to make all NAPs (Name Address Phone Numbers) the same city. What if you cant'? Does having the NAP show up different cities hurt your efforts? etc. Obviously I think you'd try to keep the NAP as consistent as possible but what do you do if the citation source changes it or only uses the major of the two cities? There isn't a right or wrong answer (or maybe there is) but I wanted to get some thoughts on it. Darin.
Algorithm Updates | | DarinPirkey0 -
Domain Name search in google not appearing
My hcg domain doesn't show up in google search. Shows up in new sand image search. If I wrap the domain name in quotes it shows up
Algorithm Updates | | noork0 -
Need help with some duplicate content.
I have some duplicate content issues on my blog I'm trying to fix. I've read lots of different opinions online about the best way to correct it, but they all contradict each other. I was hoping I could ask this community and see what the consensus was. It looks like my category and page numbers are showing duplicate content. For instance when I run the report I see things like this: http://noahsdad.com/resources/ http://noahsdad.com/resources/page/2/ http://noahsdad.com/therapy/page/2/ I'm assuming that is just the categories that are being duplicated, since the page numbers only show on the report at the end of a category. What is the best way to correct this? I don't use tags at all on my blog, using categories instead. I also use the Yoast SEO plug in. I have a check mark in the box that disables tags. However it says, "If you're using categories as your only way of structure on your site, you would probably be better off when you prevent your tags from being indexed." There is a box that allows you to disable categories also, but the description above makes it seem like I don't want to block both tags and categories. Any ideas what I should do? Thanks.
Algorithm Updates | | NoahsDad0