Could we run into issues with duplicate content penalties if we were to borrow product descriptions?
-
Hello,
I work for an online retailer that has the opportunity to add a lot of SKUs to our site in a relatively short amount of time by borrowing content from another site (with their permission). There are a lot of positives for us to do this, but one big question we have is what the borrowed content will do to our search rankings (we normally write our own original content in house for a couple thousand SKUs). Organic search traffic brings in a significant chunk of our business and we definitely don't want to do something that would jeopardize our rankings.
Could we run into issues with duplicate content penalties if we were to use the borrowed product descriptions?
Is there a rule of thumb for what proportion of the site should be original content vs. duplicate content without running into issues with our search rankings?
Thank you for your help!
-
I think Alan and EGOL have summed it up nicely for you.
I have looked at a lot of Panda hit sites and one of the most common issues were e-commerce sites that consisted of primarily of stock product descriptions. Why would Google want to rank a site highly that just contains information that hundreds of other sites have?
If you've got a large chunk of your site containing duplicate descriptions like this then you can attract a Panda flag which can cause your whole site to not rank well, not just the product pages.
You could use the duplicate product descriptions if you had a large amount of original and helpful text around it. However, no one knows what the ratio is. If you have the ability to rewrite the product descriptions this is by far the best thing to do.
-
Just adding a point to this (and with reference to the other good points left by others) - Writing good product descriptions isn't actually that expensive!
It always seems it, as they are usually done in big batches. However on a per product basis they are pretty cheap. Do it well and you will not only improve the search results, but you can improve conversions and even make it more linkable.
Pick a product at random. Would it be worth a few £/$ to sell more of that item? If not remove it from the site anyway.
-
Adding a lot of SKUs to your site in a relatively short amount of time by borrowing content from another site sounds more like a bad sales pitch than a good "opportunity". If you don't want to put in jeopardy a significant chunk of your business, then simply drip the new sku's in as you get new content for them. The thin content's not likely to win you any new search traffic, so unless their addition is going to quickly increase sales from your existing traffic sources and quantities in dramatic fashion, why go down that road?
-
adding emphasis on the danger.
Duplicate product descriptions are the single most problematic issue ecommerce sites face from an SEO perspective. Not only are most canned descriptions so short as to cause product pages to be considered thin on content, copied/borrowed descriptions are more likely to be spread across countless sites.
While it may seem like an inordinate amount of time/cost, unique quality descriptions that are long enough to truly identify product pages as being worthy will go a long way to proving a site deserves ranking, trust.
-
You can hit Panda problems doing this. If you have lots of this content the rankings of your entire site could be damaged.
Best to write your own content, or use this content on pages that are not indexed until you have replaced with original content.
Or you could publish it to get in the index and replace as quickly as possible.
The site you are getting this content from could be damaged as well.
-
You definitely could run in to trouble here. Duplicate content of this type is meant to be dealt with on a page level basis. However if Google think it is manipulative then then it can impact on the domain as a whole. By "think" I really mean "if it matches certain patterns that manipulative sites use" - there is rarely an actual human review.
It is more complex than a simple percentage. Likely many factors are involved. However.. there is a solution!
You can simply add a no index tag to the product pages that have non-original content. That;ll keep them out of the index and keep you on the safe side of dupe issues.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google Images search traffic and image thumbnail issues
Hi MOZ community! Need a little help with a strange issue we are seeing of late on our project CareerAddict.com. We have seen a sudden and significant drop in image visibility in Search Console from the 27th August onwards. I understand that Google has been updating their filters and other bits in image search, so maybe this could have impacted us? I also noticed that the images which are mapped to our articles are not the full featured article 700px wide images which we provide to Google in the Structured Data. They are instead taking the OG share 450px wide images now on many occasions. You can see this by searching for "careeraddict.com" in images. Any insight or suggestions welcome on both of these. Interested to understand if any other webmasters are experiencing other or similar problems with image visibility in Google also. Thanks!
Algorithm Updates | | dqmedia0 -
Issue with Category Ranking on Page 1 vs. Homepage Ranking on Page 2
A client has a high-volume keyword that is rendering different results, whether it is on page one or page two of Google SERPs. If the keyword is on page one, ONLY the category page is ranking. When the keyword bumps off to page two, BOTH the category AND the homepage are ranking. This is happening on our IP and theirs, incognito and personalized searches. This has been happening since February. Any thought/insights would be greatly appreciated, thank you!!!!
Algorithm Updates | | accpar0 -
Duplicate Content
I was just using a program (copyscpape) to see if the content on a clients website has been copied. I was surprised that the content on the site was displaying 70% duplicated and it's showing the same content on a few sites with different % duplicated (ranging from 35%-80%) I have been informed that the content on the clients site is original and was written by the client. My question is, does Google know or understand that the clients website's content was created as original and that the other sites have copied it word-for-word and placed it on their site? Does he need to re-write the content to make it original? I just want to make sure before I told him to re-write all the content on the site? I'm well aware that duplicate content is bad, but i'm just curious if it's hurting the clients site because they originally created the content. Thanks for your input.
Algorithm Updates | | Kdruckenbrod0 -
Multiple products with legitimate duplicate descriptions
We are redeveloping a website for a card company who have far too many products to write unique descriptions for each. Even if they could I don't think it would be beneficial to the user. However they do have unique descriptions for each range which is useful for users viewing an individual card. Which is better practice: a) Ignore the duplicate content issue and supply the user with info about the range b) Provide clear enticing links to find out more about the range which will leave the individual card page a little void of content. Many thanks
Algorithm Updates | | SoundinTheory0 -
How to deal with EMD penalty?
After some research I would say my site have been hit by an EMD penalty, it seems many other people have faced the same. It would be very useful to know how to deal with this, many topics discuss the reasons behind it, but few have a realistic response for fast action. I am in two minds - one would be to try and improve content etc, but this is subjective and could take any amount of time, or never resolve the issue. The other would be to move the content to a new URL, which poses the question, should I do a 301 redirect, or would this just transfer the penalty? If no redirect, then I am proposing starting fresh - as the sites hit by EMD penalties are deemed 'low quality' this might be the fastest way to recovery. If I move the old content to my established main site as a sub folder, would this cause any problems? Many thanks for people responses.
Algorithm Updates | | Quime1 -
Rel="alternate" hreflang="x" or Unique Content?
Hi All, I have 3 sites; brand.com, brand.co.uk and brand.ca They all have the same content with very very minor changes. What's best practice; to use rel="alternate" hreflang="x" or to have unique content written for all of them. Just wondering after Panda, Penguin and the rest of the Zoo what is the best way to run multinational sites and achieve top positions for all of them in their individual countries. If you think it would better to have unique content for each of them, please let us know your reasons. Thanks!
Algorithm Updates | | Tug-Agency0 -
Host name per content
Hello everyone. I'm in charge of the website HispaZone.com in which apart from many other things we provide free program downloads in spanish in a similar way to softpedia, tucows, cnet, softonic and others. I'm not a great SEO but I try to do my best. Several months ago based on my most important competence (softonic.com and uptodown.com) I decided that I would give a host name under the domain hispazone.com for the landing page of each program download. For downloading Nero for example the landing page would be http://nero.hispazone.com and like this for the whole of our 800 program database. The thing is that after 5-6 months since that change and after many other improvements, the traffic coming from google to these downloads dropped dramatically. We thought it could have been related to Google Panda but we recently hired an SEO consultant and he says that it's because of not having the downloads under the same host name. That we lose the page authority and the link flow from the hostname http://www.hispazone.com. The SEO consultant seems to be great, very up to date with all new changes in google. We made many improvements thanks to him and I can say that I trust him with everything. But now comes the time for deciding if we move our program download landing pages back to the www.hispazone.com hostname. I would like some second opinion about this because the fact that the biggest ones in Spain like Softonic and Uptodown have a hostname for each program download when these companies invest really a lot in their SEO makes me be unsure of going back into having all under the same hostname. Thanks a lot.
Algorithm Updates | | HispaZone0 -
Large site with faceted navigation using rel=canonical, but Google still has issues
First off, I just wanted to mention I did post this on one other forum so I hope that is not completely against the rules here or anything. Just trying to get an idea from some of the pros at both sources. Hope this is received well. Now for the question..... "Googlebot found an extremely high number of URLs on your site:" Gotta love these messages in GWT. Anyway, I wanted to get some other opinions here so if anyone has experienced something similar or has any recommendations I would love to hear them. First off, the site is very large and utilizes faceted navigation to help visitors sift through results. I have implemented rel=canonical for many months now to have each page url that is created based on the faceted nav filters, push back to the main category page. However, I still get these damn messages from Google every month or so saying that they found too many pages on the site. My main concern obviously is wasting crawler time on all these pages that I am trying to do what they ask in these instances and tell them to ignore and find the content on page x. So at this point I am thinking about possibly using robots.txt file to handle these, but wanted to see what others around here thought before I dive into this arduous task. Plus I am a little ticked off that Google is not following a standard they helped bring to the table. Thanks for those who take the time to respond in advance.
Algorithm Updates | | PeteGregory0