Duplicate content - how to diagnose duplicate content from another domain before publishing pages?
-
Hi,
My company is having new distributor contract, and we are starting to sell products on our own webshop.
Bio-technology is an industry in question and over 1.000 products. Writing product description from scratch would take many hours. The plan is to re-write it.
With permission from our contractors we will import their 'product description' on our webshop. But, I am concerned being penalies from Google for duplicate content.
If we re-write it we should be fine i guess. But, how can we be sure? Is there any good tool for comparing only text (because i don't want to publish the pages to compare URLs)?
What else should we be aware off beside checking 'product description' for duplicate content?
Duplicate content is big issue for all of us, i hope this answers will be helpful for many of us.
Keep it hard work and thank you very much for your answers,
Cheers,
Dusan
-
Thank you again Monica. The reviews are definitely be implemented. Good luck!
-
I think you should stay above 90% unique.
I would strongly encourage you to make sure you add user generated content to the pages. Even rewriting the content to be exactly unique will not be enough to guarantee your pages can rank. The content is going to be written in unique words, but it will essentially be the same thing. You will need something that is uniquely valuable to the user.
No you won't necessarily be penalized, but, it will become harder to rank in branded searches.
-
Thank you Monica, really good answer!
Now, with copyscape.com what is the percentage of uniqueness that is 'allright'?
We are selling laboratory products for cell analysis, tittles and descriptions of our products are highly scientific. Manufacturer is the only one who can write it. Re-writing is our only option.
However, i am not that scared of being penalies any more. Thank you very much!
-
I would add that if you are going to have user generated content, make sure there's a review process so it doesn't get spammed/abused.
-
Monica said basically everything I would and probably a little better. The review system is extremely helpful in generating unique content, for a few reasons. One, you don't have to write it yourself (just review it), second customers want to hear from other customers, third the way a user describes a product may actually attract keyword opportunities you may not have thought of on your own.
-
Hi Dusan,
I have a couple of suggestions for you. The first, to answer your question, is that copyscape.com will let you compare two pieces of content for free, and then it will tell you the percentage of uniqueness. This will be a good way to tell if your rewrites are adequate.
My second suggestion is to implement a review system that will stream user generated content onto your pages. The duplicate content "penalty" you are referring to is really not a penalty in the empirical sense. When it comes to ranking, Google will look at two sites with the same content and pick one to display. Usually the branded site would win in a branded search. There are many other factors, like page rank and domain authority that can influence which page is displayed, and the user query can influence that as well.
Having uniquely valuable content on your site, like user generated comments and reviews, can be the offset for your duplicate content issue. Rewriting the manufacturer's descriptions isn't really going to accomplish the goal of offering the searcher something they can't find anywhere else. In my opinion, just rewriting the content isn't enough of an advantage to beat out other sites. You have to offer something valuable, that can only be found on your site. User generated content is (in my opinion) the best content you can have. Every consumer reads reviews when they are available. They want to find out what regular people have to say about a product. Is is the right size, is the color consistent, how long did it last, is this a fair price? These are all questions that can be answered in a review system.
I added reviews to my Ecommerce site about 6 months ago and have seen great success. 80% of my content is the same as 4 other sites, except my category pages. I write extremely unique content for those pages, which helps me target long tail and branded key terms. Then my product pages have the manufacturer's descriptions, tech specs and warranty info plus the user generated content. It has been very successful whenever I have implemented it.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
301 redirect to avoid duplicate content penalty
I have two websites with identical content. Haya and ethnic Both websites have similar products. I would like to get rid of ethniccode I have already started to de-index ethniccode. My question is, Will I get any SEO benefit or Will it be harmful if I 301 direct the below only URL’s https://www.ethniccode/salwar-kameez -> https://www.hayacreations/collections/salwar-kameez https://www.ethniccode/salwar-kameez/anarkali-suits - > https://www.hayacreations/collections/anarkali-suits
Intermediate & Advanced SEO | | riyaaaz0 -
Does Google View "SRC", "HREF", TITLE and Alt tags as Duplicate Content on Home Page Slider?
Greetings MOZ Community. A keyword matrix was developed by my SEO firm. I am in the process of integrating primary, secondary and terciary phrases into the text and am also sprinkling three or four other terms. Using a keyword density tool (http://www.webconfs.com/keyword-density-checker.php) the results were somewhat unexpected after I optimized. So I then looked at the source code and noticed text from HREF, ALT and SRC tags that may be effecting how Google would interpret text on the page. Our home page (www.nyc-officespace-leader.com) contains a slider with commercial real estate listings. Would Google index the SRC, HREF, TITLE and ALT tags in these slider items? Would this be detrimental to SEO? The code for one listing (and there are 7-8 in the slider) looks like this: | href="http://www.nyc-officespace-leader.com/listings/305-fifth-avenue-office-suite-1340sf" title="Lease a Prestigious Fifth Avenue Office - Manhattan, New York">Class A Fifth Avenue Offices class="blockLeft"><a< p=""></a<> href="http://www.nyc-officespace-leader.com/listings/305-fifth-avenue-office-suite-1340sf" title="Lease a Prestigious Fifth Avenue Office - Manhattan, New York"> src="http://dr0nu3l9a17ym.cloudfront.net/wp-content/uploads/fsrep/houses/125x100/305.jpg" alt="Lease a Prestigious Fifth Avenue Office - Manhattan, New York" width="125" height="94" /> 1,340 Sq. Ft. $5,918 / month Fifth Avenue Midtown / Grand Central <a< p=""></a<> | Could the repetition of the title text ("lease a Prestigious Fifth...") trigger a duplicate content penalty? Should the slider content be blocked or set to no-index by some kind of a Java script? We have worked very hard to optimize the home page so it would be a real shame if through some technical oversight we got hit by a Google Panda penalty. Thanks, Alan Thanks
Intermediate & Advanced SEO | | Kingalan10 -
Duplicate content on subdomains
Hi All, The structure of the main website goes by http://abc.com/state/city/publication - We have a partnership with public libraries to give local users access to the publication content for free. We have over 100 subdomains (each for an specific library) that have duplicate content issues with the root domain, Most subdomains have very high page authority (the main public library and other local .gov websites have links to this subdomains).Currently this subdomains are not index due to the robots text file excluding bots from crawling. I am in the process of setting canonical tags on each subdomain and open the robots text file. Should I set the canonical tag on each subdomain (homepage) to the root domain version or to the specific city within the root domain? Example 1:
Intermediate & Advanced SEO | | NewspaperArchive
Option 1: http://covina.abc.com/ = Canonical Tag = http://abc.com/us/california/covina/
Option 2: http://covina.abc.com/ = Canonical Tag = http://abc.com/ Example 2:
Option 1: http://galveston.abc.com/ = Canonical Tag = http://abc.com/us/texas/galveston/
Option 2: http://galveston.abc.com = Canonical Tag = http://abc.com/ Example 3:
Option 1: http://hutchnews.abc.com/ = Canonical Tag = http://abc.com/us/kansas/hutchinson/
Option 2: http://hutchnews.abc.com/ = Canonical Tag = http://abc.com/ I believe it makes more sense to set the canonical tag to the corresponding city (option 1), but wondering if setting the canonical tag to the root domain will pass "some link juice" to the root domain and it will be more beneficial. Thanks!0 -
Should we move a strong category page, or the whole domain to new domain?
We are debating moving a strong category page (and subcategory, product pages) from our current older domain to a new domain vs just moving the whole domain. The older domain has DA 40+, and the category page has PA 40+. Anyone with experience on how much PR etc will get passed to a virgin domain if we just redirect olddomain/strongcategorypage/ to newdomain.com? If the answer is little to none, we might consider just moving the whole site since the other categories are not that strong anyway. We will use 301 approach either way. Thanks!
Intermediate & Advanced SEO | | Durand0 -
Frequent FAQs vs duplicate content
It would be helpful for our visitors if we were to include an expandable list of FAQs on most pages. Each section would have its own list of FAQs specific to that section, but all the pages in that section would have the same text. It occurred to me that Google might view this as a duplicate content issue. Each page _does _have a lot of unique text, but underneath we would have lots of of text repeated throughout the site. Should I be concerned? I guess I could always load these by AJAX after page load if might penalize us.
Intermediate & Advanced SEO | | boxcarpress0 -
Category Content Duplication
Does indexing category archive page for a blog cause duplications? http://www.seomoz.org/blog/setup-wordpress-for-seo-success After reading this article I am unsure.
Intermediate & Advanced SEO | | SEODinosaur0 -
What is the best way to allow content to be used on other sites for syndication without taking the chance of duplicate content filters
Cookstr appears to be syndicating content to shape.com and mensfitness.com a) They integrate their data into partner sites with an attribution back to their site and skinned it with the partners look. b) they link the image back to their image hosted on cookstr c) The page does not have microformats or as much data as their own page does so their own page is better SEO. Is this the best strategy or is there something better they could be doing to safely allow others to use our content, we don't want to share the content if we're going to get hit for a duplicate content filter or have another site out rank us with our own data. Thanks for your help in advance! their original content page: http://www.cookstr.com/recipes/sauteacuteed-escarole-with-pancetta their syndicated content pages: http://www.shape.com/healthy-eating/healthy-recipes/recipe/sauteacuteed-escarole-with-pancetta
Intermediate & Advanced SEO | | irvingw
http://www.mensfitness.com/nutrition/healthy-recipes/recipe/sauteacuteed-escarole-with-pancetta0 -
Pages with Little Content
I have a website that lists events in Dublin, Ireland. I want to provide a comprehensive number of listings but there are not enough hours in the day to provide a detailed (or even short) unique description for every event. At the moment I have some pages with little detail other than the event title and venue. Should I try and prevent Google from crawling/indexing these pages for fear of reducing the overall ranking of the site? At the moment I only link to these pages via the RSS feed. I could remove the pages entirely from my feed, but then that mean I remove information that might be useful to people following the events feed. Here is an example page with very little content
Intermediate & Advanced SEO | | andywozhere0