Duplicate content - what to do?
-
Hi,
We have a whole lot of articles on our site. In total 5232 actually.
The web crawler tells me that in the articles we have a lot of duplicate content. Which is sort of nonsense, since each article is unique. Ah, some might have some common paragraphs because they are recurring news about a weekly competition.
But, an example:
http://www.betxpert.com/artikler/bookmakere/brandvarme-ailton-snupper-topscorerprisen
AND
http://www.betxpert.com/artikler/bookmakere/opdaterede-odds-pa-sportschef-situationen-pa-vestegnen
These are "duplicate content", however the two article texts are not the same. The menu, and the widgets are all the same, but highly relevant to the article.
So what should I do? How can i rid myself of these errors?
-Rasmus
-
So! Have now made some changes.
I have removed the large number of links at the bottom and a single widget on the right. It seems to have lowerede the level of noise on the page.
We brandish our articles with keywords (but mostly new ones since the feature is new) such that only relevant articles are linked (in a widget which appears when relevant articles are found).
This change should cut it, right?
-Rasmus
-
Rasmus,
Your description and keywords tags are exactly the same on both pages.
The description should be a summary of the content of each individual page
If you fix that, it may go some way towards fixing the problem.
If that doesn't help, then consider adding more text to those stories
- say a little more about each competitor, or at least the leader or the top 3.
Another thing you could consider is using a different template for that type of page, so you don't add all that stuff under the HR tag, and maybe drop off something from the right column, so the page is a little bit shorter and the story will be a larger percentage of the text.
You also have a lot of inline javascript and a lot of external javascripts.
I don't know if the bots or the ranking algorithms care, but it would be a lot nicer if you consolidated some of that.
-
Hey Rasmus,
The menu, widgets e.t.c all count as part of content as they are page elements, so ofcourse they would be considered.
This situation is abit tricky, however if you think you can merge them do so, maybe add slightly more content to each page; otherwise do nothing.
Just make sure and see whether both pages are ranking for the appropriate keywords.
Thanks,
Vahe
-
That is almost duplicate content. The % of the unique content on the two pages you've posted as example is small - only the actual content is different - the rest of the site wide boxes and boiler plates are the same.
So matematicly the difference in those two pages is very small.
The good thing is that you have the article in a good real estate location on the page - amnd that is good.
If your "score" as far as page views, Bounce rate, time on page, loyality is good then no need to worry about the crawler telling you about the duplicate issue. if the Bounce rate, time on page etc doesn't look good then yes, you nee dot worry about it.
My suggestion will be to cut out some or most of the boiler plate boxes in your site structure - it doesn't help you, it doesn't help your readers but it dose harm you with all those links per page as far as number (% of links vs text) and to be honest I am not 100% sure it helps your users with that aggressive approach: a lot of options.
Also there is a good approach not to use the same target for the link in the article as it doesn't counts - only the first one is taken in consideration as far as anchor text - as far as seo. If it has a good CTR then keep it but remember that even if only one counts as sending SEO signal (anchor text and link love) it dose count as far as number of links per page.
The bottom line is that if you cut about 50% of the boiler plates in those pages you will gain a lot of advantages for both the SEO side and you would also help your users (usability, navigation etc).
You don't have to take my advice for granted - do this for some of the articles and monitor ctr, serps for them and so on to see the improvements.
Hope it helps. If it didn't hopefully someone else will also post an opinion
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate page content
These two URLs are being flagged as 98% similar in the code. We're a large ecommerce site, and while it would be ideal to have unique product descriptions on each page we currently don't have the bandwith. Thoughts on what else might be triggering this duplicate content? https://www.etundra.com/restaurant-parts/cooking-equipment-parts/fryers/scoops-skimmers/fmp-175-1081-fryer-crumb-scoop/ https://www.etundra.com/restaurant-equipment/concession-equipment/condiment-pumps/tablecraft-664-wide-mouth-condiment-pump/ Thanks, Natalie
On-Page Optimization | | eTundra0 -
Delete or not delete outdated content
Hi there!
On-Page Optimization | | Enrico_Cassinelli
We run a website about a region in Italy, the Langhe area, where we write about wine and food, local culture, and we give touristic informations. The website also sports a nice events calendar: in 4 years we (and our users) loaded more than 5700 events. Now, we're starting to have some troubles managing this database. The database related to events is huge both in file size and number of rows. There are a lot of images that eat up disk space, and also it's becoming difficult to manage all the data in our backend. Also, a lot of users are entering the website by landing on outdated events. I was wondering if it could be a good idea to delete events older than 6 months: the idea was to keep only the most important and yearly recurring events (which we can update each year with fresh information), and trash everything else. This of course means that 404 errors will increase, and also that our content will gettin thinner, but at the same time we'll have a more manageable database, and the content will be more relevant and "clean". What do you think? thank you 🙂 Best0 -
Prevent indexing of dynamic content
Hi folks! I discovered bit of an issue with a client's site. Primarily, the site consists of static html pages, however, within one page (a car photo gallery), a line of php coding: dynamically generates a 100 or so pages comprising the photo gallery - all with the same page title and meta description. The photo gallery script resides in the /gallery folder, which I attempted to block via robots.txt - to no avail. My next step will be to include a: within the head section of the html page, but I am wondering if this will stop the bots dead in their tracks or will they still be able to pick-up on the pages generated by the call to the php script residing a bit further down on the page? Dino
On-Page Optimization | | SCW0 -
Number of characters to duplicate content
I wonder how much characters in a page title so it can be characterized for Googleas duplicate content?
On-Page Optimization | | imoveiscamposdojordao
Sorry for the English, I used Google Translator.
I'm from Brazil 😄
Thanks.0 -
Checking Duplicate Content
Hi there, We are migrating to a new website, which we are writing lots of new content for the new website. The new website is hosted on a development site which is password protected and so on so that it cannot be indexed. What i would like to know is, how do i check for duplicate content issues out there on the world wide web with the dev site being password protected? Hope this makes sense. Kind Regards,
On-Page Optimization | | Paul780 -
Split testing and dupe content
Hi Everyone, good to be here. I'd like to do split testing in Adwords, currently with a clients site we are selling from a normal site with navigation. The site has about 5 specific products, I want to dupe one of the products and create a funnel without navigation distractions right to checkout. Then A/B test the same product pages in Adwords, one with nav and one without. Will the dupe content be ignored do you think? I'm only slightly concerned as the product pages rank well at the moment.
On-Page Optimization | | eonicWeb0 -
Duplicate content on my domain
I have several pages on my domain that are using the same content except for changing a paragraph or sentence. Do I need to create unique content, even though much of the information pertains to a feature and is related?
On-Page Optimization | | Court_H0 -
Cross Domain Duplicate Content
Hi My client has a series of websies, one main website and several mini websites, articles are created and published daily and weekly, one will go on a the main website and the others on one, two, or three of the mini sites. To combat duplication, i only ever allow one article to be indexed (apply noindex to articles that i don't wanted indexed by google, so, if 3 sites have same article, 2 sites will have noindex tag added to head). I am not completely sure if this is ok, and whether there are any negative affects, apart from the articles tagged as noindex not being indexed. Are there any obvious issues? I am aware of the canonical link rel tag, and know that this can be used on the same domain, but can it be used cross domain, in place of the noindex tag? If so, is it exactly the same in structure as the 'same domain' canonical link rel tag? Thanks Matt
On-Page Optimization | | mattys0