Duplicate Page Content
-
Hi there,
We keep getting duplicate page content issues. However, its not actually the same page.
E.G - There might be 5 pages in say a Media Release section of the website. And each URL says page 1, 2 etc etc. However, its still coming up as duplicate. How can this be fixed so Moz knows its actually different content? -
Thanks all - will give those options a try and see which works the best for us.
-
Hi!
I suggested the noindex in order to deindex pages that maybe are already indexed. But, yes, the rel="canonical" should be doing the same (the problem is that Google may not respect it).
The nofollow is order to not letting the crawler wasting budget crawl following the links of those (many) pages.
-
Gianluca,
Wouldn't be much more work to identify if the parameter is set and then add the noindex meta? Wouldn't be easier to just set the canonical? I'm sure that's a dynamic site, just one canonical cal without using any extra code (PHP or whatever).
Why the nofollow? If I just preventing that page for being indexed as it would constitute a duplicate content issue, why the nofollow?, noindex should be enough in this case.
We recently fixed a similar issue with our blog tags, showing duplicate content on about 400 pages. We fixed that by adding the noindex (they already had the canonical but it wasn't enough as the canonical couldn't point to a definite version as that changed if the tag had or not another post on it). Within a few days all those pages were deindexed, we noticed a loss in search traffic and I decided to run a small test removing the noindex tag. Results: 2 weeks later none of those pages returned to the index (I added the noindex tag back as it was just a test to see if we could regain that traffic, but ultimately decided it wouldn't help to have a duplicate content issue for that lost traffic).
-
Federico is right.
Your duplicated content issue is due to the date parameters, hence you are potentially duplicating every page having that calendar for all the possible combination of dates... and that is an huge issue.
You should implement the rel="canonical" in order to have all these kind of URLs having as canonical the URL without the parameter.
Or, even better, you should implement the meta robots "noindex,nofollow" in every date parametered URL.
Said that, the most logical thing to do was to block these URLs via robots.txt when launching the site. Unfortunately, now blocking these URLs is not enough, as they are already indexed (even if they not appear in the index because they are filtered out by Google).
-
Ah you mean that if the dates of the reservation changes then it creates a duplicate page content?
If that's the case, you should use the rel="canonical" the definite page, no dates selected, just the page that shows the property.
-
Did you try adding the rel="canonical" tag to the pages?
-
So they might look at this page: http://www.hihh.com.au/property-details?hihhpropertyId=HCP006&checkin=2013-08-06&checkout=2013-08-09&search=checkindate%3D2013-08-06%26checkoutdate%3D2013-08-09
Then the same page would come up on the error list but with different dates.
-
Can you provide us with some examples? It would make our job easier
-
Its basically all seperate pages/URL's with different information on each. However it seems to be crawled for each possible range of that page. e.g for check in/check out dates. It will search a range of dates and think that each page has different information. However, its all exactly the same.
-
Is the issue on the pagination? as sometimes some pages from categories/tags/etc can have the same content within an exact page.
If that's the issue, I would recommend you add a noindex meta to the least important pages (tags for example).
Hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content
I run a Business Directory, Where all the businesses are listed. I am having an issue.. with Duplication content. I have categories, Like A, B, C Now a business in Category A, User can filter it by different locations, by State, City, Area So If they filter it by the State and the State has 10 businesses and all of them are in one City. Both of the page
On-Page Optimization | | Adnan4SEO
The state filtered and the city filtered are same, What can i do to avoid that? canonical-url-tag or changing page Meta's and Body text? Please help 🙂0 -
Duplicate Content, http vs https
Hi All! I just discovered that a client of ours a duplicate content issue. Essentially they have approximately 20 pages that have an http and an https version. Is there a better way to handle this than a simple 301? Regards, Frank
On-Page Optimization | | FrankSweeney0 -
Duplicate Page Titles in Crawl Errors (although Google is rewriting in serps ??)
Hi Im working on a client/project and crawl report is showing thousands of dupe page titles In the case of the blog/news section its aprox 50 since aprox 50 posts and they all have the same meta-title: "Brand News | Brand" as opposed to: "Title Unique to Page/Topic/KW Relating to Content | Brand" Since these are the main content pages we want to rank (in addition to the main site category pages) then i have instructed dev must prioritise populating these pages meta-titles with the actual post/article titles, as per the latter version of the above example. (I should mention that i have requested they fix all dupe titles but main content pages are the priority). Whilst this will reduce the number of dupe titles in crawl error/warning report which is a good thing, is it actually likely to increase the ranking of these news/content pages given that Google does seem to be rewriting the titles correctly in the serps based on the page content ? Many Thanks in advance for your input
On-Page Optimization | | Dan-Lawrence0 -
I have a question about having to much content on a single page. Please help :)
I am working on a music related site. We are building a feature in our system to allow people to write information about songs on their playlist. So when a song is currently being played a user can read some cool facts or information about the song. http://imgur.com/5jFumPW ( screenshot). Some playlists have over 100 songs and could be completely random in genre and artist. I am wondering if some of these playlists have over 5,000 words of content if that is going to hurt us? We will be very strict about making sure its non spammy and good content. Also for the titles of the content is it bad to have over 100 h3 tags on one page? Just want to make sure we are on the right track. Any advice is greatly appreciated.
On-Page Optimization | | mikecrib10 -
Different pages for OS's vs 1 Page with Dynamic Content (user agent), what's the right approach?
We are creating a new homepage and the product are at different stages of development for different OS's. The value prop/messaging/some target keywords will be different for the various OS's for that reason. Question is, for SEO reasons, is it better to separate them into different pages or use 1 page and flip different content in based on the user agent?
On-Page Optimization | | JoeLin0 -
What is the best way to manage industry required duplicate Important Safety Information (ISI) content on every page of a site?
Hello SEOmozzer! I have recently joined a large pharmaceutical marketing company as our head SEO guru, and I've encountered a duplicate content related issue here that I'd like some help on. Because there is so much red tape in the pharmaceutical industry, there are A LOT of limitations on website content, medication and drug claims, etc. Because of this, it is required to have Important Safety Information (ISI) clearly stated on every page of the client's website (including the homepage). The information is generally pretty lengthy, and in some cases is longer than the non-ISI content on each page. Here is an example: http://www.xifaxan.com/ All content under the ISI header is required on each page. My questions are: How will this duplicated content on each page affect our on-page optimization scores in the eyes of search engines? Is Google seeing this simply as duplicated content on every page, or are they "smart" enough to understand that because it is a drug website, this is industry standard (and required)? Aside from creating more meaty, non-ISI content for the site, are there any other suggestions you have for handling this potentially harmful SEO situation? And in case you were going to suggest it, we cannot simply have an image of the content, as it may not be visible by all internet users. We've already looked into that 😉 Thanks in advance! Dylan
On-Page Optimization | | MedThinkCommunications0 -
Cross Domain Duplicate Content
Hi My client has a series of websies, one main website and several mini websites, articles are created and published daily and weekly, one will go on a the main website and the others on one, two, or three of the mini sites. To combat duplication, i only ever allow one article to be indexed (apply noindex to articles that i don't wanted indexed by google, so, if 3 sites have same article, 2 sites will have noindex tag added to head). I am not completely sure if this is ok, and whether there are any negative affects, apart from the articles tagged as noindex not being indexed. Are there any obvious issues? I am aware of the canonical link rel tag, and know that this can be used on the same domain, but can it be used cross domain, in place of the noindex tag? If so, is it exactly the same in structure as the 'same domain' canonical link rel tag? Thanks Matt
On-Page Optimization | | mattys0 -
Duplicate content Issue
I'm getting a report of duplicate title and content on: http://www.website.com/ http://www.website.com/index.php Of course, they're the same pages but does this need to be corrected somehow. Thanks!
On-Page Optimization | | dbaxa-2613380