Duplicate page content and Duplicate page title errors
-
Hi,
I'm new to SeoMoz and to this forum. I've started a new campaign on my site and got back loads of error.
Most of them are Duplicate page content and Duplicate page title errors. I know I have some duplicate titles but I don't have any duplicate content.
I'm not a web developer and not so expert but I have the impression that the crawler is following all my internal links (Infact I have also plenty of warnings saying "Too many on-page links".
Do you think this is the cause of my errors? Should I implement the nofollow on all internal links? I'm working with Joomla.
Thanks a lot for your help
Marco
-
Hi Marco,
I took a look at your page at http://www.beautifulpuglia.com/it/linea-costiera/isole-tremiti.html
Looks like you've got the canonical in place okay here. The next step is to add the canonical on every page that is a duplicate of this page. And you want to make sure to point to the right page. Let me be clear: Every page that is a duplicate of this page should have the same canonical. In this case:
<link rel=”canonical” href=[”http://www.beautifulpuglia.com/it/gargano/isole-tremiti.html”/](view-source:http://www.beautifulpuglia.com/linea-costiera/%E2%80%9Dhttp://www.beautifulpuglia.com/it/gargano/isole-tremiti.html%E2%80%9D/)>
You can find the other pages you need to add this tag to in your SEOmoz report. In each duplicated content report, it will list the number of other pages that are duplicates. Simply click on the number to see the URLs.
I'm not a Joomla expert, but webmasters I've talked to have expressed that other platforms such as Wordpress and Drupal are much more accommodating of these types of fixes. There are some various plugin modules you can use, but you'll have to select one appropriate to your configuration.
Here's a good resource from Dr. Pete: http://www.seomoz.org/blog/duplicate-content-in-a-post-panda-world
Hope this helps. Best of luck.
-
Elias,
I too have 'thousands' of duplicated errors in SEOmoz. Most of which are because it is returning
/abc.com as a different page to /ABC.com
Surely Google doesn't do that? Just because one URL is in capital and the other small case? I also have no idea where SEOmoz is picking that up from......possibly links internal to the page with the hyperlink using different case?
It seems to me this is too sensitive and for me to fix that would take WEEKS!!!! I fail to see if there would be any uplift if Google sees beyond that issue as its cosmetic and not functional.
Regards
Andy
-
It looks fine to me. You will need to do the same on all of your pages.
If you've just added the code you will need to wait up to a week for SEOmoz to re-crawl your website depending on when you're site crawl is scheduled.
Let me know how you get on.
Elias
-
Hi Elias, Hi Marisa,
thanks you both
you are right, in the meantime I had done this but I have the impression it is not working and I don't know what I'm doing wrong.
I'm attaching a link to a page of my site (I hope I can do this). Please have a look at the code, you will see the tag rel=”canonical” href=”http://www.beautifulpuglia.com/it/gargano/isole-tremiti.html”/> which is indicating the URL I want to use. However SeoMoz is still giving me the error. And this is happening for both the Italian and English version.
So far I've only added the tag to this page, I want to find the solution before modifying all pages currently affected.
http://www.beautifulpuglia.com/it/linea-costiera/isole-tremiti.html
Thanks a lot again
-
Hi Marco, as Marissa says - by putting the canonical tag on one page you are putting it on all of them as they are in fact the same page - they are just reached by different URLs.
-
www.site.com/ and www.site.com/index.html, site.com/index.html/, ect, are already the same page. So, there's only one page TO put the tag on. You're just telling the crawlers that you only want one of them to get the credit, and which version of the page you prefer to be displayed.
-
Hi Elias,
thanks a lot for your reply. I've read few posts about the canonical tag and Yes I'm going to try it.
Just couple of things:
-
Let's say I have 4 duplicate for one page, I presume I have to add the tag in the head of only one page right? Does it make any difference which one I pick?
-
Any idea on how this can be implemented in Joomla?It doesn't seem to be very straightforward.
Thanks a lot
Marco
-
-
Hi Marco,
It seems to me like you need to implement the canonical tag.
Site crawlers/bots will consider the following pages as different pages because of their URL and thus tell indicate to them that the content is duplicated on each page...
By implementing the following tag on each of your sites pages (changing the URL for each page) you will tell the crawler which page they should be indexing and to ignore the other.
Here's an example of a canonical tag (to be placed within the head tag of the page)
I think this will sort out your duplication issues.
You can find more information about canonical URLs here http://www.seomoz.org/blog/canonical-url-tag-the-most-important-advancement-in-seo-practices-since-sitemaps
I hope this helps!
Elias
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Title page google serp
Why does Google change the titles automatically? I have <title>Canyoning Açores - São Jorge | Discover Experience Açores</title> but google show Discover Experience Açores: Canyoning Açores - São Jorge
Intermediate & Advanced SEO | | tiagoarruda0 -
Duplicate content. Competing for rank.
Scenario: An automotive dealer lists cars for sale on their website. The descriptions are very good and in depth at 1,200 words per car. However chunks of the copy are copied from car review websites and weaved into their original copy. Q1: This is flagged in copyscape - how much of an issue is this for Google? Q2: The same stock with the same copy is fed into a popular car listing website - the dealer's website and the classifieds website often rank in the top two positions (sometimes the dealer on top other times the classifieds site). Is this a good or a bad thing? Are you risking being seen as duplicating/scraping content? Thank you.
Intermediate & Advanced SEO | | Bee1590 -
Wondering if creating 256 new pages would cause duplicate content issues
I just completed a long post that reviews 16 landing page tools. I want to add 256 new pages that compare each tool against each other. For example: Leadpages vs. Instapage Leadpages vs. Unbounce Instapage vs. Unbounce, etc Each page will have one product's information on the left and the other on the right. So each page will be a unique combination BUT the same product information will be found on several other pages (its other comparisons vs the other 15 tools). This is because the Leadpages comparison information (a table) will be the same no matter which tool it is being compared against. If my math is correct, this will create 256 new pages - one for each combination of the 16 tools against each other! My site now is new and only has 6 posts/pages if that matters. Want to make sure I don't create a problem early on...Any thoughts?
Intermediate & Advanced SEO | | martechwiz0 -
Questions about duplicate photo content?
I know that Google is a mystery, so I am not sure if there are answers to these questions, but I'm going to ask anyway! I recently realized that Google is not happy with duplicate photo content. I'm a photographer and have sold many photos in the past (but retained the rights for) that I am now using on my site. My recent revelations means that I'm now taking down all of these photos. So I've been reverse image searching all of my photos to see if I let anyone else use it first, and in the course of this I found out that there are many of my photos being used by other sites on the web. So my questions are: With photos that I used first and others have stolen, If I edit these photos (to add copyright info) and then re-upload them, will the sites that are using these images then get credit for using the original image first? If I have a photo on another one of my own sites and I take it down, can I safely use that photo on my main site, or will Google retain the knowledge that it's been used somewhere else first? If I sold a photo and it's being used on another site, can I safely use a different photo from the same series that is almost exactly the same? I am unclear what data from the photo Google is matching, and if they can tell the difference between photos that were taken a few seconds apart.
Intermediate & Advanced SEO | | Lina5000 -
Big problem with duplicate page content
Hello! I am a beginner SEO specialist and a have a problem with duplicate pages content. The site I'm working on is an online shop made with Prestashop. The moz crawl report shows me that I have over 4000 duplicate page content. Two weeks ago I had 1400. The majority of links that show duplicate content looks like bellow:
Intermediate & Advanced SEO | | ana_g
http://www.sitename.com/category-name/filter1
http://www.sitename.com/category-name/filter1/filter2 Firstly, I thought that the filtres don't work. But, when I browse the site and I test it, I see that the filters are working and generate links like bellow:
http://www.sitename.com/category-name#/filter1
http://www.sitename.com/category-name#/filter1/filter2 The links without the # do not work; it messes up with the filters.
Why are the pages indexed without the #, thus generating me duplicate content?
How can I fix the issues?
Thank you very much!0 -
Duplicate content on subdomains
Hi All, The structure of the main website goes by http://abc.com/state/city/publication - We have a partnership with public libraries to give local users access to the publication content for free. We have over 100 subdomains (each for an specific library) that have duplicate content issues with the root domain, Most subdomains have very high page authority (the main public library and other local .gov websites have links to this subdomains).Currently this subdomains are not index due to the robots text file excluding bots from crawling. I am in the process of setting canonical tags on each subdomain and open the robots text file. Should I set the canonical tag on each subdomain (homepage) to the root domain version or to the specific city within the root domain? Example 1:
Intermediate & Advanced SEO | | NewspaperArchive
Option 1: http://covina.abc.com/ = Canonical Tag = http://abc.com/us/california/covina/
Option 2: http://covina.abc.com/ = Canonical Tag = http://abc.com/ Example 2:
Option 1: http://galveston.abc.com/ = Canonical Tag = http://abc.com/us/texas/galveston/
Option 2: http://galveston.abc.com = Canonical Tag = http://abc.com/ Example 3:
Option 1: http://hutchnews.abc.com/ = Canonical Tag = http://abc.com/us/kansas/hutchinson/
Option 2: http://hutchnews.abc.com/ = Canonical Tag = http://abc.com/ I believe it makes more sense to set the canonical tag to the corresponding city (option 1), but wondering if setting the canonical tag to the root domain will pass "some link juice" to the root domain and it will be more beneficial. Thanks!0 -
Duplicate Titles caused by multiple variations of same URL
Hi. Can you please advise how I can overcome this issue. Moz.com crawle is indicating I have 100's of Duplicate Title tag errors. However this is caused because many URL's have been indexed multiple times in Google. For example. www.abc.com
Intermediate & Advanced SEO | | adhunna
www.abc.com/?b=123 www.abc.com/ www.abc.com/?b=654 www.abc.com/?b=875 www.abc.com/index.html What can I do to stop this issue being reported as duplictae Titles, as well as content? I was thinking maybe I can use Robots.txt to block various query string parameters. I'm Open to ideas and examples.0 -
Duplicate Content/ Indexing Question
I have a real estate Wordpress site that uses an IDX provider to add real estate listings to my site. A new page is created as a new property comes to market and then the page is deleted when the property is sold. I like the functionality of the service but it creates a significant amount of 404's and I'm also concerned about duplicate content because anyone else using the same service here in Las Vegas will have 1000's of the exact same property pages that I do. Any thoughts on this and is there a way that I can have the search engines only index the core 20 pages of my site and ignore future property pages? Your advice is greatly appreciated. See link for example http://www.mylvcondosales.com/mandarin-las-vegas/
Intermediate & Advanced SEO | | AnthonyLasVegas0