Duplicate Content?
-
My site has been archiving our newsletters since 2001. It's been helpful because our site visitors can search a database for ideas from those newsletters. (There are hundreds of pages with similar titles: archive1-Jan2000, archive2-feb2000, archive3-mar2000, etc.)
But, I see they are being marked as "similar content." Even though the actual page content is not the same. Could this adversely affect SEO? And if so, how can I correct it?
Would a separate folder of archived pages with a "nofollow robot" solve this issue? And would my site visitors still be able to search within the site with a nofollow robot?
-
Cool. No worries
StackOverFlow has always been awesome in helping me with my IIS rules and such.
If you Google: site:stackoverflow.com apache redirect
You will see MANY examples of how to set up 301 redirects, including redirecting from non-www to www pages, etc.
Hope this helps.
Mike
-
Yes, on Google webmaster...sorry. And it's apache.
thank u!
-
Google Analytics or Google Webmaster Tools? You will need to do that in Webmaster Tools.
That is a bummer they are having issues with your 301 redirects. If you know whether you are using Apache, IIS, etc. for your backend, you could post the code you are using in a new question and hopefully someone in the SEOMoz community can help; otherwise, there are Apache and IIS forums where you can post and get some great results and/or examples to base your redirects off of too.
Good luck Sarah! I hope you get your site in shape and back on page 1!!!
Mike
-
HI Mike,
Thank you. To change all the titles is a huge task, there are hundreds and hundreds of pages. I think I'll put them in a folder and mark the page link to that folder with a nofollow. As to the canoncalization of the two names, I have marked one of them as the top one in Google Analytics. But I have a much greater problem than that. I have several domain names that are on the same server and that all point to the one domain (same files and folders). I have been attempting to get my server techs to do a 301 redirect so that only http://www.sundayschoolnetwork.com displays in a browser. However, every time they attempt to do it, part or all of my site stops working correctly.
-
You can go back and fix all of your old title tags, making them unique, like Newsletter Archive | Month Year | Sunday School Network, which will get rid of your errors and provide a better user experience. This approach will allow you to target specific keywords on each page for ranking in Google. When you have the same title across multiple pages, the assumption is that the content is either the same or very similar.
I noticed you have a canonical issue, where you can access your site via http://sundayschoolnetwork.com as well as http://www.sundayschoolnetwork.com
The issue with this, that you have 44 relatively important links from external websites pointing to the non-www version (http://sundayschoolnetwork.com)... which means you are splitting up your potential power between two sites instead of one. There are many ways you can fix this.
As for why you are not ranking as well, it could be the market became more competitive for the keywords you were originally using. It could be that your site content does not reflect the keywords you are targeting. It could be lots of things.
Like I said in my previous post, the nofollow tells crawlers not to follow the internal and external links on those pages; however, they will still get indexed. This means that you will still have duplicate titles appearing in results. The way to remove them from the results would be to use the noindex directive - which will eventually remove them from the index and you will not have competing title tags.
If you fix your title tags, you do not need to worry about the nofollow or noindex directives.
That is about all I can help with, without knowing any additional information.
The only other thing I can suggest is to read the SEOMoz Beginners Guide to SEO - which will help a TON!
I hope that helps.
Mike
-
thank u. I'm gonna do that!
-
Hi Mike,
That was fast. I copied some of the report from Seomoz "Crawled Diagnostics." Some do have the same titles, which was an edition after many years. The early newsletters I didn't even title, so they have a "default title" of the url.
I happened on SEOmoz, because I am trying to figure out why after so many years of having been on the first or second page of Google search results, we are lucky to show up on page 10 or deeper, if at all.
So I'm trying out SEOmoz to see if this will help us get back on top!
|
The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive13_Apr10.html 1 18 1 The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive13_Apr11.html 1 18 1 The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive13_Apr12.html 1 18 1 http://sundayschoolnetwork.com/archive13_Feb06.html
http://sundayschoolnetwork.com/archive13_Feb06.html 1 18 1 http://sundayschoolnetwork.com/archive13_Feb07.html
http://sundayschoolnetwork.com/archive13_Feb07.html 1 18 1 The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive14_Apr08.html 1 18 1 The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive14_Apr09.html 1 18 1 The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive14_Apr11.html 1 18 1 The Sunday School Teacher's Network Newsletter - Great ideas for children's ministry!
http://sundayschoolnetwork.com/archive14_Apr12.html 1 18 1 http://sundayschoolnetwork.com/archive14_Feb06.html
-
Hi Sarah,
If the titles are different and the page content is different, I do not understand why you should be getting any errors.
What tool are you using that is giving you the "similar content" message?
Your site visitors will still be able to search your site with nofollow in place, because nofollow is simply a directive telling search engines to not follow the internal and external links on your page.
The noindex directive tells Google to not index the content on the selected pages.
If you can provide me with the name of the tool you are receiving the "similar content" message from and/or provide me with your website address I could take a look into things further.
... long story short, if your titles are unique and your content is unique, you should not have to worry about duplicate content.
Hope this helps,
Mike
-
The best way to go is to put all your newsletters in on folder and and disallow the folder in your robot.txt.
rel nofollow & robot.txt are only read by google bot, your visitors won't be affected and will be able to navigate & search the archives without problem.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content for Locations on my Directory Site
I have a pretty big directory site using Wordpress with lots of "locations", "features", "listing-category" etc.... Duplicate Content: https://www.thecbd.co/location/california/ https://www.thecbd.co/location/canada/ referring URL is www.thecbd.co is it a matter of just putting a canonical URL on each location, or just on the main page? Would this be the correct code to put: on the main page? Thanks Everyone!
Technical SEO | | kay_nguyen0 -
Duplicate content problem
Hi there, I have a couple of related questions about the crawl report finding duplicate content: We have a number of pages that feature mostly media - just a picture or just a slideshow - with very little text. These pages are rarely viewed and they are identified as duplicate content even though the pages are indeed unique to the user. Does anyone have an opinion about whether or not we'd be better off to just remove them since we do not have the time to add enough text at this point to make them unique to the bots? The other question is we have a redirect for any 404 on our site that follows the pattern immigroup.com/news/* - the redirect merely sends the user back to immigroup.com/news. However, Moz's crawl seems to be reading this as duplicate content as well. I'm not sure why that is, but is there anything we can do about this? These pages do not exist, they just come from someone typing in the wrong url or from someone clicking on a bad link. But we want the traffic - after all the users are landing on a page that has a lot of content. Any help would be great! Thanks very much! George
Technical SEO | | canadageorge0 -
Duplicate Content
Hello guys, After fixing the rel tag on similar pages on the site I thought that duplicate content issue were resolved. I checked HTML Improvements on GWT and instead of going down as I expected, it went up. The duplicate issues affect identical product pages which differ from each other just for one detail, let's say length or colour. I could write different meta tags as the duplicate is the meta description, and I did it for some products but still didn't have any effects and they are still showing as duplicates. What would the problem be? Cheers
Technical SEO | | PremioOscar0 -
Duplicate content issue with Wordpress tags?
Would Google really discount duplicate content created by Wordpress tags? I find it hard to believe considering tags are on and indexed by default and the vast majority of users would not know to deindex them . . .
Technical SEO | | BlueLinkERP0 -
Duplicate title/content errors for blog archives
Hi All Would love some help, fairly new at SEO and using SEOMoz, I've looked through the forums and have just managed to confuse myself. I have a customer with a lot of duplicate page title/content errors in SEOMoz. It's an umbraco CMS and a lot of the errors appear to be blog archives and pagination. i.e. http://example.com/blog http://example.com/blog/ http://example.com/blog/?page=1 http://example.com/blog/?page=2 and then also http://example.com/blog/2011/08 http://example.com/blog/2011/08?page=1 http://example.com/blog/2011/08?page=2 http://example.com/blog/2011/08?page=3 (empty page) http://example.com/blog/2011/08?page=4 (empty page) This continues for different years and months and blog entries and creates hundreds of errors. What's the best way to handle this for the SEOMoz report and the search engines. Should I rel=canonical the /blog page? I think this would probably affect the SEO of all the blog entries? Use robots.txt? Sitemaps? URL parameters in the search engines? Appreciate any assistance/recommendations Thanks in advance Ian
Technical SEO | | iragless0 -
301ed Pages Still Showing as Duplicate Content in GWMT
I thank anyone reading this for their consideration and time. We are a large site with millions of URLs for our product pages. We are also a textbook company, so by nature, our products have two separate ISBNs: a 10 digit and a 13 digit form. Thus, every one of our books has at least two pages (10 digit and 13 digit ISBN page). My issue is that we have established a 301 for all the 10 digit URLs so they automatically redirect to the 13 digit page. This fix has been in place for months. However, Google still reports that they are detecting thousands of pages with duplicate title and meta tags. Google is referring to these page URLs that I already have 301ed to the canonical version many months ago! Is there anything that I can do to fix this issue? I don't understand what I am doing wrong. Example:
Technical SEO | | dfinn
http://www.bookbyte.com/product.aspx?isbn=9780321676672
http://www.bookbyte.com/product.aspx?isbn=032167667X As you can see the 10 digit ISBN page 301s to 13 digit canonical version. Google reports that they have detected duplicate title and meta tags between the two pages and there are thousands of these duplicate pages listed. To add some further context: The ISBN is just a parameter that allows us to provide content when someone searches for a product with the 10 or 13 digit ISBN. The 13 digit version of the page is the only physical page that exists, the 10 digit is only a part of the virtual URL structure of the website. This is why I cannot simply change the title and meta tags of the 10 digit pages because they only exist in the sense that the URL redirects to the 13 digit version. Also, we submit a sitemap every day of all the 13 digit pages so Google knows exactly what our physical URL structure is. I have submitted this question to GWMT forums and received no replies.0 -
How do i deal with duplicate content on the same domain?
I'm trying to find out if there's a way we can combat similar content on different pages on the same site, without having to re write the whole lot? Any ideas?
Technical SEO | | indurain0 -
CGI Parameters: should we worry about duplicate content?
Hi, My question is directed to CGI Parameters. I was able to dig up a bit of content on this but I want to make sure I understand the concept of CGI parameters and how they can affect indexing pages. Here are two pages: No CGI parameter appended to end of the URL: http://www.nytimes.com/2011/04/13/world/asia/13japan.html CGI parameter appended to the end of the URL: http://www.nytimes.com/2011/04/13/world/asia/13japan.html?pagewanted=2&ref=homepage&src=mv Questions: Can we safely say that CGI parameters = URL parameters that append to the end of a URL? Or are they different? And given that you have rel canonical implemented correctly on your pages, search engines will move ahead and index only the URL that is specified in that tag? Thanks in advance for giving your insights. Look forward to your response. Best regards, Jackson
Technical SEO | | jackson_lo0