Duplicate Content Issues with Forum
-
Hi Everyone,
I just signed up last night and received the crawl stats for my site (ShapeFit.com). Since April of 2011, my site has been severely impacted by Google's Panda and Penguin algorithm updates and we have lost about 80% of our traffic during that time. I have been trying to follow the guidelines provided by Google to fix the issues and help recover but nothing seems to be working. The majority of my time has been invested in trying to add content to "thin" pages on the site and filing DMCA notices for copyright infringement issues. Since this work has not produced any noticeable recovery, I decided to focus my attention on removing bad backlinks and this is how I found SEOmoz.
My question is about duplicate content. The crawl diagnostics showed 6,000 errors for duplicate page content and the same for duplicate page title. After reviewing the details, it looks like almost every page is from the forum (shapefit.com/forum). What's the best way to resolve these issues? Should I completely block the "forum" folder from being indexed by Google or is there something I can do within the forum software to fix this (I use phpBB)?
I really appreciate any feedback that would help fix these issues so the site can hopefully start recovering from Panda/Penguin.
Thank you,
Kris
-
Hi Alan,
Thanks for your feedback. In regards to thin content, do you know if there is a certain minimum requirement of textual content that should be on a page? I have read a few articles that recommend having at least 300 words for each page. Do you know if there any validity to this?
I appreciate your input about the forum issues. I will look into adding/updating the robots file to block everything but the main posts.
What are your thoughts on removing bad backlinks and how critical is this for Panda/Pengiun recovery? There are so many areas to work on but only so many hours in the day
Thanks,
Kris
-
Kris,
The ideal scenario is to eliminate or consolidate content that exists on extremely thin pages on a site. When it comes to forums, consolidation is a near impossibility technically and likely to cause problems for usability. As a result of those challenges, the best course of action would to either eliminate the content entirely or block the forums from indexation via robots.txt file.
If there is any potential for any of the forum content being of value from a search engine perspective, it would likely be a case where you'd want to keep the core forum posts indexable but block the member pages.
Unfortunately there's no one answer and no true way to gauge the impact on any action you choose without actually taking that action and then waiting to see what happens.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Multi channel product descriptions & dupe content issues
Hi When filling in inventory files to upload to the likes of Amazon clients will usually be copying and pasting the product descriptions from the website product descriptions into the Amazon product description field Should they really be re-written to avoid dupe content issues ? I presume not since it is the official description of the product. Please note that i'm talking here about the manufacturers website/product descriptions and their own Amazon shop descriptions. So theirs is the original authoritative description. Cheers Dan
On-Page Optimization | | Dan-Lawrence0 -
Duplicate Pages software
Hey guys, i was told few hours ago about a system that can take few of your keywords and automatically will create new links and pages (in the map file) for your website, so a website that was build with 20 pages( for example) will be shown to SE as a site with hundreds of pages, thing that should help the SEO IS anyone heard about such a software? is it legal? any advice that you can give on this mater? Thanks i.
On-Page Optimization | | iivgi0 -
How to avoid duplicates when URL and content changes during the course of a day?
I'm currently facing the following challenge: Newspaper industry: the content and title of some (featured) articles change a couple of times during a normal day. The CMS is setup so each article can be found by only using it's specific id (eg. domain.tld/123). A normal article looks like this: domain.tld/some-path/sub-path/i-am-the-topic,123 Now the article gets changed and with it the topic. It looks like this now: domain.tld/some-path/sub-path/i-am-the-new-topic,123 I can not tell the writers that they can not change the article as they wish any more. I could implement canonicals pointing to the short url (domain.tld/123). I could try to change the URL's to something like domain.tld/some-path/sub-path/123. Then we would lose keywords in URL (which afaik is not that important as a ranking factor; rather as a CTR factor). If anyone has experiences sharing them would be greatly appreciated. Thanks, Jan
On-Page Optimization | | jmueller0 -
Are these considered duplicates?
http://www.domain.com/blog/sample-blog-post/#more-0001 http://www.domain.com/blog/sample-blog-post/ The first URL is coming from a "click here" hyperlink from the excerpt of the 2nd URL in my homepage. Thanks in advance!
On-Page Optimization | | esiow20130 -
Duplicate content with a trailing slash /
Hi, I 've pages like this: A) www.example.com/file/ B) www.example.com/file Just two questions: Does Google see this as duplicate content? Best to 301 redirect B to A? Many thanks Richard PS I read previous threads re the subject, it sounded like there was a bug in SEOMoz but I was not absolutely clear. Apologies if this is going over old ground.
On-Page Optimization | | Richard5550 -
What should I do with these duplicate mass production?
Hi, I'm reviewing somebodies site and just realized that it's overflown with duplicates. Like these: <colgroup><col width="3496"></colgroup>
On-Page Optimization | | jjtech
| www.joannalark.com/store/products/24"-Sting.html |
| www.joannalark.com/store/products/24"-Sting.html?setCurrencyId=1 |
| www.joannalark.com/store/products/24"-Sting.html?setCurrencyId=6 |
| www.joannalark.com/store/products/24"-Sting.html?setCurrencyId=7 | It also produces something like this: | <colgroup><col width="3496"></colgroup>
| www.joannalark.com/store/pages/pages/pages/pages/pages.php?pageid=8 |
| www.joannalark.com/store/pages/pages/pages/pages/pages/pages.php?pageid=8 |
| www.joannalark.com/store/pages/pages/pages/pages/pages/pages/pages.php?pageid=8 |
| www.joannalark.com/store/pages/pages/pages/pages/pages/pages/pages/pages.php?pageid=8 |
| www.joannalark.com/store/pages/pages/pages/pages/pages/pages/pages/pages/pages.php?pageid=8 |
| www.joannalark.com/store/pages/pages/pages/pages/pages/pages/pages/pages/pages/pages.php?pageid=8 | |
|
|
|
|
| I don't know what to do with that and would appreciate any help Thanks, JJ <colgroup><col width="3496"></colgroup>
| |
| |
| |
| |
| |
| |0 -
Duplicate Page Content Issue
For one of our campaigns, we have 164 errors for Duplicate Page Content. We have a website where much of the same content lives in two different places on their website. The information needs to be accessible from both areas. What is the best way to tackle this problem? Is there anything that can be done so these pages are not competing against one another? If the only solution is to edit the content on one of the pages, how much of the content has to be different? Is there a certain percentage to go by? Here is an example of what I am referring to: 1.) http://www.valleyorthopedicassociates.com/services/foot-center/preventing-sprains-and-strains 2.) http://www.valleyorthopedicassociates.com/patient-resources/service/foot-and-ankle-center/preventing-sprains-and-strains
On-Page Optimization | | cmaseattle1 -
Cross Domain Duplicate Content
Hi My client has a series of websies, one main website and several mini websites, articles are created and published daily and weekly, one will go on a the main website and the others on one, two, or three of the mini sites. To combat duplication, i only ever allow one article to be indexed (apply noindex to articles that i don't wanted indexed by google, so, if 3 sites have same article, 2 sites will have noindex tag added to head). I am not completely sure if this is ok, and whether there are any negative affects, apart from the articles tagged as noindex not being indexed. Are there any obvious issues? I am aware of the canonical link rel tag, and know that this can be used on the same domain, but can it be used cross domain, in place of the noindex tag? If so, is it exactly the same in structure as the 'same domain' canonical link rel tag? Thanks Matt
On-Page Optimization | | mattys0