Auto genrated content problem?
-
Hi all,
I operate a Dutch website (sneeuwsporter.nl), the website is a a database of European ski resorts and accommodations (hotels, chalets etc). We launched about a month ago with a database of about 1700+ accommodations. Of every accommodation we collected general information like what village it is in, how far it is from the city centre and how many stars it has. This information is shown in a list on the right of each page (e.g. http://www.sneeuwsporter.nl/oostenrijk/zillertal-3000/mayrhofen/appartementen-meckyheim/). In addition a text of this accomodation is auto generated based on some of the properties that are also in the list (like distance, stars etc).
Below the paragraph about the accommodation is a paragraph about the village the accommodation is located in, this is a general text that is the same with all the accommodations in this village. Below that is a general text about the resort area, this text is also identical on all the accommodation pages in the area. So a lot of these texts about the village and area are used many times on different pages.
Things went well at first and every day we got more Google traffic, and more and more pages. But a few days ago our organic traffic took a near 100% dive, we are hardly listed anymore and if we are at very low places. We expect the Google gave us a penalty. We expect this to be the case because of 2 reasons:
-
we have auto generated text that only vary slightly per page
-
we re-use the content about villages and area's on many pages
We quickly removed the content of the villages and resort area's because we are pretty sure that this is definitely something Google does not want. We are less sure about the auto generated content, is this something we should remove as well? These are normal readable text, they just happen to be structured more or less the same way on every page. Finally, when we made these and maybe some other fixes, what is the best and quickest ways to let Google see us again and show them we improved?
Thanks in advance!
-
-
The page that you have linked to has 3 sentences of text. When I search Google for "Appartementen Meckyheim" it looks like there is a lot of competition. 3 sentence of text is not going to add a lot fo quality to a page.
But, I do think there is more than just a poor ranking issue. I searched through 6 pages and didn't see your page at all. It's still in the index, but it's not ranking.
Also, I'm concerned that the Trail Map and Accessibility pages may look like duplicated content to Google. They really can only evaluate what they can crawl, so this page likely looks the same on every listing you have in Google's eyes.
I am suspicious that there may have been a Panda update in the last few days. Sometimes Google doesn't announce them right away.
Thin content like you have shown us as well as duplicate content are what Panda goes after.
I'm guessing that you ranked well until the Panda filter detected thin and duplicate content. It's possible that removing the duplicated pages will be enough but I'm suspicious that you'll need to have substantially more content such as a thorough review of each place in order to get back to ranking again.
If I am right and there was a Panda update then you may not see recovery after beefing the content up until Panda runs again.
-
Google has been treating sites with lots of page-to-page duplication this way for at least five or six years.
You get indexed, ranked and start getting traffic but when Google figures out that your site was made with a cookie cutter then most of your pages will be filtered from the SERPs.
In my opinion this is different from a penalty. It's simply not showing dupes in the SERPs.
I used to have a lot of autogenerated content. Entire sites with hundreds of thousands of pages dedicated to it. They were kickass for a few weeks to a few months and then tanked hard.
I found that autogenerated content (where it is mainly boiler plate or duplicated) is a continuous expense. (Get killed and replace it, get killed and replace it.)
However, genuine authorship can be an investment that might continue to pay after I am dead (I wouldn't say that if I was twenty years old because strong competitors are popping up in every niche... but since I am one of the older people posting here I can say that with a little more certainty.)
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Simple duplicate content query
Hello Community, One of my clients runs a job board website. They are having some new framework installed which will lead to them having to delete all their jobs and re-add them. The same jobs will be re-posted but with a different reference number which in turn with change each URL. I believe this will cause significant duplicate content issues, I just thought I would get a second opinion on best practice for approaching a situation like this. Would a possible solution be to delete jobs gradually and 301 re-direct old URLs to new URLs? Many thanks in advance, Adam
Technical SEO | | SO_UK0 -
Index problems
“The website http://www.vaneyckshutters.com/nl/ does not show in the index of Google (site:vaneyckshutters.com/nl/). This must be the homepage in the Netherlands. Previously, the page www.vaneyckshutters.com was redirected to /nl/. This page is accessible now with a canonical tag to http://www.vaneyckshutters.com/nl/ in the hope to let /nl/ be indexed. When we look at the SERPS for keyword ‘shutters’, the page http://www.vaneyckshutters.com/ is shown in Google.nl on #32 and in Belgium #3. Problem & question: Why is it that /nl/ has not been indexed properly and why is it that we rank with http://www.vaneyckshutters.com on ‘shutters’ instead the/nl/ page?”
Technical SEO | | Happy-SEO1 -
Duplicate content and canonicalization confusion
Hello, http://bit.ly/1b48Lmp and http://bit.ly/1BuJkUR pages have same content and their canonical refers to the page itself. Yet, they rank in search engines. Is it because they have been targeted to different geographical locations? If so, still the content is same. Please help me clear this confusion. Regards
Technical SEO | | IM_Learner0 -
Duplicate content. Wordpress and Website
Hi All, Will Google punish me for having duplicate blog posts on my website's blog and wordpress? Thanks
Technical SEO | | Mike.NW0 -
Google description problem
Hi all, My website is www.ipbskinning.com I'm having a problem with how my site is appearing in google. I have this in the head of my website: <meta name='<a class="attribute-value">description</a>' content='<a class="attribute-value">Free and Custom IPB Skins for Invision Power Board.</a>'/> Yet when I google 'ipbskinning' it says: Solid Skins. 1We test all our skins in all browsers to insure that they are compatible. This ensures that your users have the best user experience. which is random text from the content of my site. Any idea why this is happening? Thanks a lot all
Technical SEO | | pezza34340 -
Linking to unrelated content
Hi, Just wanted to know, linking to unrelated content will harm the site? I know linking to unrelated content is not good. But wanted to know weather any chances are there or not. I have a site related to health and the other one related to technology. The technology site is too good having PR 6 and very good strong backlinks. And the health related site has very much tough competition, So i wanted to know may be i could link this health site to technology site to get good link from it. Can you suggest me about it. waiting for your replies...
Technical SEO | | Dexter22387874870 -
Blocking AJAX Content from being crawled
Our website has some pages with content shared from a third party provider and we use AJAX as our implementation. We dont want Google to crawl the third party's content but we do want them to crawl and index the rest of the web page. However, In light of Google's recent announcement about more effectively indexing google, I have some concern that we are at risk for that content to be indexed. I have thought about x-robots but have concern about implementing it on the pages because of a potential risk in Google not indexing the whole page. These pages get significant traffic for the website, and I cant risk. Thanks, Phil
Technical SEO | | AU-SEO0