Site been plagiarised - duplicate content
-
Hi,
I look after two websites, one sells commercial mortgages the other sells residential mortgages.
We recently redesigned both sites, and one was moved to a new domain name as we rebranded it from being a trading style of the other brand to being a brand in its own right.
I have recently discovered that one of my most important pages on the residential mortgages site is not in Google's index. I did a bit of poking around with Copyscape and found another broker has copied our page almost word-for-word.
I then used copyscape to find all the other instances of plagiarism on the other broker's site and there are a few! It now looks like they have copied pages from our commercial mortgages site as well.
I think the reason our page has been removed from the index is that we relaunced both these sites with new navigation and consequently new urls. Can anyone back me up on this theory?
I am 100% sure that our page is the original version because we write everything in-house and I check it with copyscape before it gets published, Also the fact that this other broker has copied from several different sites corroborates this view.
Our legal team has written two letters (not sent yet) - one to the broker and the other to the broker's web designer. These letters ask the recipient to remove the copied content within 14 days. If they do remove our content from our site, how do I get Google to reindex our pages, given that Google thinks OUR pages are the copied ones and not the other way around? Does anyone have any experience with this? Or, will it just happen automatically? I have no experience of this scenario!
In the past, where I've found duplicate content like this, I've just rewritten the page, and chalked it up to experience but I don't really want to in this case because, frankly, the copy on these pages is really good! And, I don't think it's fair that someone else could potentially be getting customers that were persuaded by OUR copy.
Any advice would be greatly appreciated.
Thanks,
Amelia
-
Hi David,
I hope you had a good weekend?
Thank you for all your help! I reported them to Google using the link you posted and already the other site's URLs that had copied us have been removed and our pages have been put back in the index.
I have to say I am absolutely astounded that Google responded so quickly!
Yes, that is us on Google + and my personal Google + is here: https://plus.google.com/u/0/+AmeliaVargo/posts/.
Thank you again for your help thus far, and for your kind offer of more help should we need it!
Have a great day,
Amelia
-
Glad I could help. I really hope you get this all sorted out. Good news is, you found the problem and are working to fix it, which is much better than most people would have been able to do. Have high hopes!
"the two pages they've copied are really important sales pages (remortgage and first time buyer) so for us, it's a massive shame. "
There is still a way to promote those pages, just not using Google organic to do so. Modify some of the content, create a press release, promote that page using social networks, and drive interest to that page and your site the old fashioned way. PPC is always an option as well. Remember, there are many ways to get traffic, don't lose hope or the vision.
On a side note, is this your company?
https://plus.google.com/u/0/+TurnkeymortgagesCoUk/postsI can add you to my circles, so if you have any more issues or need additional help just let me know.
-
I just wanted to post up a message to everyone who has helped me with this problem.
First of all, please accept my sincere thanks. I REALLY appreciate everyone's contribution.
Now, I just wanted to tell you all what, as a company, we've decided to do.
- We've written letters to: The company that copied us, their web designer and their host, asking them to remove the copied content within 14 days of the letters.
- We've 'reported' them to Google, via one of the links that David posted (https://support.google.com/legal/troubleshooter/1114905?hl=en)
- We've reported them for scraping, using the link that Paddy posted
Hopefully, this problem will go away, but I hate to think how much business we may have lost as a result - the two pages they've copied are really important sales pages (remortgage and first time buyer) so for us, it's a massive shame.
Best wishes, and I hope you all have a great weekend!
Amelia
-
Thank you David.
-
Once their version is removed/rewritten, resubmit your site to Google in every way that you can.
1. Fetch as Google
2. Change sitemap created dates to current day
3. Change crawl frequency in sitemap to daily
4. Check for proper 301 redirects from old pages, when you moved/modified the site to separate branding.
5. Submit the URL in question to Google, and letting them know that someone has copied your site's content. They should be able to see that your was created first.Here are a few links to help:
https://www.google.com/webmasters/tools/dmca-notice <<< start there
https://support.google.com/legal/troubleshooter/1114905?hl=en
http://blog.kissmetrics.com/find-remove-stolen-content/
http://www.orclage.com/report-remove-stolen-duplicate-content-google/
-
Thank you Paddy! Much appreciated, and thank you for helping me again!
-
Ahh, good one.
-
Don't forget about this:
https://docs.google.com/forms/d/1Pw1KVOVRyr4a7ezj_6SHghnX1Y6bp1SOVmy60QjkF0Y/viewform
-
Thank you, you've helped me no end.
Have a great weekend
-
It really depends on the web host whether they will follow it or not. Some that are soley based in the UK might not. If they have US based servers or the site is hosted in the US more than likely they will. It is worth a shot though, I try to rattle as many cages as possible. Here is a little info on filing them in the UK https://www.teneric.co.uk/marketing/copyright-infringement.html
-
Hi Lesley,
Yes, I redirected everything using 301 redirects - page to page. I also used the change of address tool in webmaster tools for the site that changed domains.
I don't know if using DMCA will be appropriate - isn't that a US-only thing or can site owners in the UK use it too? If I can, I will use it.
Thank you for responding - I really do appreciate your help.
Best wishes,
Amelia
-
After they drop out of the searches google will index your site as a the canonical site with that content on it. So that part happens manually. Also, when you relaunched, did you redirect everything from the old site? That helps preserve link juice and at the same time gives search engines a pointer that the address of a page has changed to this new address.
One thing I would suggest is having a DMCA take down notice draft and sent to the host as well. If the other people you send letters to tell you to go pound sand, normally the host does not.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Cached version of my site is not showing content?
Hi mozzers, I am a bit worried since I looked a cache version of my site and somehow content is partially showing up and navigation has completely disappeared. Where could this come from? What should I be doing? Thanks!
Intermediate & Advanced SEO | | Taysir0 -
Possible duplicate content issue
Hi, Here is a rather detailed overview of our problem, any feedback / suggestions is most welcome. We currently have 6 sites targeting the various markets (countries) we operate in all websites are on one wordpress install but are separate sites in a multisite network, content and structure is pretty much the same barring a few regional differences. The UK site has held a pretty strong position in search engines the past few years. Here is where we have the problem. Our strongest page (from an organic point of view) has dropped off the search results completely for Google.co.uk, we've picked this up through a drop in search visibility in SEMRush, and confirmed this by looking at our organic landing page traffic in Google Analytics and Search Analytics in Search Console. Here are a few of the assumptions we've made and things we've checked: Checked for any Crawl or technical issues, nothing serious found Bad backlinks, no new spammy backlinks Geotarggetting, this was fine for the UK site, however the US site a .com (not a cctld) was not set to the US (we suspect this to be the issue, but more below) On-site issues, nothing wrong here - the page was edited recently which coincided with the drop in traffic (more below), but these changes did not impact things such as title, h1, url or body content - we replaced some call to action blocks from a custom one to one that was built into the framework (Div) Manual or algorithmic penalties: Nothing reported by search console HTTPs change: We did transition over to http at the start of june. The sites are not too big (around 6K pages) and all redirects were put in place. Here is what we suspect has happened, the https change triggered google to re-crawl and reindex the whole site (we anticipated this), during this process, an edit was made to the key page, and through some technical fault the page title was changed to match the US version of the page, and because geotargetting was not turned on for the US site, Google filtered out the duplicate content page on the UK site, there by dropping it off the index. What further contributes to this theory is that a search of Google.co.uk returns the US version of the page. With country targeting on (ie only return pages from the UK) that UK version of the page is not returned. Also a site: query from google.co.uk DOES return the Uk version of that page, but with the old US title. All these factors leads me to believe that its a duplicate content filter issue due to incorrect geo-targetting - what does surprise me is that the co.uk site has much more search equity than the US site, so it was odd that it choose to filter out the UK version of the page. What we have done to counter this is as follows: Turned on Geo targeting for US site Ensured that the title of the UK page says UK and not US Edited both pages to trigger a last modified date and so the 2 pages share less similarities Recreated a site map and resubmitted to Google Re-crawled and requested a re-index of the whole site Fixed a few of the smaller issues If our theory is right and our actions do help, I believe its now a waiting game for Google to re-crawl and reindex. Unfortunately, Search Console is still only showing data from a few days ago, so its hard to tell if there has been any changes in the index. I am happy to wait it out, but you can appreciate that some of snr management are very nervous given the impact of loosing this page and are keen to get a second opinion on the matter. Does the Moz Community have any further ideas or insights on how we can speed up the indexing of the site? Kind regards, Jason
Intermediate & Advanced SEO | | Clickmetrics0 -
Duplicate content hidden behind tabs
Just looking at an ecommerce website and they've hidden their product page's duplicate content behind tabs on the product pages - not on purpose, I might add. Is this a legitimate way to hide duplicate content, now that Google has lowered the importance and crawlability of content hidden behind tabs? Is this a legitimate tactic to tackle duplicate content? Your thoughts would be welcome. Thanks, Luke
Intermediate & Advanced SEO | | McTaggart0 -
Duplicate content in external domains
Hi,
Intermediate & Advanced SEO | | teconsite
I have been asking about this case before, but now my question is different.
We have a new school that offers courses and programs . Its website is quite new (just a five months old) It is very common between these schools to publish the courses and programs in training portals to promote those courses and to increase the visibility of them. As the website is really new, I found when I was doing the technical audit, that when I googled a text snipped from the site, the new school website was being omitted, and instead, the course portals are being shown. Of course, I know that the best recommendation would be to create a different content for that purpose, but I would like to explore if there is more options. Most of those portals doesn't allow to place a link to the website in the content and not to mention canonical. Of course most of them are older than the new website and their authority is higher. so,... with this situation, I think the only solution is to create a different content for the website and for the portals.
I was thinking that maybe, If we create the content first in the new website, send it to the index, and wait for google to index it, and then send the content to the portals, maybe we would have more opportunites to not be ommited by Google in search results. What do you think? Thank you!0 -
Product Page on Eccomerce Site ranking very poorly - Unique Product description but duplicate content on other tabs.
Hi All, I have a query regarding my Product pages on my eCommerce site. I have unique Product descriptions but some of the other page content on the other tabs i.e Hire Terms , Delivery , About the Hire Company - Are duplicated across ALL my products. Is that okay or how should I deal with them ? See example url of one of my products below below - http://goo.gl/aSFPqP My products currently rank very badly... 200 + so Any advice would be greatly appreciated thanks Peter
Intermediate & Advanced SEO | | PeteC120 -
Noindexing Duplicate (non-unique) Content
When "noindex" is added to a page, does this ensure Google does not count page as part of their analysis of unique vs duplicate content ratio on a website? Example: I have a real estate business and I have noindex on MLS pages. However, is there a chance that even though Google does not index these pages, Google will still see those pages and think "ah, these are duplicate MLS pages, we are going to let those pages drag down value of entire site and lower ranking of even the unique pages". I like to just use "noindex, follow" on those MLS pages, but would it be safer to add pages to robots.txt as well and that should - in theory - increase likelihood Google will not see such MLS pages as duplicate content on my website? On another note: I had these MLS pages indexed and 3-4 weeks ago added "noindex, follow". However, still all indexed and no signs Google is noindexing yet.....
Intermediate & Advanced SEO | | khi50 -
Syndicating duplicate content descriptions - Can these be canonicalised?
Hi there, I have a site that contains descriptions of accommodation and we also use this content to syndicate to our partner sites. They then use this content to fill their descriptions on the same accommodation locations. I have looked at copyscape and Google and this does appear as duplicate content across these partnered sites. I do understand as well that certain kinds of content will not impact Google's duplication issue such as locations, addresses, opening times those kind of things, but would actual descriptions of a location around 250 words long be seen and penalised as duplicate content? Also is there a possible way to canonicalise this content so that Google can see it relates back to our original site? The only other way I can think of getting round a duplicate content issue like this is ordering the external sites to use tags like blockquotes and cite tags around the content.
Intermediate & Advanced SEO | | MalcolmGibb0 -
How should we handle syndicated content on a partner site?
Say we have a subdomain with resources (resources.site.com) and a partner site (partner.com) and have an agreement to share content (I know - this isn't ideal but it's what I've got to work with). Please comment on the following: the use of cross-domain canonicals on "shared" articles an intro and/or conclusion paragraph that is unique on the site that re-publishes that could say something like "our partner over at resources.site.com recently published the following report ... yada, yada....." other meta tags to let Google know that we are not scraping, e.g. author tags any other steps we can take to ensure neither site gets "dinged" by the search engines. Thanks a bunch in advance! AK26
Intermediate & Advanced SEO | | akim260