Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
What can I do if my reconsideration request is rejected?
-
Last week I received an unnatural link warning from Google. Sad times.
I followed the guidelines and reviewed all my inbound links for the last 3 months. All 5000 of them! Along with several genuine ones from trusted sites like BBC, Guardian and Telegraph there was a load of spam. About 2800 of them were junk. As we don't employ any SEO agency and don't buy links (we don't even buy adwords!) I know that all of this spam is generated by spam bots and site scrapers copying our content.
As the bad links have not been created by us and there are 2800 of them I cannot hope to get them removed. There are no 'contact us' pages on these Russian spam directories and Indian scraper sites. And as for the 'adult book marking website' who have linked to us over 1000 times, well I couldn't even contact that site in company time if I wanted to! As a result i did my manual review all day, made a list of 2800 bad links and disavowed them.
I followed this up with a reconsideration request to tell Google what I'd done but a week later this has been rejected "We've reviewed your site and we still see links to your site that violate our quality guidelines." As these links are beyond my control and I've tried to disavow them is there anything more to be done?
Cheers
Steve
-
Tom has given you good advice. I'll put in my 2 cents' worth as well.
There are 3 main reasons for a site to fail at reconsideration:
1. Not enough links were assessed by the site owner to be unnatural.
2. Not enough effort was put into removing links and documenting that to Google.
3. Improper use of the disavow tool.
In most cases #1 is the main cause. Almost every time I do a reconsideration request my client is surprised at what kind of links are considered unnatural. From what I have seen, Google is usually pretty good at figuring out whether you have been manually trying to manipulate the SERPS or whether links are just spam bot type of links.
Here are a few things to consider:
Are you being COMPLETELY honest with yourself about the spammy links you are seeing? How did Russian and porn sites end up linking to you? Most sites don't just get those by accident. Sometimes this can happen when sites use linkbuilding companies that use automated methods to build links. Even still, do all you can to address those links, and then for the ones that you can't get removed, document your efforts, show Google and then disavow them.
Even if these are foreign language sites, many of them will have whois emails that you can contact.
Are you ABSOLUTELY sure that your good links are truly natural? Just because they are from news sources is not a good enough reason. Have you read all the interflora stuff recently? They had a pile of links from advertorials (amongst other things) that now need to be cleaned up.
-
Hi Steve
If Google is saying there are still a few more links, then it might be an idea to manually review a few others that you haven't disavowed. I find the LinkDetox tool very useful for this. It's free with a tweet and will tell you if a link from a site is toxic (the site is deindexed) or if it's suspicious (and why it's suspicious). You still need to use your own judgement on these, but it might help you to find the extra links you're talking about.
However, there is a chance you have gone and disavowed every bad link, but still got the rejection. In this case, I'd keep trying but make your reconsideration request more detailed. Create an excel sheet and list the bad URLs and/or domains and give a reason explaining why you think they're bad links. Then provide information on how you found their contact details. If there are no contact us pages, check the whois registrar's email. After that, say when you contacted them (give a sample of your letter to them too), and if they replied, along with a follow up date if you got silence. If there are no details in the whois, explicitly mention that there are no contact details and so you have proceeded straight to disavowing.
Then list the URLs you've disavowed (upload the .txt file with your reconsideration email). You've now told Google that you've found bad links, why you think their bad (also include how you discovered them), that you've contacted the webmaster on numerous occasions and, if no removal was made, you've disavowed as a last resort. This is a very thorough process and uses the disavow tool in the way that Google wants us to - as a last resort to an unresponsive or anonymous webmaster.
Please forgive me if you've already done all this and it seems like repetition. I only mention it because I've found it's best to be as thorough as possible with Google in these situations. Remember, a reconsideration request is manual and if they see that you've gone through all this effort to be reinstated, you've got a better chance of being approved.
Keep trying, mate. It can be disheartening, but if you think it's worth the time and effort, then keep going for it. I would bear in mind the alternatives, however, such as starting fresh on a new domain. If you find yourself going round the bend with endless reconsiderations, sometimes your time, effort and expertise can be better put elsewhere.
All the best!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Why can't google mobile friendly test access my website?
getting the following error when trying to use google mobile friendly tool: "page cannot be reached. This could be because the page is unavailable or blocked by robots.txt" I don't have anything blocked by robots.txt or robots tag. i also manage to render my pages on google search console's fetch and render....so what can be the reason that the tool can't access my website? Also...the mobile usability report on the search console works but reports very little, and the google speed test also doesnt work... Any ideas to what is the reason and how to fix this? LEARN MOREDetailsUser agentGooglebot smartphone
Technical SEO | | Nadav_W0 -
How can I stop a tracking link from being indexed while still passing link equity?
I have a marketing campaign landing page and it uses a tracking URL to track clicks. The tracking links look something like this: http://this-is-the-origin-url.com/clkn/http/destination-url.com/ The problem is that Google is indexing these links as pages in the SERPs. Of course when they get indexed and then clicked, they show a 400 error because the /clkn/ link doesn't represent an actual page with content on it. The tracking link is set up to instantly 301 redirect to http://destination-url.com. Right now my dev team has blocked these links from crawlers by adding Disallow: /clkn/ in the robots.txt file, however, this blocks the flow of link equity to the destination page. How can I stop these links from being indexed without blocking the flow of link equity to the destination URL?
Technical SEO | | UnbounceVan0 -
Can you use multiple videos without sacrificing load times?
We're using a lot of videos on our new website (www.4com.co.uk), but our immediate discovery has been that this has a negative impact on load times. We use a third party (Vidyard) to host our videos but we also tried YouTube and didn't see any difference. I was wondering if there's a way of using multiple videos without seeing this load speed issue or whether we just need to go with a different approach. Thanks all, appreciate any guidance! Matt
Technical SEO | | MattWatts1 -
What punctuation can you use in meta tags? Are there any Google does not like?
So I know you can use dashes and | in meta tags, but can anyone tell me what other punctuation you can use? Also, it'd be great to know what punctuation you can't use. Thanks!
Technical SEO | | Trevorneo1 -
How can I Style Long "List Posts" in Wordpress?
Hi All, I have been working on a list-post which spans over 100 items. Each item on the list has a quick blurb to explain it, an image and a few resource links. I am trying to find an attractive way to present this long list post in Wordpress. I have seen several sites with long list posts however; they place their items one on top of the other which yields a VERY long page and the end user has to do a lot of scrolling. Others turn their lists into slideshows, but I have no data on how slides perform against 10-mile-long-lists which load in 1 page. I would like to do something similar to what List25.com does as they present about 5-10 items per page and they seem to have pagination. The pagination part I understand however; is there a shortcode plugin to format lists in an attractive way just like list25?
Technical SEO | | IvanC0 -
Can I mark up breadcrumbs without showing them? (responsive design)
I am working on a site that has responsive design. We use faceted search for the desktop version but implemented a style of breadcrumbs for the mobile version as sidebars take up too much screen real estate. On the desktop design we are putting a display:none in front of the breadcrumbs. If we mark up those breadcrumbs and they are behind a display none, can we still get the rich snippets? Will Google see this is cloaking? In follow up, is there a way to markup breadcrumbs in the or somewhere else that is constant?
Technical SEO | | MarloSchneider0 -
Can I Disallow Faceted Nav URLs - Robots.txt
I have been disallowing /*? So I know that works without affecting crawling. I am wondering if I can disallow the faceted nav urls. So disallow: /category.html/? /category2.html/? /category3.html/*? To prevent the price faceted url from being cached: /category.html?price=1%2C1000
Technical SEO | | tylerfraser
and
/category.html?price=1%2C1000&product_material=88 Thanks!0 -
What can I do about missing Meta Description for category pagest etc.?
On all my campaigns I'm returning high levels of 'Missing Meta Description Tags'. The problem with fixing this is they're all for category, tag and author pages. Is there a way to add a meta description to these pages (there are hundreds) or will it not really have any ranking effect?
Technical SEO | | SiliconBeachTraining0