How to find out if I have been penalized?
-
I have launched a new website beginning January this year and have seen slowly more and more traffic coming from google to the website until the 20th of March where suddenly there are no more visitors from the google search engine. The only traffic left is from google images, social networks or other search engines. Without visitors from google search this reduces our overall traffic by ~66%.
I can't easily find anymore our website in the search results of google by using terms which we usually ranked quite well. Nevertheless, the website is still indexed as I can find it using the "site:" search query. In google webmaster tools there are no messages and we have only been doing a bit of link building on website and blog directories (nothing excessive and nothing paid neither).
Is there any way to find out if google penalized my website? I guess it has... and what would be the best thing to do right now?
The website is hellasholiday (dot) com
Thanks in advance for your idea and suggestions
-
I am not a fan of CMS, i realize there are pros and cons, but when you try to do too much and be all things to all people you tend to have a lot of compromises.
There is one other reason i dont like to use robots,txt, i remeber Matt Cutts saying that it is a spam signal because they can not see what you are hiding, not that it is going to get you flaged by itself, but with other signals it can. If i remember correctly he was talking about hiding malware in scripts blocked by robots.
If you are interested, the best CMS for SEO i had found was Orchard CMS but even that has some silly errors, it puts more then one H1 tag in pages, but is still the best solution I have looked at. It is more customizable via code.
-
After having read your post and all the linked articles you have recommended I understand the issue and have adapted the robots.txt accordingly. Basically only leaving one single Disallow for the WordPress plugins. I hope this will help but I suppose I will see this in the next few days...
Now regarding WordPress I would suggest them to adapt their documentation as it is really misleading. Also I think they should implement all these noindex meta tags where necessary natively into wordpress and not by having to use a plugin for that, but this is another story.
-
Wordpress do many things that are not recommened, and blocking by robots is not recomened, what they are suggesting is a extream messure to solve the softewares problems. there are better ways to solve duplicate content without giveing away your link juice
Read this section "WordPress Robots.txt blocking Search results and Feeds"
on this page http://yoast.com/example-robots-txt-wordpress/
These plug-ins like yoast and word press itself, do not produse very good results. I have crawled many wordpress sites and they all have the same old problems many caused by the yoast plugin.
What google is refereing to in the link, is not getting pages of little value into their index, this is for their advantage not yours.
Its quite simple, if you block a page, the links pointing to that page waste their link juice, if you dont, or at least allow follow with a meta tag, you will get the link juice back.
See this article where Dr Pets calls it an extream messure, search for robots.txt you will see many comments refering to my point http://www.seomoz.org/blog/duplicate-content-in-a-post-panda-world
See Dr pets comments here http://www.seomoz.org/blog/serious-robotstxt-misuse-high-impact-solutions
-
I thought it would be no use for google to index and cache small icons, logos and cached resized images which have no meaningful name or so. So now I have at least removed the Disallow for these but for WordPress blog I want to keep the Disallow rules as recommended by WordPress itself for SEO purposes as documented here http://codex.wordpress.org/Search_Engine_Optimization_for_WordPress#Robots.txt_Optimization assuming they know what they are speaking about.
Anyhow I don't have the feeling this is really the problem why my website doesn't show up anymore in the google search engine results...
-
The question should be why block them?
its like cutting off your hand, because you have a splinter.
If duplicate content is a problem, then you can (in order of prefrance) fix it, use a canonical, a noindex,follow meta tag, but not robots
-
Many thanks Alan for your answer!
Regarding the robots.txt, basically I just would like to block/disallow some cached images and small icons/pictures from the website as well as some stuff for the associated WordPress blog which is also host on the same website. For the blog I am disallowing the admin pages, feeds, comments, trackbacks, content theme files etc. Here wold be the complete list just in case:
Disallow /wp-admin
Disallow: /wp-includes
Disallow: /wp-content/plugins
Disallow: /wp-content/cache
Disallow: /wp-content/themes
Disallow: /trackback
Disallow: /feed
Disallow: /comments
Disallow: /category//
Disallow: /*/trackback
Disallow: /*/feed
Disallow: /*/comments
Disallow: /?
Disallow: /*?
So maybe I should change my question to "what URLs should I disallow for a WordPress blog?"
Also where can I see all the pages which are blocked by my robots.txt file?
-
You can ask for reconsideration from google though webmaster tools. But since you have no warnings and you are still in the index, i have doubts that you have been flagegd manualaly, but you may have been algorthmicly.
I notived that you have blokced hundreds of pages with robots.txt, thios had led to thousonds of links pointing to pages that are not indexed, this means these links are puiring away link juice into nowhere.
You should not use robots text to block pages that are linked to, its a waste of valuable link juice.
if you must no-index the pages, use a meta noindex,follow tag, this way you will get most of the link juice back though the pages outlinks
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Ranking subpage for main keyword instead of homepage – somehow I feel like our homepage is penalized
Hello guys, I hope this question isn't too broad – but I run out of ideas and I feel like there is some technical issue with our site which sees us penalized: Search: German Google
Technical SEO | | Maggiathor
Website: www.elementspure-hotel.de
Keyword: Hotel Bremen Right now we're ranking between 40 and 50 with the subpage /zimmer. a lot of sites with lesser DA scores and a lot more SEO Issues are ranked higher. The homepage is optimized for hotel bremen, yet /zimmer page ranks first, which I just don't understand
(Page Authority is much higher, Keyword density)
– I've run the site through a lot of tools and there aren't any severe issues Search Console doesn't say anything in regards to errors My suggestions which I can't confirm:
– There are some redirection issues (http/https/www/nonww/)
– Backlink issues (a lot of backlinks are from www.elements-pure.de)
– Over optimization penalties? Tank you! Magnus0 -
Best tool to find Related keywords with a Keyword provided.
Hi, Best tool to find Related keywords with a Keyword provided. Basically i want to give a keyword and find all related keywords we can use to write articles. Also any way we can find what keyword a page is getting traffic based on? Thanks
Technical SEO | | skandlikp90 -
Find all old links from a site to 301
We worked on this site a while ago - http://www.electric-heatingsupplies.co.uk/ Whilst we did a big 301 redirect exercise, I wanted to check that we "got" all of them. Is there a historical way I can check all the old indexed links to make sure they correlate to the new links? Thanks!!
Technical SEO | | lauratagdigital0 -
What might make Bing.bot find a URL that looks like this on our site?
I have been doing something Richard Baxter recently suggested and reviewing our server logs. I have found an oddity that hopefully some of you smart Mozzers can help me figure out. Here is the line from the server log (there are many more like this): 157.55.32.166 - - [04/Mar/2013:08:00:59 -0800] "GET /StoreFront/category/www.ccisolutions.com/StoreFront/category/shure-se-earphones HTTP/1.1" 200 94133 "-" "Mozilla/5.0 (compatible; bingbot/2.0; +http://www.bing.com/bingbot.htm)" "-" See how the www.ccisolutions.com appears after /StoreFront/category/ ? We used to see weird URLs reported in GWT that looked like this, but ever since we fixed our canonical tags to be absolute instead of relative URLs, they no longer appeared in our Webmaster Tools reports. However, it seems there is still a problem. Where/how could Bingbot be seeing URLs configured this way? Could it be a server issue, or is it most likely a data problem? Thanks in advance! Dana P.S. Could this be resulting from our massive use of relative URLs all over the site?
Technical SEO | | danatanseo0 -
404 error - but I can't find any broken links on the referrer pages
Hi, My crawl has diagnosed a client's site with eight 404 errors. In my CSV download of the crawl, I have checked the source code of the 'referrer' pages, but can't find where the link to the 404 error page is. Could there be another reason for getting 404 errors? Thanks for your help. Katharine.
Technical SEO | | PooleyK0 -
How can I best find out which URLs from large sitemaps aren't indexed?
I have about a dozen sitemaps with a total of just over 300,000 urls in them. These have been carefully created to only select the content that I feel is above a certain threshold. However, Google says they have only indexed 230,000 of these urls. Now I'm wondering, how can I best go about working out which URLs they haven't indexed? No errors are showing in WMT related to these pages. I can obviously manually start hitting it, but surely there's a better way?
Technical SEO | | rango0 -
Very Weird Type of Penguin Penalization
One of my client's sites has a bunch of bad links from blog networks with exact-match anchor text. Since Penguin, they have been completely removed from Google for that keyword. But here's the weird part: It's only the homepage that has been removed, and only for that keyword. If I put other keywords into Google, our homepage comes up. So the site hasn't been banned, and that page hasn't even been banned because it still comes up with all of our other keywords. It's only when you put in the keyword that has all the anchor text that the homepage doesn't come up anywhere. (I went all the way to the end). Has this happened to anyone else, and does it warrant a re-inclusion request since the site and even that page haven't technically been banned?
Technical SEO | | UnderRugSwept0 -
Parameter handling (where to find all parameters to handle)?
Google recently said they updated their parameter handling, but I was wondering what is the best way to know all of the parameters that need "handling"? Will Google Webmaster find them? Should the company know based on what is on their site? Thanks!
Technical SEO | | nicole.healthline0