Are there discrepancies between GWT and SEOMoz?
-
In our keyword rank tracking report, we've dominated a keyword in Google and have secured the slot for years. All evidence points in this direction. In Google Webmaster Tools, however, this particular keyword averages a rank of 6.5. Is anyone else experience these kinds of discrepancies? What is your take on it?
-
That makes MUCH more sense. I'll Google it a bit and see what others have to say. Thanks Keri!
-
If I remember correctly, Google is doing an average of all of the pages that rank for that term. So if your privacy policy ranks in 20th position for your company name, that gets factored in. I think that's how it works, but could be wrong. It should be a place to start looking for information though.
-
Everything is way off. Even our company name according to GWT is ranking at an average of 4.5. Does GWT also include universal results in their ranking report?
-
The scope of data Google has ( i would not expect them to show all that to user through GWT ) and OSE has are not the same so there will be some discrepancies, here and there .
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
2.3 million 404s in GWT - learn to live with 'em?
So I’m working on optimizing a directory site. Total size: 12.5 million pages in the XML sitemap. This is orders of magnitude larger than any site I’ve ever worked on – heck, every other site I’ve ever worked on combined would be a rounding error compared to this. Before I was hired, the company brought in an outside consultant to iron out some of the technical issues on the site. To his credit, he was worth the money: indexation and organic Google traffic have steadily increased over the last six months. However, some issues remain. The company has access to a quality (i.e. paid) source of data for directory listing pages, but the last time the data was refreshed some months back, it threw 1.8 million 404s in GWT. That has since started to grow progressively higher; now we have 2.3 million 404s in GWT. Based on what I’ve been able to determine, links on this particular site relative to the data feed are broken generally due to one of two reasons: the page just doesn’t exist anymore (i.e. wasn’t found in the data refresh, so the page was simply deleted), or the URL had to change due to some technical issue (page still exists, just now under a different link). With other sites I’ve worked on, 404s aren’t that big a deal: set up a 301 redirect in htaccess and problem solved. In this instance, setting up that many 301 redirects, even if it could somehow be automated, just isn’t an option due to the potential bloat in the htaccess file. Based on what I’ve read here and here, 404s in and of themselves don’t really hurt the site indexation or ranking. And the more I consider it, the really big sites – the Amazons and eBays of the world – have to contend with broken links all the time due to product pages coming and going. Bottom line, it looks like if we really want to refresh the data on the site on a regular basis – and I believe that is priority one if we want the bot to come back more frequently – we’ll just have to put up with broken links on the site on a more regular basis. So here’s where my thought process is leading: Go ahead and refresh the data. Make sure the XML sitemaps are refreshed as well – hopefully this will help the site stay current in the index. Keep an eye on broken links in GWT. Implement 301s for really important pages (i.e. content-rich stuff that is really mission-critical). Otherwise, just learn to live with a certain number of 404s being reported in GWT on more or less an ongoing basis. Watch the overall trend of 404s in GWT. At least make sure they don’t increase. Hopefully, if we can make sure that the sitemap is updated when we refresh the data, the 404s reported will decrease over time. We do have an issue with the site creating some weird pages with content that lives within tabs on specific pages. Once we can clamp down on those and a few other technical issues, I think keeping the data refreshed should help with our indexation and crawl rates. Thoughts? If you think I’m off base, please set me straight. 🙂
Intermediate & Advanced SEO | | ufmedia0 -
Discrepancy in keyword ranking from webmasters and actual ranking.
I have been tracking ranks of some keywords important to my business since the last 2 months. Recently I have observed that, for one of my keywords, google webmasters is giving the avg position as 8 but when i search in google it comes in the 6th page. I know that webmasters tools gives the average position but i do not think there will be such big difference in the ranks. Please help.Thanks.
Intermediate & Advanced SEO | | seomoz12320 -
Unnatural Inbound Links Warning in GWT
Hi all, A bit of a long questions so apologies in advance but please bear with me... My client has received an 'Unnatural Inbound Links' warning and it is now my task to try and resolve through a process of; Highlighting the unnatural links Requesting that the links be removed (via webmaster requests) Possibly using the Disavow Tool Submitting a Reconsideration Request So I downloaded my clients link profile from both OSE and GWT in CSV format and compared - the amount of links returned was considerably more in GWT than it was in OSE...? So I set about going through the links, first filtering into order so that I could see blocks of links from the same URL - I highlighted in colours; Red - Definitely need to be removed Orange - Suspect, need to investigate further Yellow - Seem to be ok but may revisit Green - Happy with the link, no further action So to my question which relates to, is it 'black & white' - is it a case of 'good link v 'bad link' or could there be some middle ground? (am I making this process even more confusing than it actually is?) As an example, here are some 'Orange' URL's; http://www.24searchengines.com/ (not exact URL as it goes to the travel section which is my clients niche) - this to me looks spammy and I would normally 'paint it red' and look to remove, however, when I go to the 'contact us' page; (http://www.24searchengines.com/texis/open/allthru?area=contactus) and follow the link to remove from directory, it takes me here; http://www.dmoz.org/docs/en/help/update.html DMOZ??? My clients has a 'whole heap' of these type of links; http://www.25searchengines.com/ http://www.26searchengines.com/ http://www.27searchengines.com/ http://www.28searchengines.com/ ...and many many more!! Here is another example; http://foodys.eu/ http://foodys.eu/2007/01/04/the-smoke-ring-bbq-community/ ...plus many more... My client is in the 'cruise niche' and as there is a 'cruise' section on the site I'm not sure whether this constitutes a good, bad or indifferent link! Finally, prior to me working with this client (1 month) they moved their site from a .co.uk to a .com domain and redirected all links from the .co.uk to the .com (according to GWT, over 16k have been redirected) - a lot of these 'spammy' links were to the .co.uk and have thus been redirected, should I even consider removing the redirection or will that have severe consequences? Apologies for the long (long) post, I know I'm heading in the right direction but some assurance wouldn't go amiss! 🙂 Many thanks Andy <colgroup><col width="1317"></colgroup>
Intermediate & Advanced SEO | | TomKing
| |0 -
Severe health issues are found on your site. - Check site health (GWT)
Hi, We run a Magento website - When i log in to Google Webmaster Tools, I am getting this message: Severe health issues are found on your site. - <a class="GNHMM2RBFH">Check site health
Intermediate & Advanced SEO | | bjs2010
</a>Is robots.txt blocking important pages? Some important page is blocked by robots.txt. Now, this is the weird part - the page being blocked is the admin page of magento - under
www.domain.com/index.php/admin/etc..... Now, this message just wont go away - its been there for days now - so why does Google think this is an "important page"? It doesnt normally complain if you block other parts of the site ?? Any ideas? THanks0 -
SEOMoz Paid Directories List.... Are paid links always bad?
We all know paid links are bad.... so it begs the question.... why does the SEOMoz directory list have so many directories that only submit businesses for a fee.... i.e. a paid link. Are they worth paying for or not? Interested to hear peoples thoughts and experiences with some of the directories listed on the SEOMoz directory list....
Intermediate & Advanced SEO | | JohnW-UK0 -
How accurate are the index figures in GWT?
I've been looking at a site in GWT and the number of indexed urls is very low when compared with the number or submitted urls on the xml sitemaps. The site has several stores which are all submitted using different sitemaps. When you perform a search in Google, eg site:domain.com/store1 site:domain.com/store2 site:domain.com/store3 The results are similar to the webmaster urls. However, looking in the analytics for landing pages used for organic traffic from Google shows a much higher number of pages. If these pages aren't indexed as reported in GMT, how could they be found in the results and be recorded as landing pages?
Intermediate & Advanced SEO | | edwardlewis0 -
How can I export SEOmoz ranking reports to google spreadsheet
How can I export SEOmoz website rankings to Google Spreadsheet? I have applied other SEOmoz API's and Google Spreadsheet combos effectively but cannot find anything online for this. I would like to display current ranking and ranking history for specific keywords in Google Spreadsheet and have them update automatically using the SEOmoz API.
Intermediate & Advanced SEO | | Michael_Rock0 -
SeoMoz Crawler Shuts Down The Website Completely
Recently I have switched servers and was very happy about the outcome. However, every friday my site shuts down (not very cool if you are getting 700 unique visitors per day). Naturally I was very worried and digged deep to see what is causing it. Unfortunately, the direct answer was that is was coming from "rogerbot". (see sample below) Today (aug 5) Same thing happened but this time it was off for about 7 hours which did a lot of damage in terms of seo. I am inclined to shut down the seomoz service if I can't resolve this immediately. I guess my question is would there be a possibility to make sure this doesn't happen or time out like that because of roger bot. Please let me know if anyone has answer for this. I use your service a lot and I really need it. Here is what caused it from these error lines: 216.244.72.12 - - [29/Jul/2011:09:10:39 -0700] "GET /pregnancy/14-weeks-pregnant/ HTTP/1.1" 200 354 "-" "Mozilla/5.0 (compatible; rogerBot/1.0; UrlCrawler; http://www.seomoz.org/dp/rogerbot)" 216.244.72.11 - - [29/Jul/2011:09:10:37 -0700] "GET /pregnancy/17-weeks-pregnant/ HTTP/1.1" 200 51582 "-" "Mozilla/5.0 (compatible; rogerBot/1.0; UrlCrawler; http://www.seomoz.org/dp/rogerbot)"
Intermediate & Advanced SEO | | Jury0