3724 pages submitted, 3591 indexed
-
You probably know what I mean, the report in Google Webmaster Tools > Sitemaps.
So how do I locate the pages that are NOT indexed?
Thanks,
Ben
-
Thanks.
-
...and check the last reply in this http://moz.com/community/q/how-to-determine-which-pages-are-not-indexed. I have not tried but it looks promising.
-
This may go some way to answering your question - http://support.google.com/webmasters/bin/answer.py?hl=en&answer=2642366
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should I implement 301 Redirects vs 410 in Removing Product Pages?
I manage an ecommerce site and had a question about 301 redirects vs 410 when removing product pages. What we have in place now is a 410 because we have removed products that are no longer available and the content itself no longer exists. In all my research, SEO Best Practices is to have a 301 in place. Should we replace our 410 with a 301 redirect or keep it?
Industry News | | UAScrubs0 -
Dex Landing Pages!
Hello Moz-ers! Have a question for the community that I'd like some feedback on, here is the situation My client owns a local business, which has a website that we built for them last year. Rankings and traffic have been great, and they're doing well in a competitive market. This client also works with Dex Media, who manages some of the marketing for this business. In this effort Dex has purchased a few variations of the branded domain names, and has placed landing pages at these URL's with basic information, phone number, etc. Seems like a standard page for one of these sites. These sites have different phone numbers, used for tracking leads from those particular pages. To me, these seem totally unnecessary and also as though it could be something pulling visitors away from the main website, which details all of their services, etc. It also seems like it could pretty easily confuse or frustrate potential visitors with the 4 similar domain names, each with different numbers of the same business. I told the client that I thought this might be the case, and at the very least they should ask Dex about acquiring ownership of those extra domains containing their brand in addition to seeing if they would be able to just redirect those URLs to the main domain. They responded by saying that Google doesn't care about any of this, citing this article: https://support.google.com/adwordspolicy/answer/2643759?hl=en# To me, that support link appears to not address the issue at hand, as it focuses mainly on proxy URL's being used giving the example of proxy.example.com Can anyone weigh in here with an opinion? Is there any reason to think that these extra domains would not have a potential negative impact on client traffic/rank for the main domain? Thanks! Tyler
Industry News | | kbaltzell0 -
Can't seem to submit my xml sitemap to Baidu, can someone help?
Hello Mozers and SEM's, I'm sure all of you know and understand the importance of uploading your sitemap to search engines. Once search engine in particular is Baidu. For some reason, i can't find a way to submit my link www.mysite.com/sitemap.xml. For Google, Bing, and Yandex it was easy, but Baidu is giving me problems. Don't tell me to "wait until they crawl your site", i have over 1000+ pages of unique content that those other 3 search engines found because of my sitemap; Baidu can't be that slow. Their robots only found 147 pages. 😞 If you take a look at the image attachment, you can see why I'm stuck, i can't read chinese! (and it's an image, so i couldn't translate it). Has anybody had any luck submitting their .xml link into Baidu? Can someone walk me through it? Let me know! Shawn xILP69w
Industry News | | Shawn1240 -
100's of versions of the same page. Is rel=canonical the solution???
Hi, I am currently working with an eCommerce site that has a goofy set up for their contact form. Basically, their are hundreds of "contact us" pages that look exactly the same but have different URLs and are used to help the store owner determine which product the user contacted them about. So almost every product has it's own "contact us" URL. The obvious solution is to do away with this set up but if that is not an option, would a rel=canonical tag linked back to the actually "contact us" page be a possible solution? Or is the canonical tag only used to show the difference between www vs non-www? Thanks!
Industry News | | RossFruin0 -
Get Google To Crawl More Pages Faster on my Site
We opened our database of about 10 million businesses to be crawled by Google. Since Wednesday, Google has crawled and indexed about 2,000 pages. Google is crawling us at about 1,000 pages a day now. We need to substantially increase this amount. Is it possible to get Google to crawl our sites at a quicker rate?
Industry News | | Intergen0 -
Two Industry giant has occupied the Google SE First Page.
I have seen just two websites on Google SE first page one with 4 and second one is with 6 internal pages on Coupon related question/niche . What it means Is this a Google Dance or temporary fluctuation. ? Why Google don't consider other sites to show on first page? Many pages from askvila.amazon, yahoo answers and facebook pages are shown on second and third pages.? Is this meant that there are a lack of quality websites/pages for this industry or query? What do you say? Here is the SERPs from Google US.(as on July 09) http://bitly.com/Md8JFW http://bitly.com/LaT9yD Thanks Alex
Industry News | | alexgray0 -
Why Does OSE (Open Site Explorer) have such little backlink data on russian sites in the google.ru index?
OK this seems v strange, but google.ru are indexing far more BLs in their SERPS for a widget than OSE reports. Very little data is found in OSE for russian based sites. Is this the marketing intention? (I could send raw data if needed!) What is filtering this vast google.ru data list out? Is OSE only catered for US/UK?
Industry News | | Turkey0 -
What is the best method for getting pure Javascript/Ajax pages Indeded by Google for SEO?
I am in the process of researching this further, and wanted to share some of what I have found below. Anyone who can confirm or deny these assumptions or add some insight would be appreciated. Option: 1 If you're starting from scratch, a good approach is to build your site's structure and navigation using only HTML. Then, once you have the site's pages, links, and content in place, you can spice up the appearance and interface with AJAX. Googlebot will be happy looking at the HTML, while users with modern browsers can enjoy your AJAX bonuses. You can use Hijax to help ajax and html links coexist. You can use Meta NoFollow tags etc to prevent the crawlers from accessing the javascript versions of the page. Currently, webmasters create a "parallel universe" of content. Users of JavaScript-enabled browsers will see content that is created dynamically, whereas users of non-JavaScript-enabled browsers as well as crawlers will see content that is static and created offline. In current practice, "progressive enhancement" in the form of Hijax-links are often used. Option: 2
Industry News | | webbroi
In order to make your AJAX application crawlable, your site needs to abide by a new agreement. This agreement rests on the following: The site adopts the AJAX crawling scheme. For each URL that has dynamically produced content, your server provides an HTML snapshot, which is the content a user (with a browser) sees. Often, such URLs will be AJAX URLs, that is, URLs containing a hash fragment, for example www.example.com/index.html#key=value, where #key=value is the hash fragment. An HTML snapshot is all the content that appears on the page after the JavaScript has been executed. The search engine indexes the HTML snapshot and serves your original AJAX URLs in search results. In order to make this work, the application must use a specific syntax in the AJAX URLs (let's call them "pretty URLs;" you'll see why in the following sections). The search engine crawler will temporarily modify these "pretty URLs" into "ugly URLs" and request those from your server. This request of an "ugly URL" indicates to the server that it should not return the regular web page it would give to a browser, but instead an HTML snapshot. When the crawler has obtained the content for the modified ugly URL, it indexes its content, then displays the original pretty URL in the search results. In other words, end users will always see the pretty URL containing a hash fragment. The following diagram summarizes the agreement:
See more in the....... Getting Started Guide. Make sure you avoid this:
http://www.google.com/support/webmasters/bin/answer.py?answer=66355
Here is a few example Pages that have mostly Javascrip/AJAX : http://catchfree.com/listen-to-music#&tab=top-free-apps-tab https://www.pivotaltracker.com/public_projects This is what the spiders see: view-source:http://catchfree.com/listen-to-music#&tab=top-free-apps-tab This is the best resources I have found regarding Google and Javascript http://code.google.com/web/ajaxcrawling/ - This is step by step instructions.
http://www.google.com/support/webmasters/bin/answer.py?answer=81766
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
Some additional Resources: http://googlewebmastercentral.blogspot.com/2009/10/proposal-for-making-ajax-crawlable.html
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
http://www.google.com/support/webmasters/bin/answer.py?answer=357690