Google Cached "Text Only" version
-
Is there a way to test what a page would look like in Google "Text Only" version before a page is indexed in Google? Is there a tool out there to help with this?
-
Thanks Dan, that works. We were having some issues because our advertising page skin was blocking all of the written content on our pages. When the pages got cached in google and you clicked "text only", there was no text there because the page skin was blocking it. Just fetched as googlebot thought and all text seems to come through now, but we'll se what happens when the pages get indexed.
Thanks!
-
Kevin
I bet if you have Webmaster Tools setup and you run a "fetch as Googlebot" this will get you what you need. You'll have the header response info in there too, but you'll see how Google will view all the source HTML.
-Dan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Now that Google is no longer publicly displaying Page Rank updates, how will this effect Moz's ability to calculate DA and PA?
Hi, How much more important do you guys think that Moz's Page Authority and Domain Authority metrics are going to become now that Google has stopped giving people public access to a site or pages Page Rank? And how accurate is PA and DA as a measurement in comparison to Page Rank..so for example if I was seeking a guestposting opportunity and saw a site as having a PR of 4....if I now looked to Moz's Page and Domain Authority metrics instead...would that still give me equivalent information on the strength of that domain and thus make a judgement on whether it will be a worthy site for a guestpost.. I guess what I am asking is, how close is now looking at Moz's metrics (ie. a third party company) to the info on PageRank that was being updated by Google themselves? Also will the lack of updated public PR info from Google effect the ability for Moz to calculate PA and DA?? Look forward to your replies on this,
Industry News | | sanj50500 -
Manual action penalty by Google
Hello, We have a big well-known brand - www.titanbet.com. This brand is well established and the site has been live for almost 4 years now ranking very well on some very strong KWs. we received a message from Google on Aug 29<sup>th</sup> saying “Google has detected a pattern of artificial or unnatural links pointing to your site” and that “Google has applied a manual spam action to titanbet.com/” The past 2 weeks since the penalty was received we saw some of our major KWs drop in rankings. BUT all brand related KWs were still ranked 1<sup>st</sup> Over the last weekend the penalty has worsen and we no longer rank on any of the brands KWs (we find the site in 5<sup>th</sup> page at best). Moreover, when searching for a sentence from the any of the page on the site in Google, we see other sites ahead of us in the SERPs. Based on the message we originally received from Google we have started cleaning some of the bad links to the site. We found a lot of links from bad sites, some of them are not indexed and probably penalized as well, some are from affiliate websites and some are from some automatic indexation websites based in China and Russia
Industry News | | Tit
we have started reaching out to some of these sites to try and have them remove our links. We are also worried about the duplication of our site. We have found many other sites (mostly affiliate websites) have copied and in some cases completely duplicated our content. Google for some reason has chosen to penalize us for this. Although we do not have control over these other sites. We have run copyscape to try and figure out which pages are the most problematic and we will try to re-write the content on these pages. But what if the other sites copy us again? Any suggestions on the above would be appreciated as we try to understand why Google has penalized us. thank you Titan Bet Team0 -
How can i discover how many of my pages have been indexed by google?
I am currently in the process of trying to produce a report for my corporation and this is a metric that i cannot seem to find on OpenSiteExplorer. Could anyone help?
Industry News | | CF20150 -
What are the click-through rates by position for organic search on Google by industry?
I'm interested in knowing what the average CTRs are by industry (specifically for the credit card industry). Here's an example of what I want in aggregate (not by industry): http://searchengineland.com/organic-click-thru-rates-tumbling-study-97338 Is this information available?
Industry News | | wepayinc0 -
When will Rand put out "Art of SEO 2nd Edition"? (ANSWER: IN ABOUT 2 WEEKS)
First edition was printed in the end of 2009. Great Book. Needs updating of course. I would buy the next edition if it was updated in an awesome way that I know Rand and the others would do.
Industry News | | stubby0 -
What is the best method for getting pure Javascript/Ajax pages Indeded by Google for SEO?
I am in the process of researching this further, and wanted to share some of what I have found below. Anyone who can confirm or deny these assumptions or add some insight would be appreciated. Option: 1 If you're starting from scratch, a good approach is to build your site's structure and navigation using only HTML. Then, once you have the site's pages, links, and content in place, you can spice up the appearance and interface with AJAX. Googlebot will be happy looking at the HTML, while users with modern browsers can enjoy your AJAX bonuses. You can use Hijax to help ajax and html links coexist. You can use Meta NoFollow tags etc to prevent the crawlers from accessing the javascript versions of the page. Currently, webmasters create a "parallel universe" of content. Users of JavaScript-enabled browsers will see content that is created dynamically, whereas users of non-JavaScript-enabled browsers as well as crawlers will see content that is static and created offline. In current practice, "progressive enhancement" in the form of Hijax-links are often used. Option: 2
Industry News | | webbroi
In order to make your AJAX application crawlable, your site needs to abide by a new agreement. This agreement rests on the following: The site adopts the AJAX crawling scheme. For each URL that has dynamically produced content, your server provides an HTML snapshot, which is the content a user (with a browser) sees. Often, such URLs will be AJAX URLs, that is, URLs containing a hash fragment, for example www.example.com/index.html#key=value, where #key=value is the hash fragment. An HTML snapshot is all the content that appears on the page after the JavaScript has been executed. The search engine indexes the HTML snapshot and serves your original AJAX URLs in search results. In order to make this work, the application must use a specific syntax in the AJAX URLs (let's call them "pretty URLs;" you'll see why in the following sections). The search engine crawler will temporarily modify these "pretty URLs" into "ugly URLs" and request those from your server. This request of an "ugly URL" indicates to the server that it should not return the regular web page it would give to a browser, but instead an HTML snapshot. When the crawler has obtained the content for the modified ugly URL, it indexes its content, then displays the original pretty URL in the search results. In other words, end users will always see the pretty URL containing a hash fragment. The following diagram summarizes the agreement:
See more in the....... Getting Started Guide. Make sure you avoid this:
http://www.google.com/support/webmasters/bin/answer.py?answer=66355
Here is a few example Pages that have mostly Javascrip/AJAX : http://catchfree.com/listen-to-music#&tab=top-free-apps-tab https://www.pivotaltracker.com/public_projects This is what the spiders see: view-source:http://catchfree.com/listen-to-music#&tab=top-free-apps-tab This is the best resources I have found regarding Google and Javascript http://code.google.com/web/ajaxcrawling/ - This is step by step instructions.
http://www.google.com/support/webmasters/bin/answer.py?answer=81766
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
Some additional Resources: http://googlewebmastercentral.blogspot.com/2009/10/proposal-for-making-ajax-crawlable.html
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
http://www.google.com/support/webmasters/bin/answer.py?answer=357690 -
Google's New Release "What do you love"
Is this gonna be a game changer for SEO http://www.wdyl.com/ Regards, Shailendra Sial
Industry News | | IM_Learner0 -
How to achieve the highest global and local relevance in google?
Let's say I have a company that has its main business in Europe for thefollowing languages: English German Portugese French Italian And let's say some other markets (e.g. the Portugese one in south america) is also important. The question now is how should we structure the Domain if we want onlyone top level domain (www.company.com)? a) By using subdomains to target users with Google Webmaster Tools for the relevant country: portugal.company.com/pt (same content) brasil.company.com/pt (same content) germany.company.com/de england.company.com/en etc. or b) by using virtual folders www.company.com/pt www.company.com/de www.company.com/en
Industry News | | imsi
etc. or c) something completely different I do not know about? What do you reckon is best? I appreciate all suggestions!0