I just found something weird I can't explain, so maybe you guys can help me out.
-
I just found something weird I can't explain, so maybe you guys can help me out.
In Google http://www.google.nl/#hl=nl&q=internet. The number 3 result is a big telecom provider in the Netherland called Ziggo. The ranking URL is https://www.ziggo.nl/producten/internet/. However if you click on it you'll be directed to https://www.ziggo.nl/#producten/internet/
HttpFox in FF however is not showing any redirects. Just a 200 status code.
The URL https://www.ziggo.nl/#producten/internet/ contains a hash, so the canonical URL should be https://www.ziggo.nl/. I can understand that. But why is Google showing the title and description of https://www.ziggo.nl/producten/internet/, when the canonical URL clearly is https://www.ziggo.nl/?
Can anyone confirm my guess that Google is using the bulk SEO value (link juice/authority) of the homepage at https://www.ziggo.nl/ because of the hash, but it's using the relevant content of https://www.ziggo.nl/producten/internet/ resulting in a top position for the keyword "internet".
-
The site you've pointed to uses ajax to load its content. When the page loads there's a javascript snippet which takes over and adds the # to the page (hence why you're not seeing it as a httpd header). If you click on any other link you'll see that the base URL stays the same with some extra parameters on the end.
There are potential crawling issues with this and a number of fixes (some Google documentation here, although this isn't the fix that the site in question is using: http://code.google.com/intl/en-US/web/ajaxcrawling/).
So, in short, there's nothing fishy going on - it's just good old ajax content loading
- Matt
-
This is actually a fairly crude attempt of loading AJAX content. I say 'crude' because it's not quite using Google's documented AJAX protocol using the hashbang (#!). There was an SEOmoz post about Google's protocol a while back that had some good examples:
http://www.seomoz.org/blog/how-to-allow-google-to-crawl-ajax-content
For this specific website, there actually is a JavaScript redirect involved. The original URL will load, then some JS will do some work and eventually do a document.location.replace() to do the redirect to the URL with the hash. As far as GoogleBot is concerned it won't necessarily do the redirect and will index the original page.
One thing I want to caution is to again remember that this site is not exactly adhering to Google's recommendations on AJAX content. Coupled with the fact that there is a JS redirect going on I would say that there might be a risk of cloaking. On the front end, the content looks the same and I would kinda hope that Google would just treat this scenario similar to their hashbang solution because this site is not intending to do some tricky stuff here. But we can't trust that Google will always give a free pass.
-
This looks more like a dynamic site using AJAX, rather than anchors in the page like you're thinking.
See: http://code.google.com/web/ajaxcrawling/docs/getting-started.html
No funny stuff. The page you see is the page google intended to show you, with all the SEO value for the page itself being responsible for its spot in the SERPs.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
My Home Page meta title on Google isn't what it should be
Hey guys My website is http://www.oxfordmeetsfifth.com According to SEOcentro, my website should appear to Google as Fashion Tips for Women | Oxford Meets Fifth. I have used the Yoast plugin and force rewrote titles to ensure that is the home page meta title. It also appears correctly in browser. Could anyone advise why this is the case? Thanks in advance!
Technical SEO | | OxfordMeetsFifth0 -
What to do with 404 errors when you don't have a similar new page to 301 to ??
Hi If you have 404 errors for pages that you dont have similar content pages to 301 them to, should you just leave them (the 404's are optimised/qood quality with related links & branding etc) and they will eventually be de-indexed since no longer exist or should you 'remove url' in GWT ? Cheers Dan
Technical SEO | | Dan-Lawrence0 -
Creating in-text links with ' 'target=_blank' - helping/hurting SEO!?!
Good Morning Mozzers, I have a question regarding a new linking strategy I'm trying to implement at my organization. We publish 'digital news magazines' that oftentimes have in-text links that point to external sites. More recently, the editorial department and me (SEO) conferred on some ways to reduce our bounce rate and increase time on page. One of the suggestions I offered is to add the 'target=_blank" attribute to all the links so that site visitors don't necessarily have to leave the site in order to view the link. It has, however, come to my attention that this can have some very negative effects on my SEO program, most notably, (fake or inaccurate) time(s) on-page. Is this an advisable way to create in-text links? Are there any other negative effects that I can expect from implementing such a strategy?
Technical SEO | | NiallSmith0 -
I always get this error "We have detected that the domain or subfolder does not respond to web requests." I don't know why. PLEASE help
subdomain www.nwexterminating.com subfolder www.nwexterminating.com/pest_control www.nwexterminating.com/termite_services www.nwexterminating.com/bed_bug_services
Technical SEO | | NWExterminating0 -
Found a Typo in URL, what's the best practice to fix it?
Wordpress 3.4, Yoast, Multisite The URL is supposed to be "www.myexample.com/great-site" but I just found that it's "www.myexample.com/gre-atsite" It is a relatively new site but we already pointed several internal links to "www.myexample.com/gre-atsite" What's the best practice to correct this? Which option is more desirable? 1.Creating a new page I found that Yoast has "301 redirect" option in the Advanced tap Can I just create a new page(exact same page) and put noindex, nofollow and redirect it to http://www.myexample.com/great-site OR 2. htacess redirect rule simply change the URL to http://www.myexample.com/great-site and update it, and add Options +FollowSymLinks RewriteEngine On
Technical SEO | | joony2008
RewriteCond %{HTTP_HOST} ^http://www.myexample.com/gre-atsite$ [NC]
RewriteRule ^(.*)$ http://www.myexample.com/great-site$1 [R=301,L]0 -
Merged old wordpress site to new theme and have crazy amount of 4xx and duplicate content that wasn't there before?
URL is awardrealty.com We have a new website that we merged into a new wordpress theme. I just crawled the site with my seomoz crawl tool and it is showing a ridiculous amount of 4xx pages (200+) and we cant find the 4xx pages in the sitemap or within wordpress. Need some help? Am i missing something easy?
Technical SEO | | Mark_Jay_Apsey_Jr.0 -
Site: search doesn't return homepage first
When searching for site:myclient.com their homepage doesn't appear first. I know some SEOs have reported this was a warning sign that there was a penalty. Here is what I've checked/found: Toolbar pagerank remains strong. Homepage is indexed. SEO traffic is falling, but its been gradually falling for a year now, mainly due to the client neglecting any type of marketing campaigns or link building, I believe. There was not a specific drop that could be tied to a penalty. Site remains well indexed. 62,742 of 63,021 URLs in the sitemap are indexed. Site is a large ecommerce site, so many pages are duplicate content (product descriptions). Homepage does rank #1 when searching for string of text present on the homepage. Nothing unusual in Google Webmaster Tools Search for myclient.com returns homepage with 6 expanded sitelinks under it. Google safe browsing check shows no malware. Anything else I should check?
Technical SEO | | AdamThompson0 -
How do I use the Robots.txt "disallow" command properly for folders I don't want indexed?
Today's sitemap webinar made me think about the disallow feature, seems opposite of sitemaps, but it also seems both are kind of ignored in varying ways by the engines. I don't need help semantically, I got that part. I just can't seem to find a contemporary answer about what should be blocked using the robots.txt file. For example, I have folders containing site comps for clients that I really don't want showing up in the SERPS. Is it better to not have these folders on the domain at all? There are also security issues I've heard of that make sense, simply look at a site's robots file to see what they are hiding. It makes it easier to hunt for files when they know the directory the files are contained in. Do I concern myself with this? Another example is a folder I have for my xml sitemap generator. I imagine google isn't going to try to index this or count it as content, so do I need to add folders like this to the disallow list?
Technical SEO | | SpringMountain0