Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
AJAX requests and implication for SEO
-
Hi,
I got a question in regard to webpages being served via AJAX request as I couldn't find a definitive answer in regard to an issue we currently face:
When visitors on our site select a facet on a Listing Page, the site doesn't fully reload. As a consequence only certain tags of the content (H1, description,..) are updated, while other tags like canonical URLs, meta noindex,nofollow tag, or the title tag are not updating as long as you don't refresh the page.
We have no information about how this will be crawled and indexed yet but I was wondering if anyone of you knows, how this will impact SEO?
-
Hi - right, then if the URL changes for the user, you'll want to probably use the PushState method (linked above) to convey this to Google. They likely can't see the URL change by default.
You can check by trying to crawl the site with Screaming Frog SEO Spider with the user agent set to Googlebot. Then go to "outlinks" for the page with the facet links, and see if they are listed.
Hope that helps some more! Let me know if you need further direction.
-Dan
-
Hi Dan,
Yes, the URL would change in both cases (AJAX request as well as for a complete refresh).
Thanks
-
Hi - right, I should have answered your specific situation too

When the user selects a facet - does this change the URL too? Meaning, it's supposed to be a totally different page?
-
Hi Dan,
Although it doesn't address the specific question I had I think your link is a good start for me to dig into the topic in more detail.
Many Thanks,
-
Google recently updated how they prefer AJAX is handled:
https://webmasters.googleblog.com/2015/10/deprecating-our-ajax-crawling-scheme.html
They now recommend using the PushState Method - I won't pretend to know all the intricacies of how it works for implementation, but that's the best method to go with. If you need any more help let us know, and I'll have another associate jump in and take a look.
-Dan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
SEO for multiple languages [Arabic]
Hello all, I am currently managing a Marketplace that comes in two different languages: English & Arabic. The English website is, fortunately, doing quite well in terms of SEO performances but, not the Arabic one. The website has two kinds of content: Static content: controlled by me. It includes menu items, navigation, static pages etc which is properly translated among the two languages User-uploaded content: It includes ads/news posted by the user which may not be translated to Arabic if they chose not to do it. Now if somebody goes to the Arabic website and check a news item that doesn't have an Arabic translation, it will show the English title. I am assuming, serving content in a different language that is specified in the hreflang is a straight no, right?
Intermediate & Advanced SEO | | MozammilStorat0 -
Mass Removal Request from Google Index
Hi, I am trying to cleanse a news website. When this website was first made, the people that set it up copied all kinds of articles they had as a newspaper, including tests, internal communication, and drafts. This site has lots of junk, but this kind of junk was on the initial backup, aka before 1st-June-2012. So, removing all mixed content prior to that date, we can have pure articles starting June 1st, 2012! Therefore My dynamic sitemap now contains only articles with release date between 1st-June-2012 and now Any article that has release date prior to 1st-June-2012 returns a custom 404 page with "noindex" metatag, instead of the actual content of the article. The question is how I can remove from the google index all this junk as fast as possible that is not on the site anymore, but still appears in google results? I know that for individual URLs I need to request removal from this link
Intermediate & Advanced SEO | | ioannisa
https://www.google.com/webmasters/tools/removals The problem is doing this in bulk, as there are tens of thousands of URLs I want to remove. Should I put the articles back to the sitemap so the search engines crawl the sitemap and see all the 404? I believe this is very wrong. As far as I know this will cause problems because search engines will try to access non existent content that is declared as existent by the sitemap, and return errors on the webmasters tools. Should I submit a DELETED ITEMS SITEMAP using the <expires>tag? I think this is for custom search engines only, and not for the generic google search engine.
https://developers.google.com/custom-search/docs/indexing#on-demand-indexing</expires> The site unfortunatelly doesn't use any kind of "folder" hierarchy in its URLs, but instead the ugly GET params, and a kind of folder based pattern is impossible since all articles (removed junk and actual articles) are of the form:
http://www.example.com/docid=123456 So, how can I bulk remove from the google index all the junk... relatively fast?0 -
Is CloudFlare bad for SEO?
I have been hit by DDoS attacks lately...not on a huge scale, but probably done by some "script kiddies" or competitors of mine. Still, I need to take some action in order to protect my server and my site against all of this spam traffic that is being sent to it. In the process of researching the tools available for defending a website from a DDoS attack, I came across the service offered by CloudFlare.com. According to the CloudFlare website, they protect your site against a DDoS attack by showing users/visitors they find suspicious an interstitial that asks them if they are a real user or a bot...this interstitial contains a Captcha that suspicious users are asked to enter in order to visit the site. I'm just wondering what kind of an effect such an interstitial could have on my Google rankings...I can imagine that such a thing could add to increased click-backs to the SERPs and, if Google detects this, to lower rankings. Has anyone had experience with the DDoS protection services offered by CloudFlare, who can say a word or two regarding any effects this may have on SEO? Thanks
Intermediate & Advanced SEO | | masterfish1 -
Server Migration, Does it effect SEO?
About to go through a server migration. My intitial thought is that a change in servers shouldn't really change my rankings. But I've heard rumors... Can a server migration change rankings? Why?
Intermediate & Advanced SEO | | Thos0030 -
What should I cover in a SEO proposal ?
What should I cover in a SEO proposal? Is there any sample SEO Proposal template in SEOMoz?
Intermediate & Advanced SEO | | kashyaplakkad1 -
Is DOCTYPE important for SEO?
Hello fellow Mozzers. I am just having a brief look at a potential clients website before speaking to them tomorrow and whilst looking at the source I noticed that they don't appear to have a clear definition for their Doctype. All the have at the top of each page is I have to admit that Doctypes aren't my strong point but I know that they are normally slightly more descriptive than this. Can this have any effect on rankings? or is this just an issue for W3C validation? Thanks 🙂
Intermediate & Advanced SEO | | AdeLewis0 -
Iframe redirect bad for SEO?
Hi, I have a website (http://www.blowingminds.de) wich I put a spreadshirt shop into via iframe. The thing is I am not sure on how the iframe effects my SEO? Can I just optimise the main domain for search? Well I want the spreadshirt shop to be found under the domain name (www.blowingminds.de) but the only real way to do it is by implementing an iframe because each spreadshirt shop has its own subdomain eg.: blowingminds.spreadshirt.de but the only real way to do it is via iframe, as they do not offer a complete domain redirect. (Or have I overseen some other way?) I hope you guys can help me on this one 🙂 Thanks in advance. Malte
Intermediate & Advanced SEO | | wellbo1 -
Subdomains and SEO - Should we redirect to subfolder?
A new client has mainsite.com and a large numer of city specific sub domains i.e. albany.mainsite.com. I think that these subdomains would actually work better as subfolders i.e mainsite.com/albany rather than albany.mainsite.com. The majority of links on the subdomains link to the main site anyway i.e. mainsite.com/contactus rather than albany.mainsite.com/contactus. Having mostly main domain links on a subdomain doesnt seem like clever link architecture to me and maybe even spammy. Im not overly familiar with redirecting subdomains to subfolders. If we go the route of 301'ing subdomains to subfolders any advice/warnings?
Intermediate & Advanced SEO | | AndyMacLean0