GA and Ajax Forms
-
Hi Mozzers,
Gravity Forms for Wordpress provides an unsurpassed user experience. The form submits and the form response replaces it without having to reload or redirect to another page.
The problem is that without a redirect there's no way to track this event in Google Analytics.
I'm interested in hearing from anyone who tried to track and ajax event in Google Analytics.
-
...ask Gravity — why didn't I think of that? Good advice, thank you both.
-
-
You could tag an event when a user hits the submit button .... then set it as a goal but you should ask gravity forms im sure they have the best way to track Eli
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
We are writing 5,000 word long form content that is relevant and engaging. It is too long?
We are writing a series of relevant and informative "power pages" on our site. In the past these have been 2,000 to 3,000 words and our audience has shown to be highly engaged with these pages and they converted well. We have decided to expand our new pages to capture more relevant keywords/topics and the result is they are a bit over 5,000 words. Is there a point where long content, even if highly relevant and engaging, is too long to benefit SEO? Is there any reason we would limit ourselves to 2,000-ish word long form content? I ask because I have read multiple blog posts that suggest that long form content that has ranked well in Google ranges between 2,000 and 3,000 words.
Intermediate & Advanced SEO | | Cutopia0 -
Is Google able to see child pages in our AJAX pagination?
We upgraded our site to a new platform the first week of August. The product listing pages have a canonical issue. Page 2 of the paginated series has a canonical pointing to page 1 of the series. Google lists this as a "mistake" and we're planning on implementing best practice (https://webmasters.googleblog.com/2013/04/5-common-mistakes-with-relcanonical.html) We want to implement rel=next,prev. The URLs are constructed using a hashtag and a string of query parameters. You'll notice that these parameters are ¶meter:value vs ¶meter=value. /products#facet:&productBeginIndex:0&orderBy:&pageView:grid&minPrice:&maxPrice:&pageSize:& None of the URLs are included in any indexed URLs because the canonical is the page URL without the AJAX parameters. So these results are expected. Screamingfrog only finds the product links on page 1 and doesn't move to page 2. The link to page 2 is AJAX. ScreamingFrog only crawls AJAX if its in Google's deprecated recommendations as far as I know. The "facet" parameter is noted in search console, but the example URLs are for an unrelated URL that uses the "?facet=" format. None of the other parameters have been added by Google to the console. Other unrelated parameters from the new site are in the console. When using the fetch as Google tool, Google ignores everything after the "#" and shows only the main URL. I tested to see if it was just pulling the canonical of the page for the test, but that was not the case. None of the "#facet" strings appear in the Moz crawl I don't think Google is reading the "productBeginIndex" to specify the start of a page 2 and so on. One thought is to add the parameter in search console, remove the canonical, and test one category to see how Google treats the pages. Making the URLs SEO friendly (/page2.../page3) is a heavy lift. Any ideas how to diagnose/solve this issue?
Intermediate & Advanced SEO | | Jason.Capshaw0 -
How to switch from URL based navigation to Ajax, 1000's of URLs gone
Hi everyone, We have thousands of urls generated by numerous products filters on our ecommerce site, eg./category1/category11/brand/color-red/size-xl+xxl/price-cheap/in-stock/. We are thinking of moving these filters to ajax in order to offer a better user experience and get rid of these useless urls. In your opinion, what is the best way to deal with this huge move ? leave the existing URLs respond as before : as they will disappear from our sitemap (they won't be linked anymore), I imagine robots will someday consider them as obsolete ? redirect permanent (301) to the closest existing url mark them as gone (4xx) I'd vote for option 2. Bots will suddenly see thousands of 301, but this is reflecting what is really happening, right ? Do you think this could result in some penalty ? Thank you very much for your help. Jeremy
Intermediate & Advanced SEO | | JeremyICC0 -
Is it OK to have Search Engines Skip Ajax Content Execution?
I recently added some ajax pages to automatically fill in small areas of my site upon page loading. That is, the user doesn't have to click anything. Therefore when Google and Bing crawl the site the ajax is executed too. However, my understanding is that does not mean Google and Bing are also crawling the ajax content. I actually would prefer that the content would be not be executed OR crawled by them. In the case of Bing I would prefer that the content not even be executed because indications are that the program exits the ajax page for Bing because Bing isn't retaining session variables which that page uses, which makes me concerned that perhaps when that happens Bing isn't able to even crawl the main content..dunno..So, ajax execution seems potentially risky for normal crawling in this case. I would like to simply have my program skip the ajax execution for Google and Bing by recognizing them in the useragent and using an If robot == Y skip ajax approach. I assume I could put the ajax program in the robots.txt file but that wouldn't keep Bing from executing it (and having that exit problem mentioned above). It would be simpler to just have them skip the ajax execution altogether. Is that ok or is there a chance the search engines will penalize my site if they find out (somehow) that I have different logic for them than for the actual users? In the past this surely was not a concern but I understand that Google is increasingly trying to become like a browser so may increasingly have a problem with this approach. Thoughts?
Intermediate & Advanced SEO | | friendoffood0 -
How to make AJAX content crawlable from a specific section of a webpage?
Content is located in a specific section of the webpage that are being loaded via AJAX.
Intermediate & Advanced SEO | | zpm20140 -
Is SEO as Effective on AJAX Sites?
Hey Everyone, I had a potential client contact me about doing SEO for their site and I see that they have an AJAX site where all the content is rendered dynamically via AJAX. I've been doing SEO for years, but never had a client with an AJAX site. I did a little research and see how you can setup alternative pages (or snapshots as Google calls them) with the actual content so the pages are crawlable and will get indexed, but I'm wondering if that is as effective as optimizing static HTML pages or if Google treats AJAX page alternatives as less trustworthy/valuable. Also, does having the site in AJAX effect link building and social sharing? With the link structure, it seems there could be some issues with pointing links and passing link juice to internal pages Thanks! Kurt
Intermediate & Advanced SEO | | Kurt_Steinbrueck1 -
I formed a network with 12 domains in 3 different data centers.
I formed a network with 12 domains in 3 different data centers.How should I link them to my main Domain? I did something : all minor domains link to main domain site-wide (nofollow) and 3-5 internal links from each minor domains go to sub pages of the major domain (dofollow linking from content). Shall I link each other like A>>c>d......
Intermediate & Advanced SEO | | onlinegiyim0 -
Canonical tags and GA tracking on premium sub-domain?
Hello! I'm launching a premium service on my site that will deliver two fairly distinct user experiences, but with nearly identical page content across the two. I'm thinking of placing the "upgraded" version on a subdomain, e.g. www.mysite.com, premium.mysite.com. Simple enough. I've run into two obstacles, however: -I don't want the premium site crawled separately, so I'd like to use canonical tags to pull all premium.* back to their www.* parents. --How different can page content be before canonical tags backfire? --Is there any other danger in using canonicals across subdomains like this? -Less importantly: with Google Analytics, if I track against the subdomain my visits will split naturally, and it should generate a second cookie for a new registrant who crosses subdomains. I could also use a visitor-level custom var. Good idea? Bad idea? Thanks! -m
Intermediate & Advanced SEO | | grumbles0