Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Query string parameters always bad for SEO?
-
I've recently put some query string parameters into links leading to a 'request a quote' form which auto-fill the 'product' field with the name of the product that is on the referring product page.
E.g.
Red Bicycle product page >>> Link to RFQ form contains '?productname=Red-Bicycle' >>>> form's product field's default value becomes 'Red-Bicycle'
I know url parameters can lead to keyword cannibalisation and duplicate content, we use sub-domains for our language changer. BUT for something like this, am I potentially damaging our SEO?
Appreciate I've not explained this very well. We're using Kentico by the way, so K# macros are a possibility (I use a simple one to fill the form's Default Field).
-
No, I would make sure it's the best use case for you. Sometimes you can store this data in a cookie. But there are still great ways for just SEO that can help you with making sure that it won't hurt. Usually the best way is still having a canonical tag on the page that matches with the page that the content is originally from.
-
Hi Martijn,
Thanks for the reply. Am I going about this completely the wrong way? Would you recommend using local storage instead?
All the best,
Michael
-
Well it would depend on how you set up the data. In the end you can transfer the data from one page to the other in multiple ways (local data storage, cookies, POST). So in most cases you wouldn't even need a parameter like this so you can keep your URLs as clean as possible.
-
Thanks!
I've defined it in Google Search Console and asked it to not be crawled.
It isn't actually for tracking, simply to auto-fill a form for the customer, hopefully to enhance conversions (we sell quite wordy and complex products, some people visit the form, get in a muddle and prefer to ring us than complete the form).
If I was to 301 it back to the original URL, wouldn't that immediately move the user to a URL without the query string?
-
Hi,
Parameters are definitely not always a bad thing, if you use them for filtering or pagination they have a great use case. Also for tracking purposes which I think is what you're talking about here it would be fine. But you probably want to put some support in place to make sure you don't mess up your SEO. You have a few ways of doing that:
- Define the URL parameters in Google Search Console, read more about it here.
- In some cases if it's just for tracking purposes but would trigger an actual pageview or 200 status code. Then it might be good to implement a canonical URL that would link back to the original page.
- Robots.txt, way more aggressive but if you don't want search engines to look at the pages at all you can exclude the parameters through the robots.txt. I would advise against using this if you don't have great SEO knowledge.
Hope this was useful!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Shopify SEO - Double Filter Pages
Hi Experts, Single filter page: /collections/dining-chairs/black
Technical SEO | Sep 1, 2023, 1:56 AM | williamhuynh
-- currently, canonical the same: /collections/dining-chairs/black
-- currently, index, follow Double filter page: /collections/dining-chairs/black+fabric
-- currently, canonical the same: /collections/dining-chairs/black+fabric
-- currently, noindex, follow My question is about double filter page above:
if noindexing is the better option OR should I change the canonical to /collections/dining-chairs/black Thank you0 -
Are on-site content carousel bad for SEO?
Hi, I didn't find an answer to my question in the Forum. I attached an example of content carousel, this is what I'm talking about. I understand that Google has no problem anymore with tabbed contents and accordeons (collapsible contents). But now I'm wondering about textual carousels. I'm not talking about an image slider, I'm talking about texts. Is text carousel harder to read for Google than plain text or tabs? Of course, i'm not talking about a carousel using Flash. Let's say the code is proper... Thanks for your help. spfra5
Technical SEO | May 28, 2020, 11:22 AM | Alviau0 -
Domain Masking SEO Impact
I hope I am explaining this correctly. If I need to provide any clarity please feel free to ask. We currently use a domain mask on an external platform that points back to our site. We are a non-profit and the external site allows users to create peer-to peer fundraisers that benefit our ministry. Currently we get many meta issues related to this site as well as broken links when fundraisers expire etc. We do not have a need to rank for the information from this site. Is there a way to index these pages so that they are not a part of the search engine site crawls as it relates to our site?
Technical SEO | Feb 18, 2019, 9:38 AM | SamaritansPurse0 -
JavaScript page loader - SEO impact
Hello all,
Technical SEO | Feb 16, 2016, 9:17 PM | Lvet
I am working on a site that has a bizarre page load system. All pages get loaded trough the same Javascript snippet, for example: Changing the values in the form changes the page that is loaded. The most incredible thing is that, against my expectations, pages do get indexed by Google.
My question is: "Does loading pages dynamically using JavaScript affect the overall SEO performance?" Why are pages getting indexed? Thank you for shedding light on this.
Cheers
Luca0 -
Schema for Banks and SEO
I'm researching Schema opportunities for a bank, but besides the shema markup available today (like bankorcreditunion) and developments with FIBO, I can find no answer as to the effect of tagging interest rates and such in terms of SERP/CTR performance or visibility. Does anyone have a case study to share or some insight on the matter?
Technical SEO | Jan 29, 2016, 12:23 PM | Netsociety0 -
Should I block robots from URLs containing query strings?
I'm about to block off all URLs that have a query string using robots.txt. They're mostly URLs with coremetrics tags and other referrer info. I figured that search engines don't need to see these as they're always better off with the original URL. Might there be any downside to this that I need to consider? Appreciate your help / experiences on this one. Thanks Jenni
Technical SEO | Aug 9, 2012, 1:05 PM | ShearingsGroup0 -
Changing DNS -- SEO implications?
Hey Moz, We're migrating an old site on an old server over to a new server/DNS. The plan is to keep the same URL structure and reuse our existing URL's. As long as we make minimal changes to each page's content, we should be able to update our DNS entry and get all the pages recreated and assigned to their correct URLs without any reduction in SEO rankings. Is this correct? This site gets a lot of organic traffic and ranks highly on some challenging keywords, so it's key that we retain our rankings as much as possible. I've read that it's wise to lower the DNS time-to-live to one hour, about a day before the move, to help Google crawl the DNS a little quicker. Are there any other recommendations you guys can offer or past experiences?
Technical SEO | Jul 31, 2014, 5:26 PM | stephen_reply0 -
How to handle sitemap with pages using query strings?
Hi, I'm working to optimize a site that currently has about 5K pages listed in the sitemap. There are not in face this many pages. Part of the problem is that one of the pages is a tool where each sort and filter button produces a query string URL. It seems to me inefficient to have so many items listed that are all really the same page. Not to mention wanting to avoid any duplicate content or low quality issues. How have you found it best to handle this? Should I just noindex each of the links? Canonical links? Should I manually remove the pages from the sitemap? Should I continue as is? Thanks a ton for any input you have!
Technical SEO | Apr 1, 2011, 4:34 AM | 5225Marketing0