Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Query string parameters always bad for SEO?
-
I've recently put some query string parameters into links leading to a 'request a quote' form which auto-fill the 'product' field with the name of the product that is on the referring product page.
E.g.
Red Bicycle product page >>> Link to RFQ form contains '?productname=Red-Bicycle' >>>> form's product field's default value becomes 'Red-Bicycle'
I know url parameters can lead to keyword cannibalisation and duplicate content, we use sub-domains for our language changer. BUT for something like this, am I potentially damaging our SEO?
Appreciate I've not explained this very well. We're using Kentico by the way, so K# macros are a possibility (I use a simple one to fill the form's Default Field).
-
No, I would make sure it's the best use case for you. Sometimes you can store this data in a cookie. But there are still great ways for just SEO that can help you with making sure that it won't hurt. Usually the best way is still having a canonical tag on the page that matches with the page that the content is originally from.
-
Hi Martijn,
Thanks for the reply. Am I going about this completely the wrong way? Would you recommend using local storage instead?
All the best,
Michael
-
Well it would depend on how you set up the data. In the end you can transfer the data from one page to the other in multiple ways (local data storage, cookies, POST). So in most cases you wouldn't even need a parameter like this so you can keep your URLs as clean as possible.
-
Thanks!
I've defined it in Google Search Console and asked it to not be crawled.
It isn't actually for tracking, simply to auto-fill a form for the customer, hopefully to enhance conversions (we sell quite wordy and complex products, some people visit the form, get in a muddle and prefer to ring us than complete the form).
If I was to 301 it back to the original URL, wouldn't that immediately move the user to a URL without the query string?
-
Hi,
Parameters are definitely not always a bad thing, if you use them for filtering or pagination they have a great use case. Also for tracking purposes which I think is what you're talking about here it would be fine. But you probably want to put some support in place to make sure you don't mess up your SEO. You have a few ways of doing that:
- Define the URL parameters in Google Search Console, read more about it here.
- In some cases if it's just for tracking purposes but would trigger an actual pageview or 200 status code. Then it might be good to implement a canonical URL that would link back to the original page.
- Robots.txt, way more aggressive but if you don't want search engines to look at the pages at all you can exclude the parameters through the robots.txt. I would advise against using this if you don't have great SEO knowledge.
Hope this was useful!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Capturing Source Dynamically for UTM Parameters
Does anyone have a tutorial on how to dynamically capture the referring source to be populated in UTM parameters for Google Analytics? We want to syndicate content and be able to see all of the websites that provided referral traffic for this specific objective. We want to set a specific utm_medium and utm_campaign but have the utm_source be dynamic and capture the referring website. If we set a permanent utm_source, it would appear the same for all incoming traffic. Thanks in advance!
Technical SEO | | peteboyd0 -
Domain Masking SEO Impact
I hope I am explaining this correctly. If I need to provide any clarity please feel free to ask. We currently use a domain mask on an external platform that points back to our site. We are a non-profit and the external site allows users to create peer-to peer fundraisers that benefit our ministry. Currently we get many meta issues related to this site as well as broken links when fundraisers expire etc. We do not have a need to rank for the information from this site. Is there a way to index these pages so that they are not a part of the search engine site crawls as it relates to our site?
Technical SEO | | SamaritansPurse0 -
Breadcrumbs on Mobile How important are they for SEO?
Due to Poor unsightly look of breadcrumbs and the space it takes up above the fold we only employ breadcrumbs on our desktop version. Breadcrumbs are hidden from view on mobile version. However as mobile first indexing is now in play what technical SEO impacts will this have? one thing that comes to mind is crawling deeper pages where breadcrumbs made them accessible in less than 3 link clicks? But i am unsure now of the impacts of not having breadcrumbs visible for mobile version of our site.
Technical SEO | | oceanstorm0 -
Can a CMS affect SEO?
As the title really, I run www.specialistpaintsonline.co.uk and 6 months ago when I first got it it had bad links which google had put a penalty against it so losts it value. However the penalty was lift in Sept, the site corresponds to all guidelines and seo work has been done and constantly monitored. the issue I have is sales and visits have not gone up, we are failing fast and running on 2 or 3 sales a month isn't enough to cover any sort of cost let alone wages. hence my question can the cms have anything to do with it? Im at a loss and go grey any help or advice would be great. thanks in advance.
Technical SEO | | TeamacPaints0 -
Are links in menus to external sites bad for SEO?
We're building a blog on a subdomain of the main site. The main site is on Shopify and the blog will be on wordpress. I'd like to keep the user experience as simple as possible so I'd like to make the blog look exactly like the main Shopify site. This means having a menu in the blog that duplicates the Shopify menu. So is it bad for SEO to have someone click on the 'about us' button in the blog subdomain (blog.mainsite.com) which takes you to the 'about us page' on the main shopify website (mainsite.com)?
Technical SEO | | acs1110 -
Are 404 Errors a bad thing?
Good Morning... I am trying to clean up my e-commerce site and i created a lot of new categories for my parts... I've made the old category pages (which have had their content removed) "hidden" to anyone who visits the site and starts browsing. The only way you could get to those "hidden" pages is either by knowing the URLS that I used to use or if for some reason one of them is spidering in Google. Since I'm trying to clean up the site and get rid of any duplicate content issues, would i be better served by adding those "hidden" pages that don't have much or any content to the Robots.txt file or should i just De-activate them so now even if you type the old URL you will get a 404 page... In this case, are 404 pages bad? You're typically not going to find those pages in the SERPS so the only way you'd land on these 404 pages is to know the old url i was using that has been disabled. Please let me know if you guys think i should be 404'ing them or adding them to Robots.txt Thanks
Technical SEO | | Prime850 -
No indexing url including query string with Robots txt
Dear all, how can I block url/pages with query strings like page.html?dir=asc&order=name with robots txt? Thanks!
Technical SEO | | HMK-NL0 -
How to handle sitemap with pages using query strings?
Hi, I'm working to optimize a site that currently has about 5K pages listed in the sitemap. There are not in face this many pages. Part of the problem is that one of the pages is a tool where each sort and filter button produces a query string URL. It seems to me inefficient to have so many items listed that are all really the same page. Not to mention wanting to avoid any duplicate content or low quality issues. How have you found it best to handle this? Should I just noindex each of the links? Canonical links? Should I manually remove the pages from the sitemap? Should I continue as is? Thanks a ton for any input you have!
Technical SEO | | 5225Marketing0