How can I best handle parameters?
-
Thank you for your help in advance! I've read a ton of posts on this forum on this subject and while they've been super helpful I still don't feel entirely confident in what the right approach I should take it. Forgive my very obvious noob questions - I'm still learning!
The problem: I am launching a site (coursereport.com) which will feature a directory of schools. The directory can be filtered by a handful of fields listed below. The URL for the schools directory will be coursereport.com/schools. The directory can be filtered by a number of fields listed here:
- Focus (ex: “Data Science”)
- Cost (ex: “$<5000”)
- City (ex: “Chicago”)
- State/Province (ex: “Illinois”)
- Country (ex: “Canada”)
When a filter is applied to the directories page the CMS produces a new page with URLs like these:
- coursereport.com/schools?focus=datascience&cost=$<5000&city=chicago
- coursereport.com/schools?cost=$>5000&city=buffalo&state=newyork
My questions:
1) Is the above parameter-based approach appropriate? I’ve seen other directory sites that take a different approach (below) that would transform my examples into more “normal” urls.
coursereport.com/schools?focus=datascience&cost=$<5000&city=chicago
VERSUS
coursereport.com/schools/focus/datascience/cost/$<5000/city/chicago (no params at all)
2) Assuming I use either approach above isn't it likely that I will have duplicative content issues? Each filter does change on page content but there could be instance where 2 different URLs with different filters applied could produce identical content (ex: focus=datascience&city=chicago OR focus=datascience&state=illinois). Do I need to specify a canonical URL to solve for that case? I understand at a high level how rel=canonical works, but I am having a hard time wrapping my head around what versions of the filtered results ought to be specified as the preferred versions. For example, would I just take all of the /schools?focus=X combinations and call that the canonical version within any filtered page that contained other additional parameters like cost or city?
-
Should I be changing page titles for the unique filtered URLs?
-
I read through a few google resources to try to better understand the how to best configure url params via webmaster tools. Is my best bet just to follow the advice on the article below and define the rules for each parameter there and not worry about using rel=canonical ?
https://support.google.com/webmasters/answer/1235687
An assortment of the other stuff I’ve read for reference:
http://www.wordtracker.com/academy/seo-clean-urls
http://www.practicalecommerce.com/articles/3857-SEO-When-Product-Facets-and-Filters-Fail
http://www.searchenginejournal.com/five-steps-to-seo-friendly-site-url-structure/59813/
http://googlewebmastercentral.blogspot.com/2011/07/improved-handling-of-urls-with.html
-
I think you have your answer then on how you want to focus your URLs and your site!
-
Absolutely helpful. I really appreciate it. I think one real use case that I may want to solve for is the "focus" plus "city" combo. Ie: "data science schools in chicago". Based on the research I've done thus far I think that may be the only permutation really worth worrying about. Again - thanks a lot!
-
I am not going to be very helpful here.
Looking at those parameters and all the options you would have for URLs, yes you are ripe for duplicate content issue and a whole mess of search engine problems/confusion.
I read this the other day in the QNA forum here at Moz and I wish I could remember to give them credit for the quote, they said "Don't submit search results to the search engines" - so true - so true ....
Why? You end up with an almost infinite number of thin, duplicate pages that Google then does not know which ones to rank. Even if you put all the parameters into a static URL you still have the same problem.
I think you need to step back a sec
Are people searching for "data science schools in Chicago Illinois that cost less than $5000"?
Why would you even want to attempt to setup pages that could potentially rank for those terms based on the URL?
Launch the search function on the site, but hide all the search URLs behind robots.txt
Just setup things like
/search/?focus=datascience&cost=$<5000&city=chicago
/search/focus/datascience/cost/$<5000/city/chicago
put /search/ in robots and you are set
Another option (from one of my favorite WBF http://moz.com/blog/whiteboard-friday-using-the-hash)
Hide all the parameters behind the hash and they stay hidden from the search engines
/schools#?focus=datascience&cost=$<5000&city=chicago
Then go back, do your keyword research and build helpful static URL pages around what your users are searching for and then get those pages to rank. If that ultimately is the type of page above, I would bet you $3,141 plus an apple pie that you need to setup a simpler organization of pages and urls around location say /il/chicago/school-name or type /data-science/school-name and then all the other iterations, you would hide behind a hash etc.
Maybe this did help - I hope so.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
URL Parameters for Product Variation
Hey there Mozzers, So I have a website that uses URL Parameters for product variations. So for example when I am in the product "test" that product has 5 different variations. So the url is as followed /product/test /products/test?variant=10271886529 /products/test?variant=10271886530 etc. Does Google understand that this is the same page ? Does it automatically exclude the variable in the url?
Technical SEO | | AngelosS0 -
Best strategy to handle over 100,000 404 errors.
I recently been given a site that has over one-hundred thousand 404 error codes listed in Google Webmasters. It is really odd because according to Google Webmasters, the pages that are linking to these 404 pages are also pages that no longer exist (they are 404 pages themselves). These errors were a result of site migration that had occurred. Appreciate any input on how one might go about auditing and repairing large amounts of 404 errors. Thank you.
Technical SEO | | SEO_Promenade0 -
How to fix google index filled with redundant parameters
Hi All This follows on from a previous question (http://moz.com/community/q/how-to-fix-google-index-after-fixing-site-infected-with-malware) that on further investigation has become a much broader problem. I think this is an issue that may plague many sites following upgrades from CMS systems. First a little history. A new customer wanted to improve their site ranking and SEO. We discovered the site was running an old version of Joomla and had been hacked. URL's such as http://domain.com/index.php?vc=427&Buy_Pinnacle_Studio_14_Ultimate redirected users to other sites and the site was ranking for buy adobe or buy microsoft. There was no notification in webmaster tools that the site had been hacked. So an upgrade to a later version of Joomla was required and we implemented SEF URLs at the same time. This fixed the hacking problem, we now had SEF url's, fixed a lot of duplicate content and added new titles and descriptions. Problem is that after a couple of months things aren't really improving. The site is still ranking for adobe and microsoft and a lot of other rubbish and the urls like http://domain.com/index.php?vc=427&Buy_Pinnacle_Studio_14_Ultimate are still sending visitors but to the home page as are a lot of the old redundant urls with parameters in them. I think it is default behavior for a lot of CMS systems to ignore parameters it doesn't recognise so http://domain.com/index.php?vc=427&Buy_Pinnacle_Studio_14_Ultimate displays the home page and gives a 200 response code. My theory is that Google isn't removing these pages from the index because it's getting a 200 response code from old url's and possibly penalizing the site for duplicate content (which don't showing up in moz because there aren't any links on the site to these url's) The index in webmaster tools is showing over 1000 url's indexed when there are only around 300 actual url's. It also shows thousands of url's for each parameter type most of which aren't used. So my question is how to fix this, I don't think 404's or similar are the answer because there are so many and trying to find each combination of parameter would be impossible. Webmaster tools advises not to make changes to parameters but even so I don't think resetting or editing them individually is going to remove them and only change how google indexes them (if anyone knows different please let me know) Appreciate any assistance and also any comments or discussion on this matter. Regards, Ian
Technical SEO | | iragless0 -
Can I disallow my subdomain for penguin recover?
Hi, I have a site like BannerBuzz.com, before last penguin my site's all keywords were in good position in google, but after penguin hit on my website, my all keywords are going down and down day by day, i have done some changes in my website for improvement, but in 1 change i have some confusion. i have one sub domain (http://reviews.bannerbuzz.com/), which display my websites all keywords user reviews, in which every category's 15 reviews are display in my website http://www.bannerbuzz.com so are those user reviews consider as duplicate content between sub domain and main website. can i disallow sub domain from all search engine? currently sub domain is open for all search engine, is that helpful to block it? Thanks
Technical SEO | | CommercePundit0 -
How do I handle soft 404s on category pages?
I have a site that provides a service where listings are displayed on site for 30 days, then they expire. These listings are categorized by type. On occasion, categories have no listings available, and Google Webmaster Tools is listing them as Soft 404 errors. It's not possible to remove these categories and 301 redirect to another page. Any suggestions on how to work around the soft 404s?
Technical SEO | | ang0 -
Lost ranking and can't figure out why
My page http://www.drschulmanplasticsurgery.com/body/buttock-lift-augmentation-new-york-city/ recently moved from first page to past the 15th. I was never penalized on the last update and have very few links pointing to this page. I can't figure out why i just moved so far back. Can anyone offer some advice?
Technical SEO | | Roots70 -
What is the best way to optimize a page for a magazine
Hi i have a serious problem with a website that i am building http://www.cheapflightsgatwick.com/ with reference to letting the search engines know what the magazine is about. I am building a holiday magazine which will focus on holiday news, cheap deals and holiday reviews. I am wanting the home page to feature for the following keywords holiday news, holiday magazine, holiday ideas, best holiday deals, but the problem i have is, i have tried putting an introduction on the home page but it looks out of place, so what is the best way for me to let google know about what the site is about and to get it ranking well in the search engines any help and advice would be great
Technical SEO | | ClaireH-1848860 -
Best place for new relevant content ?
Hi everyone, The background to my question is that one of the companies I work for has just produced a load of videos answering many of the common questions about their products, for the ecommerce part of their site. I've got these videos transcribed which I want to use as the backbone to some of the main pages on the site (some of which are already ranking reasonably well). I was thinking of putting this content on their department page (so for example their departments would be kitesurfing, land kites, toy kites etc.) above links to their sections (e.g. kite surf kites, kite surf boards etc.). Do you think this would be a good place to put a large amount of text ? (the text will be in a scrollable div) - i.e. could it get in the way of the links which are currently passing juice deeper in to the site ? Alternatives are to put the videos under the links to the sections (which would place them below the fold), or to place each video on its own page and simply link to these pages from the departments page (although it is the department page we wish to rank). Another alternative is to put links to the sections both before and after the new content. Hope that all makes sense! Thanks in advance
Technical SEO | | stukerr0