Should I Add Location to ALL of My Client's URLs?
-
Hi Mozzers,
My first Moz post! Yay! I'm excited to join the squad
My client is a full service entertainment company serving the Washington DC Metro area (DC, MD & VA) and offers a host of services for those wishing to throw events/parties. Think DJs for weddings, cool photo booths, ballroom lighting etc.
I'm wondering what the right URL structure should be. I've noticed that some of our competitors do put DC area keywords in their URLs, but with the moves of SERPs to focus a lot more on quality over keyword density, I'm wondering if we should focus on location based keywords in traditional areas on page (e.g. title tags, headers, metas, content etc) instead of having keywords in the URLs alongside the traditional areas I just mentioned. So, on every product related page should we do something like:
example.com/weddings/planners-washington-dc-md-va
example.com/weddings/djs-washington-dc-md-va
example.com/weddings/ballroom-lighting-washington-dc-md-vaOR
example.com/weddings/planners
example.com/weddings/djs
example.com/weddings/ballroom-lightingIn both cases, we'd put the necessary location based keywords in the proper places on-page. If we follow the location-in-URL tactic, we'd use DC area terms in all subsequent product page URLs as well. Essentially, every page outside of the home page would have a location in it.
Thoughts?
Thank you!!
-
No website in particular that springs to mind, I'm afraid. But it's not uncommon practice, and I'm sure you'll find plenty within your industry from a little competitor research.
Good luck!
-
This is great stuff. Thank you! Would you happen to have an example of a site that does this well? I think you're spot on in your suggestions and would love to see it in practice.
-
(I had posted my response, but Moz didn't fancy saving it for some reason and it's just gone. So I'll try and remember what I typed and repost it...)
I wouldn't dilute the site authority by using subdomains for your locations.
As a user, I would recommend your main site navigation lists the different event types (weddings, parties, corporate, etc) and branch your locations from there.
e.g.
-
Weddings - /weddings/ (Weddings)
-
Miami - /weddings/miami/ (Weddings in Miami)
-
Planners - /weddings/miami/planners/ (Wedding Planners in Miami)
-
DJs - /weddings/miami/djs/ (Wedding DJs in Miami)
-
Ballroom Lighting - /weddings/miami/ballroom-lighting/ (Ballroom Lighting for Weddings in Miami)
That structure seems the most logical to me, but you should do your own research to back this up. Conduct thorough keyword research for each service in each location and structure your landing page content accordingly. For example, main category pages broadly targeting root keyword, but display "cards" or sections that link to each location without optimising those main category pages for the locations - save this for the location-based landing pages. So this sub-navigation is in the body, rather than in the main navigation, for user-friendliness.
I think with something like events, you don't want to shove the locations in the user's face first thing. Let them see what you offer (the different event types), then delve down into the locations, and the specific services within those locations.
People are free to disagree with me, and I welcome critique on these thoughts. I do think with SEO, it gets to a point after "best practices" that it comes down to more of personal preferences.
-
-
Excellent advice Ria. I'll likely give that advice to the client.
Another question that brewed from this: how then should main navigation be handled as we expand? obviously we can't have D.C. centric keywords in the main navigation as the business expands. I think we could create unique content and landing pages for each individual service and location, but how would that be incorporate into the overall user flow and URL structure?
Would it be more of a sitemap play? If someone goes to www.example.com, should they be given an option to choose their location then be routed to that specific city's subdomain and yhenbrowse from there?
I guess my main question is, how exactly should we structure the site navigation for users from multiple cities to both please UX and the big G?
Thank you!
-
For a handful of different locations, it's quite common to structure them as different subdirectories, as you said. site.com/weddings/miami/planners or /miami/weddings/planners - whichever makes the most sense for your customer base and how you're targeting the content.
Just ensure that these are not considered doorway pages or appear to be too templated. Make each landing page for each location unique, and tailored specifically to your customers in each location. If you have nothing unique to say, then you don't need separate pages. It would be best to target the different locations on the same landing pages. But you being the expert in the industry, I can imagine it'll be easy enough to cater toward each audience specifically. Especially when you're not dealing with tens if not hundreds or thousands of different towns.
If you are certain on expanding to different cities soon, then it might be best to begin the URL structuring with /washington-dc/ subdirectory somewhere, so you don't have to change this later.
-
Thank you, Ria. That's very helpful.
Im curious, when the business expands to different cities in the coming months (for example, Miami and Chicago are being considered, not yet finalized), then in that case I would assume we need to have location in the URL path for the sake of designation and differentiation. This may be a sub folder in and of itself though. Thoughts?
-
I'd avoid adding the location in the URL if you only work with those services for a single location. It looks messy to the user, and can look spammy to Google. And it would save you from having to change the URL and set up redirects, if you need to remove the location keywords from the URL at a later date in order to please the Big G. Optimising for location within the content, title and meta can be easily tweaked with time. Tweaking URLs can be a lot messier.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can 'follow' rather than 'nofollow' links be damaging partner's SEO
Hey guys and happy Monday! We run a content rich website, 12+ years old, focused on travel in a specific region, and advertisers pay for banners/content etc alongside editorial. We have never used 'nofollow' website links as they're no explicitly paid for by clients, but a partner has asked us to make all links to them 'nofollow' as they have stated the way we currently link is damaging their SEO. Could this be true in any way? I'm only assuming it would adversely affect them if our website was peanalized by Google for 'selling links', which we're not. Perhaps they're just keen to follow best practice for fear of being seen to be buying links. FYI we now plan to change to more full use of 'nofollow', but I'm trying to work out what the client is refering to without seeming ill-informed on the subject! Thank you for any advice 🙂
Intermediate & Advanced SEO | | SEO_Jim0 -
Google's Knowledge Panel
Hi Moz Community. Has anyone noticed a pattern in the websites that Google pulls in to populate knowledge Panels? For example, for a lot of queries Google keeps pulling data from a specific source over and over again, and the data shown in the Knowledge Panel isn't on the target page. Is it possible that Google simply favors some sites over others and no matter what you do, you'll never make it into the Knowledge box? Thanks.
Intermediate & Advanced SEO | | yaelslater0 -
How necessary is it to disavow links in 2017? Doesn't Google's algorithm take care of determining what it will count or not?
Hi All, So this is a obvious question now. We can see sudden fall or rise of rankings; heavy fluctuations. New backlinks are contributing enough. Google claims it'll take care of any low quality backlinks without passing pagerank to website. Other end we can many scenarios where websites improved ranking and out of penalty using disavow tool. Google's statement and Disavow tool, both are opposite concepts. So when some unknown low quality backlinks are pointing and been increasing to a website? What's the ideal measure to be taken?
Intermediate & Advanced SEO | | vtmoz0 -
HTML5: Changing 'section' content to be 'main' for better SEO relevance?
We received an HTML5 recommendation that we should change onpage text copy contained in 'section" to be listed in 'main' instead, because this is supposedly better for SEO. We're questioning the need to ask developers spend time on this purely for a perceived SEO benefit. Sure, maybe content in 'footer' may be seen as less relevant, but calling out 'section' as having less relevance than 'main'? Yes, it's true that engines evaluate where onpage content is located, but this level of granular focus seems unnecessary. That being said, more than happy to be corrected if there is actually a benefit. On a side note, 'main' isn't supported by older versions of IE and could cause browser incompatibilities (http://caniuse.com/#feat=html5semantic). Would love to hear others' feedback about this - thanks! 🙂
Intermediate & Advanced SEO | | mirabile0 -
Will Canonical tag on parameter URLs remove those URL's from Index, and preserve link juice?
My website has 43,000 pages indexed by Google. Almost all of these pages are URLs that have parameters in them, creating duplicate content. I have external links pointing to those URLs that have parameters in them. If I add the canonical tag to these parameter URLs, will that remove those pages from the Google index, or do I need to do something more to remove those pages from the index? Ex: www.website.com/boats/show/tuna-fishing/?TID=shkfsvdi_dc%ficol (has link pointing here)
Intermediate & Advanced SEO | | partnerf
www.website.com/boats/show/tuna-fishing/ (canonical URL) Thanks for your help. Rob0 -
.GOV Link - same impact on my site's rankings whether link to home or Gov related category?
I own a job site and I am about to get a link from a .GOV. My site has a category called "State Jobs". Should I ask the ".Gov" to link to my homepage or to the state job page and use the anchor text "State Jobs". I understand "State Jobs" page would get a big kick by that being the anchor text and linking to that specific page, but the question I have is this: for my site as a whole (homepage and various categories) would they get around the same "push up" whether the linking is to 1) my homepage with anchor text being my site's name or 2) to the state job specific page and in this case the anchor text would be "State Jobs"? thank you
Intermediate & Advanced SEO | | knielsen0 -
What's the best method for segmenting HTML sitemap?
Hello all, I was wondering if anyone can help me. Currently I'm trying to set up a HTML sitemap for our website and am having trouble with the 500+ pages of content under each category. How do you segment your HTML sitemap in a case like this, keeping in mind the less than 100 links per page rule? For example, http://www.careerbliss.com/salary/ allows our users to search salaries under company, job title, and location. You can imagine how many thousands of pages we need to represent. Any help will be greatly appreciated! Cheers! Reyna
Intermediate & Advanced SEO | | CareerBliss0 -
Best solution to get mass URl's out the SE's index
Hi, I've got an issue where our web developers have made a mistake on our website by messing up some URL's . Because our site works dynamically IE the URL's generated on a page are relevant to the current URL it ment the problem URL linked out to more problem URL's - effectively replicating an entire website directory under problem URL's - this has caused tens of thousands of URL's in SE's indexes which shouldn't be there. So say for example the problem URL's are like www.mysite.com/incorrect-directory/folder1/page1/ It seems I can correct this by doing the following: 1/. Use Robots.txt to disallow access to /incorrect-directory/* 2/. 301 the urls like this:
Intermediate & Advanced SEO | | James77
www.mysite.com/incorrect-directory/folder1/page1/
301 to:
www.mysite.com/correct-directory/folder1/page1/ 3/. 301 URL's to the root correct directory like this:
www.mysite.com/incorrect-directory/folder1/page1/
www.mysite.com/incorrect-directory/folder1/page2/
www.mysite.com/incorrect-directory/folder2/ 301 to:
www.mysite.com/correct-directory/ Which method do you think is the best solution? - I doubt there is any link juice benifit from 301'ing URL's as there shouldn't be any external links pointing to the wrong URL's.0