Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
SEO - New URL structure
-
Hi,
Currently we have the following url structure for all pages, regardless of the hierarchy: domain.co.uk/page, such as domain/blog name.
Can you, please confirm the following:
1. What is the benefit of organising the pages as a hierarchy, i.e. domain/features/feature-name or domain/industries/industry-name or domain/blog/blog name etc.
2. This will create too many 301s - what is Google's tolerance of redirects? Is it worth for us changing the url structure or would you only recommend to add breadcrumbs?
Many thanks
Katarina
-
Hey all!
I am asking question in replies as i don't have free trial any more. Well, my question is about technical and off-page seo. I get confused about the both more often. Can someone please clarify the difference between the two? I am new to seo and applying my learnings to my technology blog for improving the search ranking. -
How wonderful Adam. I am currenctly working on a smaller <a href="https://cryptocasinosverige.com/">Bitcoin Casino</a> site and I hope that eventually Google will notice it. It is not easy but Moz gives great insight on how all SEO related things work.
-
I had a really working and profitable website made by qualified employees of the company https://seotwix.com/ . I liked their professionalism, efficiency and friendly attitude to clients. There was a truly impressive work - created a unique design with original findings and innovations. Adequate understanding of the needs of my company and careful attention to all my complex needs, especially in the design of the structure of the site, as well as its further promotion on the Internet.
-
Hi there! There seems to be a bit of confusion in this thread between URL structure and Information Architecture. Having more folders in a URL doesn't reduce the authority but pages with more folders in the URL tend to be deeper in the sites linking architecture, which means they tend to have less authority because they aren't as close to the surface. The difference between internal links and url format is an important one. There's a blog post here which explains in more depth.
From my perspective, here are the benefits of having pages within folders;
- There is an opportunity to put more relevant keywords in the URL without stuffing
- Easier folder-level reporting in Google Analytics, Search Console etc.
- Some increased understanding for Google of how pages hang together - there is some evidence that Google uses folder structure for ranking before it knows much about the page for example.
In terms of managing authority for pages and signals of relevance I'd be looking much more towards the internal linking to those pages. I wouldn't rely on Google intuitively understanding the topical connection between two pages unless both of those pages target that topic or have relevant links between them. So for example, say you have two pages;
If those pages are both subcategories of trinkets you could reformat them to be;
Having "trinkets" in the url might help both pages rank for "trinkets" type keywords, like "doodad trinkets" for example. However, I wouldn't rely on this change to help Google understand that widgets are related to doodads - you can handle that much more effectively with relevant internal links between /widgets and /doodads that make the relation clear.
In terms of whether there is a risk to making this change - this is essentially a migration and definitely comes with risks associated, even if all of your redirects are 1:1 and direct. It'll take time for Google to find the redirects and new pages, and as a rule of thumb, link equity isn't passed perfectly along a 301 redirect so I wouldn't expect these new pages to just inherit the strength of the old ones.
I think it comes down to weighing up whether the benefits I listed above outweigh the risk of an in-site migration. If you think the keyword targeting opportunities will make enough of a difference then great but I wouldn't rely on url structure as a way to get Google to understand your site differently - the impact of internal links is going to be a far greater factor.
-
Google's tolerance for 301 redirects is pretty high as long as you use speedy ones (implement via NginX - 'engine X', not via .htaccess lines). If the redirects are logical and they don't chain or contact with incorrect redirect types (Meta refreshes, 302s etc) then usually you're ok. Still it will take Google time to digest all the changes and you could see a small interim performance dip
Flat URL structure tends to build the 'authority' of URLs better, making them more powerful. Deeper and more nested URL structures serve 'relevance' better as they give much more context. If your domain's overall SEO authority is low to begin with, then a flatter structure may be better for now. If you have lots of SEO authority then you may be able to 'irrigate' more deeply nested URLs more effectively, thus reaping long-tail gains (so each structure has strengths and weaknesses, depending upon your current standing on the web)
Flatter structures rank better for larger terms, but only if you have the SEO authority to power them. Deeper structures rank better for longer-tail terms (but thousands of them) - again though without the right SEO authority metrics, there will be very few droplets of 'SEO juice' which end up reaching the lower-level pages
In the end most sites evolve to a point where they adopt the more deeply nested structure, but they usually suffer growing pains as they transition. In the long run it can be superior, but only for sites which can make good use of it (e.g: eCommerce web stores with categories, products, collections, product variants etc). If a site is services based it often doesn't have so much SEO authority and also - the deeper structure isn't really so relevant! A services based site will usually offer far fewer services than an eCommerce store offers products (tens vs hundreds of thousands)
A strong publisher with lots of ranking power (online magazines, newspaper digital editions) will often switch to the deeper structure for listing their content and (in the long run) see a lot of benefit from that. For smaller publications (blogs, blog or news pages on business / non-publisher sites) - it's often not worth the move
-
Hi,
Thanks for your answer. We sell B2B software.
The website is structured as global, /us, /au etc. It's just the urls appear all equal atm.
Thanks
Katarina
-
It all depends what you're selling & where you're selling it, also if your hierarchy structure allows for the inclusion of keywords including geo locations, all the better.
Somewhat dated but useful article https://moz.com/learn/seo/url
-
One additional thought to add extra complexity, adding hierarchy is fine, but try to avoid increasing page depth while doing so.
John Mueller discussed this in a few places in the past year that page depth > URL structure.
-
Thanks for your time.
Excellent! Now I'm super scared haha But I understand what you are saying and will share your advice with the team.
Many thanks.
Katarina
-
Hi
No your sweet on the redirects/301's - many sites have 95% redirects from http to https for example. So no chains and you are fine.
Well my view on above is that advice on a hierarchical structure is dangerous. Our job is to always adopt a "first do no harm" approach. We have many clients - no hierarchical structure and awesome rankings. Do we very slowly build hierarchical structures into them - yes. It makes life easier for all. But would we touch the top traffic driving pages - 100% no. It is too high a risk. So you need to do a proper evaluation of the site and what pages are ranking - getting clicks and what are not. There may be sections, a low risk that can move into a hierarchical structure - start there. But do not make a change for change sake to follow what is now good practice.
Hope that helps.
-
Hi,
Thanks for your answer.
'...if the site is ranking well under current strategy...' - I mean, we don't know as there is nothing to compare with. Recently we have been presented with an idea of creating subfolders and clearly showing the site hierarchy via urls. Apparently, it should make an instant difference and should improve our ranks. I'm really unsure if this is guaranteed.
FYI - we would never 301 one url more than just once so no chain. However, I wonder if we had 95% of all site urls redirected if this would impact us negatively.
Also - one more thing we are doing now (and we never used to have) is creating portfolio pages - very relevant pages linking from one main page to demonstrate the hierarchy further.
I'm trying to find out if adding so many 301s and putting all the effort into creating a hierarchy via additional articles, pages, breadcrumbs etc would definitely result in a positive outcome.
Thanks
Katarina
-
Hi
Not clear 100% on the question. Firstly if the site is ranking well under current strategy then recommend where appropriate that continues. It sounds like every page hangs straight off the root domain? However, if the opportunity presents to build out a hierarchical structure then we would recommend same.
The benefit of a hierarchical structure is it builds out topical authority or makes it easier for search engines to interpret the site. All google has done is roll the old dewy library system into the site maps. By analogy the more books you have hanging off the History section (parent subfolder) the better the site should be seen in the context of providing answers to history. Then it comes down to the quality of pages hanging off the subfolder and how much shared.
So in short to answer your question a hierarchical structure makes your site easier for Google to understand and builds out topical authority which long term is future proofing against voice search.
Onto the second part of the question, there is no problems with 301's per se as long as it is one hop.. so to redirect a page more than 3 times is a big negative as Google often does not crawl those pages. Recommended practice to change the redirects from page 1 > page 4 and page 2 > page 4, page 3 > 4, etc so all old redirects point in one hop to the final destination page.
Hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What happens when you replace a page with a new version that has the same URL?
a new page template was created the plan is to publish the new page (which has the same URL as before) to web and delete the old page that has the URL , will that have an SEO implications ?
Technical SEO | | lina_digital1 -
Japanese URL-structured sitemap (pages) not being indexed by Bing Webmaster Tools
Hello everyone, I am facing an issue with the sitemap submission feature in Bing Webmaster Tools for a Japanese language subdirectory domain project. Just to outline the key points: The website is based on a subdirectory URL ( example.com/ja/ ) The Japanese URLs (when pages are published in WordPress) are not being encoded. They are entered in pure Kanji. Google Webmaster Tools, for instance, has no issues reading and indexing the page's URLs in its sitemap submission area (all pages are being indexed). When it comes to Bing Webmaster Tools it's a different story, though. Basically, after the sitemap has been submitted ( example.com/ja/sitemap.xml ), it does report an error that it failed to download this part of the sitemap: "page-sitemap.xml" (basically the sitemap featuring all the sites pages). That means that no URLs have been submitted to Bing either. My apprehension is that Bing Webmaster Tools does not understand the Japanese URLs (or the Kanji for that matter). Therefore, I generally wonder what the correct way is to go on about this. When viewing the sitemap ( example.com/ja/page-sitemap.xml ) in a web browser, though, the Japanese URL's characters are already displayed as encoded. I am not sure if submitting the Kanji style URLs separately is a solution. In Bing Webmaster Tools this can only be done on the root domain level ( example.com ). However, surely there must be a way to make Bing's sitemap submission understand Japanese style sitemaps? Many thanks everyone for any advice!
Technical SEO | | Hermski0 -
Best Web-site Structure/ SEO Strategy for an online travel agency?
Dear Experts! I need your help with pointing me in the right direction. So far I have found scattered tips around the Internet but it's hard to make a full picture with all these bits and pieces of information without a professional advice. My primary goal is to understand how I should build my online travel agency web-site’s (https://qualistay.com) structure, so that I target my keywords on correct pages and do not create a duplicate content. In my particular case I have very similar properties in similar locations in Tenerife. Many of them are located in the same villa or apartment complex, thus, it is very hard to come up with the unique description for each of them. Not speaking of amenities and pricing blocks, which are standard and almost identical (I don’t know if Google sees it as a duplicate content). From what I have read so far, it’s better to target archive pages rather than every single property. At the moment my archive pages are: all properties (includes all property types and locations), a page for each location (includes all property types). Does it make sense adding archive pages by property type in addition OR in stead of the location ones if I, for instance, target separate keywords like 'villas costa adeje' and 'apartments costa adeje'? At the moment, the title of the respective archive page "Properties to rent in costa adeje: villas, apartments" in principle targets both keywords... Does using the same keyword in a single property listing cannibalize archive page ranking it is linking back to? Or not, unless Google specifically identifies this as a duplicate content, which one can see in Google Search Console under HTML Improvements and/or archive page has more incoming links than a single property? If targeting only archive pages, how should I optimize them in such a way that they stay user-friendly. I have created (though, not yet fully optimized) descriptions for each archive page just below the main header. But I have them partially hidden (collapsible) using a JS in order to keep visitors’ focus on the properties. I know that Google does not rank hidden content high, at least at the moment, but since there is a new algorithm Mobile First coming up in the near future, they promise not to punish mobile sites for a collapsible content and will use mobile version to rate desktop one. Does this mean I should not worry about hidden content anymore or should I move the descirption to the bottom of the page and make it fully visible? Your feedback will be highly appreciated! Thank you! Dmitry
Technical SEO | | qualistay1 -
Coming soon SEO
Hi, I was wondering what is the best practice to redirect all the links juice by redirecting all the pages of your website to a coming soon page. The coming soon page will point to the domain.com, not to a subfolder. Should I move the entire website to a subfolder and redirect this folder to the coming soon page? Thanks
Technical SEO | | bigrat950 -
How Does Dynamic Content for a Specific URL Impact SEO?
Example URL: http://www.sja.ca/English/Community-Services/Pages/Therapy Dog Services/default.aspx The above page is generated dynamically depending on what province the visitor visits from. For example, a visitor from BC would see something quite different than a visitor from Nova Scotia; the intent is that the information shown should be relevant to the user of that province. How does this effect SEO? How (or from what location) does Googlebot decide to crawl the page? I have considered a subdirectory for each province, though that comes with its challenges as well. One such challenge is duplicate content when different provinces may have the same information for some pages. Any suggestions for this?
Technical SEO | | ey_sja0 -
Special characters in URL
Hi There, We're in the process of changing our URL structure to be more SEO friendly. Right now I'm struggling to find a good way to handle slashes that are part of a targeted keyword. For example, if I have a product page and my product title is "1/2 ct Diamond Earrings in 14K Gold" which of the following URLs is the right way to go if I'm targeting the product title as the search keyword? example.com/jewelry/1-2-ct-diamond-earrings-in-14k-gold example.com/jewelry/12-ct-diamond-earrings-in-14k-gold example.com/jewelry/1_2-ct-diamond-earrings-in-14k-gold example.com/jewelry/1%2F2-ct-diamond-earrings-in-14k-gold Thanks!
Technical SEO | | Richline_Digital0 -
Optimal Structure for Forum Thread URL
For getting forum threads ranked, which is best and why? site.com**/topic/**thread-title-goes-here site.com**/t/**thread-title-goes-here site.com**/**thread-title-goes-here I'd take comfort in knowing that SEOmoz uses the middle version, except that "q" is more meaningful to a human than "t". The last option seems like the best bet overall, except that users could potentially steal urls that I may want to use in the future. My old structure was site.com/forum/topic/TOPIC_ID-thread-title-goes-here so obviously any of those would be a vast improvement, but I might as well make the best choice now so I only have to change once.
Technical SEO | | PatrickGriffith0 -
Is there a great tool for URL mapping old to new web site?
We are implementing new design and removing some pages and adding new content. Task is to correctly map and redirect old pages that no longer exist.
Technical SEO | | KnutDSvendsen0