Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Looking to remove dates from URL permalink structure. What do you think of this idea?
-
I know most people who remove dates from their URL structure usually do so and then setup a 301 redirect. I believe that's the right way to go about this typically. My biggest fear with doing a global 301 redirect implementation like that across an entire site is that I've seen cases where this has sort of shocked Google and the site took a hit in organic traffic pretty bad.
Heres what I'm thinking a safer approach would be and I'd like to hear others thoughts. What if...
- Changed permalink structure moving forward to remove the date in future posts.
- All current URLs stay as is with their dates
- Moving forward we would go back and optimize past posts in waves (including proper 301 redirects and better URL structure). This way we avoid potentially shocking Google with a global change across all URLs.
Do you know of a way this is possible with a large Wordpress website? Do you see any conplications that could come about in this process? I'd like to hear any other thoughts about this please.
Thanks!
-
Hey Jeff,
thank you for your input. So you just globally changed the permalink structure, put global redirects in place and you didn't see permanent loss in trafic? And you did that on multiple sites?
If so I'll most probably follow your path.
Thanks again,
Julien
-
Hey Julien -
I wouldn't go this route. Since asking this question I have had dates removed from 30+ domains, many with 5-10 million+ pageviews per month. We haven't seen this as a risk and are now very in favor of removing dates from URLs on most sites we work with. We work with sites that have very evergreen content, and republishing is a very strong SEO strategy.
The process is very similar to moving your site to HTTPS from HTTP. Since Google has started recommending HTTPS we haven't seen any issue with removing dates as well.
Hope that helps
-
Hey Thomas,
Interesting thought! Could you go in a little more details as to how that regex would work? Would that randomize the redirects to only a portion of the posts?
Thanks!
Julien
-
I think only do 10% of pages watch them if you like what you see do the next 20%
RedirectMatch301^/([0-9]{4})/([0-9]{2})/(.*)$ http://yourwebsite.com/$3
-
Garrett -
I never got a clear answer, but I have since gone forward making changes on 20+ Wordpress blogs without any ill-effect. The changes we made were only to sites that had dates in the permalink structure and 301 redirects were put in place (on the server, not through a plugin). Trying to change the permalink structure going forward but not back was too much of a hassle. It appears Google sees this as a positive change for users because it cleans up the permalink structure and allows site owners to keep their content updated and continue sharing.
Not sure how this will apply in other scenarios such as removing folder structure (categories and tags) from the permalink, but I've had only positive results removing the dates. I work with some very high profile mom and food blogs so I have some pretty solid evidence and data supporting my decisions now.
I hope that helps. Cheers!
-
Hi Jeff,
Did you end up making these changes? How is it going? I found your post as I was researching and rethinking how to structure WordPress blog permalinks.
I have a few e-commerce clients with blog posts that are several years old and still popular in organic search. I'd like to turn some of them into evergreen content that is regularly updated, but I feel like we should do something about the permalinks first.
There are some great insights here. Thank you to all who contributed.
Garrett
-
No problem, glad to help! Best of luck with whichever route you go with!
-
It was worth a shot. Thanks for sharing your thoughts. Cheers!
-
Unfortunately, I don't have any examples for ya. Never come across this particular topic for a client.
-
Know of any site that has used the canonical to do anything like this? It seems like the safest option, I just haven't seen this to this scale is all.
-
Yes, I'm saying you should keep URLs as they are. I'm always an advocate for not changing URL structure unless there's a really good, highly beneficial reason for doing so. I don't know of a way to change only new URL structures while keeping old ones the same, but I'm no WP expert.
-
Although I haven't strongly considered that approach, it did cross my mind to utilize the canonical. Do you know of any way to change WordPress permalink structure going forward but not backwards? Or are you suggesting we keep the dates in the URL going forward? I just think that eventually we'll have to think about updating that URL structure.
-
OK, now that I understand the reasoning...
I believe there's a better, less-risky approach. What I would do is write a completely new post based on information from the old post. At the same time you publish the new post, go back to the old version and add these two things: a canonical tag pointing to the new version, and a bit of _very readable _text at the top linking to the new post. Something like "Hey, thanks for your interest in our content. Feel free to read on, but we thought you should know we've updated this post which can be found here: link"
This accomplishes a few important things. It eliminates the need for a risky project that could affect your entire site just for the ability to update posts (which I'm guessing doesn't happen too often, what percent of posts get updated?). The canonical tag removes the dupe content risk so you're not cannibalizing your own content. And leaving the old post there gives people the opportunity to discover old content that, while possibly not relevant anymore, still demonstrates you've been a trustworthy source of information for a long time.
-
Logan,
By not being able to remove the dates we're not able to go back to a 5-year old post, make updates, and then republish the content. This is a "mom blog" and the topics can be recycled, but if we create a new post that we also covered 5 years ago we would be competing with ourself instead of using something that already has some authority and rank to it.
That's why we were thinking to somehow make it possible (in WordPress) to keep all current URLs as is, change the permalink structure moving forward so that future posts don't have date, and then be able to update posts as we go and 301 them manually over time. Does that make sense?
I agree with your last 2 statements, it is a HUGE risk to 301 this entire site to do away with those dates. Even though redirects supposedly pass all link juice we all know that a big change like that across an entire site could have ill-effect with search engines.
I'd like to know if anyone has gone about the URL structure change like I'm outlining here. Am I crazy to think that is a logical way to go about it? I haven't been able to find anywhere that someone has done this though.
-
Jeff,
Based on the traffic you say this blog gets, I'm assuming its rather large and has hundreds, if not thousands of posts. Which leads me to one simple question:
Why? This seems like a HUGE amount of risk and a pretty decent amount of work to go into something that's really not going to provide any benefit.
*edit: It should also be noted that just because Google has recently stated that redirects now pass all link juice doesn't mean you should go needlessly add a massive amount of redirects. There are other implications that redirects have, like load time for example. If you have 1,000 redirects, every single one of those is going to be checked before any page on your site loads, which takes a lot of time.
-
Thanks for your response. I actually agree with most, if not all of what you are saying.
The problem is that this is a larger blog with 5-7 million page views on average per month. 1 million+ just from organic. I agree with your statement about postponing and never getting done. With a large blog I still think it would be easier (less stressful, not necessarily easier) to manage it in waves in order to pause or correct when there is a larger than normal dip that maybe doesn't come back up. With a business it makes sense, but with these bloggers sites it seems like too big of a risk when it's what brings in almost all the income. Does that make sense?
That tweet you're referring to, I thought that was mainly in regard to HTTP to HTTPS migrations. I need to look more into that I guess.
Thanks!
-
I'm not a fan of your plan.
There can be many reasons why a site might "take a hit". For example, if page-to-page redirects were not implemented or the sitemap was not updated, updated correctly, or resubmitted to search engines. I wouldn't assume that will happen in your case. In my experience, if the transition is done correctly and there's a hit, it's short-lived.
If you're thinking the redirects will cause you to lose SEO equity, that is no longer the case. Gary Illyes, a Google webmaster trends analyst, tweeted on July 26, 2016 "30x redirects don’t lose PageRank anymore."
One of the biggest risks (in my mind) of staging the migration the way you suggest is that the "waves" never happen. I see that a lot - a situation where an organization agrees to postpone work to a future date that never arrives. New and competing priorities take precedence resulting in an endless postponement. If you have the management commitment, funding and resources to do the work now, I say bite the bullet and go for it. Make a plan. Stick to it. Check and double check your work.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Removing .html from URLs - impact of rankings?
Good evening Mozzers. Couple of questions which I hope you can help with. Here's the first. I am wondering, are we likely to see ranking changes if we remove the .html from the sites URLs. For example website.com/category/sub-category.html Change to: website.com/category/sub-category/ We will of course make sure we 301 redirect to the new, user friendly URLs, but I am wondering if anyone has had previous experience of implementing this change and how it has effected rankings. By having the .html in the URLs, does this stop link juice being flowed back to the root category? Second question: If one page can be loaded with and without a forward slash "/" at the end, is this a duplicate page, or would Google consider this as the same page? Would like to eliminate duplicate content issues if this is the case. For example: website.com/category/ and website.com/category Duplicate content/pages?
Intermediate & Advanced SEO | | Jseddon920 -
URL Injection Hack - What to do with spammy URLs that keep appearing in Google's index?
A website was hacked (URL injection) but the malicious code has been cleaned up and removed from all pages. However, whenever we run a site:domain.com in Google, we keep finding more spammy URLs from the hack. They all lead to a 404 error page since the hack was cleaned up in the code. We have been using the Google WMT Remove URLs tool to have these spammy URLs removed from Google's index but new URLs keep appearing every day. We looked at the cache dates on these URLs and they are vary in dates but none are recent and most are from a month ago when the initial hack occurred. My question is...should we continue to check the index every day and keep submitting these URLs to be removed manually? Or since they all lead to a 404 page will Google eventually remove these spammy URLs from the index automatically? Thanks in advance Moz community for your feedback.
Intermediate & Advanced SEO | | peteboyd0 -
Linking to URLs With Hash (#) in Them
How does link juice flow when linking to URLs with the hash tag in them? If I link to this page, which generates a pop-over on my homepage that gives info about my special offer, where will the link juice go to? homepage.com/#specialoffer Will the link juice go to the homepage? Will it go nowhere? Will it go to the hash URL above? I'd like to publish an annual/evergreen sort of offer that will generate lots of links. And instead of driving those links to homepage.com/offer, I was hoping to get that link juice to flow to the homepage, or maybe even a product page, instead. And just updating the pop over information each year as the offer changes. I've seen competitors do it this way but wanted to see what the community here things in terms of linking to URLs with the hash tag in them. Can also be a use case for using hash tags in URLs for tracking purposes maybe?
Intermediate & Advanced SEO | | MiguelSalcido0 -
Remove URLs that 301 Redirect from Google's Index
I'm working with a client who has 301 redirected thousands of URLs from their primary subdomain to a new subdomain (these are unimportant pages with regards to link equity). These URLs are still appearing in Google's results under the primary domain, rather than the new subdomain. This is problematic because it's creating an artificial index bloat issue. These URLs make up over 90% of the URLs indexed. My experience has been that URLs that have been 301 redirected are removed from the index over time and replaced by the new destination URL. But it has been several months, close to a year even, and they're still in the index. Any recommendations on how to speed up the process of removing the 301 redirected URLs from Google's index? Will Google, or any search engine for that matter, process a noindex meta tag if the URL's been redirected?
Intermediate & Advanced SEO | | trung.ngo0 -
How to detect a bad link and remove ?
As per google penguin, all the low quality back links are going to affect the website SERPS hugely. So, we need to find all the bad back links and then remove them one by one. What I would like to know is, what tool do you use to find all the bad back links ? And how do we know which is a bad back link or bad website, where our link should not be there ? Then what service what do you suggest for back links removal. I contacted LinkDelete.com and they quoted me 97$ for a month to remove all links in less than 3 weeks.
Intermediate & Advanced SEO | | monali123
Let me know, what you suggest.0 -
Best way to permanently remove URLs from the Google index?
We have several subdomains we use for testing applications. Even if we block with robots.txt, these subdomains still appear to get indexed (though they show as blocked by robots.txt. I've claimed these subdomains and requested permanent removal, but it appears that after a certain time period (6 months)? Google will re-index (and mark them as blocked by robots.txt). What is the best way to permanently remove these from the index? We can't use login to block because our clients want to be able to view these applications without needing to login. What is the next best solution?
Intermediate & Advanced SEO | | nicole.healthline0 -
URL Error or Penguin Penalty?
I am currently having a major panic as our website www.uksoccershop.com has been largely dropped from Google. We have not made any changes recently and I am not sure why this is happening, but having heard all sorts of horror stories of penguin update, I am fearing the worst. If you google "uksoccershop" you will see that the homepage does not rank. We previously ranked in the top 3 for "football shirts" but now we don't, although on page 2, 3 and 4 you will see one of our category pages ranking (this didn't used to happen). Some rankings are intact, but many have disappeared completely and in some cases been replaced by other pages on our site. I should point out our existing rankings have been consistently there for 5-6 years until today. I logged into webmaster tools and thankfully there is no warning message from Google about spam, etc, but what we do have is 35,000 URL errors for pages which are accessible. An example of this is: | URL: | http://www.uksoccershop.com/categories/5_295_327.html | | Error details In Sitemaps Linked from Last crawled: 6/20/12First detected: 6/15/12Googlebot couldn't access the contents of this URL because the server had an internal error when trying to process the request. These errors tend to be with the server itself, not with the request. Is it possible this is the cause of the issue (we are not currently sure why the URL's are being blocked) and if so, how severe is it and how recoverable?If that is unlikely to cause the issue, what would you recommend our next move is?All help is REALLY REALLY appreciated 🙂
Intermediate & Advanced SEO | | ukss19840 -
/%category%/%postname%/ Permalink structure
Mostly everyone seems to agree that /%category%/%postname%/ is the best blog structure. I'm thinking of changing my structure to that because now it's structured by date which is bad. But almost all of my posts are assigned to more than one category. Won't this create duplicate pages?
Intermediate & Advanced SEO | | UnderRugSwept0