Question on 301s
-
Hi Everyone,
I have a questions on 301 redirects, i hope someone can give me some help on this.
There was some 301 redirects made on some of the URLs at the beginning of the year, however we are now re-structuring the whole website, which means the URLs which had been given a 301 redirect are now getting another 301.
The question is, should i delete the first 301 redirect from the htaccess file?
Kind Regards
-
Ryan your analogy is fantastic. I totally understand this now and it really makes sense to do it this way.
Thanks for being patient with me
Again thanks all for your feedback on this.
Kind Regards
-
Every URL which is no longer active would require a 301 redirect to the proper page. In the situation you describe:
/a should redirect to /abc
/ab should redirect to /abc
I recognize this seems confusing so forget it's a website for a moment. Think of it as mail after you move.
You lived at 100 Main Street. That is where you received your mail. Now you move to 200 Elm Street. You put in a forward order with the post office (a real world equivalent to a 301 redirect). Now any mail addressed to 100 Main Street will be received at 200 Elm Street.
Now you move again to 300 Wall Street. You would put in another forwarding order so your mail from 200 Elm Street gets delivered to your new address. This solution is fine BUT, your mail from 100 Main Street would be delayed. First it would get forwarded to the 200 Elm Street post office, who would then have to forward it to 300 Wall Street. This process is inefficient (in seo terms, you lose link juice).
You want to change your 100 Main Street forward order to direct your mail to the 300 Wall Street address. Now all of your mail is taken to the proper location in a single hop.
I hope this analogy helps!
-
What happens to the URL
If there are external backlinks going to the URL, are these not going to get lost?
Because as we have mentioned on these 301s, there has been 3 URLs in question.
Hope that makes sense.
-
In the simplest terms, the old page should always be directed to the new page. Think of it as a non-stop flight.
-
Hi Ryan,
Thanks for your feedback, however I am getting a little lost
So what your are saying if I understand is, the 301 should be this:
example.com/a is redirected to example.com/abc
Kind Regards
-
The only thing that concerns me is what CafePress had said "Google stops crawling a link after the 5th redirect or so."
You can offer 100 links on a page. All the links can be to "seomoz.org" and they will all be crawled even though the real URL is "www.seomoz.org" and all 100 links will get redirected.
What CafePress referred to is redirects for a single URL.
www.example.com/a redirects to /ab which redirects to /abc and so forth. A crawler will only follow a single URL so far through a chain of redirects before the PR is completely gone and it stops.
Therefore the preferred solution is to redirect any old or broken URLs to their new URL in a single redirect. I'll share an example based on your site:
Very old URL: example.com/a. It is redirected to example.com/ab
Old URL: example.com/ab. It is redirected to example.com/abc
You could leave these two redirects in place, as-is, and they will work, but it is not recommended. The reason is any traffic to /a will have a double re-direct. First the traffic will go to /ab then to the final destination of /abc. This double redirect is an unnecessary delay, it adds extra points of vulnerability and is a waste of SEO link juice. The preferred solution would be to modify the /a redirect to point to the /abc page directly.
I hope that makes sense.
-
Also, if a page is indexed, which is highly likely (due to XML sitemaps, Google Analytics, Google Toolbar etc), then just removing the 301 redirect (links or no links) means that when this page disappears due to the site changes then you will have an indexed page resulting in a 404 error.
I maintain that you should have single hop 301 redirects on all of the pages that will not be there or will have been moved due to the site updated.
I also agree with what Ryan Kent says about links - you may have some links that have been discovered but not yet recognized pr picked up. If there is a chance that the content has been indexed then it should have an appropriate redirect.
-
Hi Ryan,
The only thing that concerns me is what CafePress had said "Google stops crawling a link after the 5th redirect or so."
I have another issue regarding the 301 re-directs:
We have:
/abcd http://www.example.com/abcde this is actually a 301 on a product page, however we have the same product in a shop page /shop/abcd which we have decided to do away with the shop directory, is it best practice to also do a 301 from the /shop/abcd to /abcde?
Hope that makes sense.
Kind Regards
-
I don't agree with the recommendation to simply delete the 301 due to no visible links. There are two reasons why:
1. It is more work for you to go and research the links to each page
2. There can always be links you are not aware of such as bookmarks, e-mail links, links which don't show up for various reasons, etc.
Just simply modify the 301 to point to the correct URL and you are all set.
-
Thanks for the fantastic feedback.
An example of what has happened on the .htaccess:
/abc http://www.example.com/abcd - This is the 301 that was made in March this year.
/abcd http://www.example.com/abcde - This is the new 301
If i notice that there are no links going to /abc using Open Site Explorer should i just delete this 301?
Kind Regards
-
I would change the original 301 redirect to the new location.
I would then add an additional 301 redirect to the secondary page (the old redirect) to the new location.
So you will have your original URL and the older redirected URL both 301 redirected to where the content now resides. This way you only have one hop on the 301 redirects and you have both old URLs pointing to the new one.
-
should i delete the first 301 redirect from the htaccess file?
The best results would be achieved if each URL had a single 301 redirect to the target page. To that end, yes, you should delete the old 301 redirect and create a new one.
-
+1
Totally forgot about mentioning the inbound links part. Thanks for picking it up, Rick!
-
Hey Gary,
I partially agree with Cafe. However, I wouldn't remove any redirects for URLs which may have backlinks. Maybe it would be a good idea to figure out if any of the redirects which you are removing are from URLs that have earned links? An Open Site Explorer link export would help you figure out if any of those URLs still have value.
-
Hi Gary,
Yes, it is always a good idea to cut down the number of 301 redirects (or any redirects in general) because if I remember correctly, Google stops crawling a link after the 5th redirect or so. You also lose another 10% link juice for each additional redirect.
Lastly, don't forget to 301 redirect the URLs from the beginning of the year to the new re-structured website.
Hope that helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Questions about canonicals
Howdy Moz community, I had a question regarding canonicals. I help a business with their SEO, and they are a service company. They have one physical location, but they serve multiple cities in the state. My question is in regards to canonicals and unique content. I hear that a page with slightly differing content for each page won't matter as much, if most of the content is relevantly the same. This business wants to create service pages for at least 10 other cities they service. The site currently only have pages that are targeting one city location. I was wondering if it was beneficial to use a template to service each city and then put a canonical there to say that it is an identical page to the main city page? Example: our first city was san francisco, we want to create city pages for santa rosa, novato, san jose and etc. If the content for the 2nd, 3rd, 4th, city were the same content as the 1st city, but just had the slight change with the city name would that hurt? Would putting a canonical help this issue, if i alert that it is the same as the 1st page? The reason I want to do this, is because I have been getting concerns from my copywriter that after the 5th city, they can't seem to make the services pages that much different from the first 4 cities, in terms of wording of the content and its structure. I want to know is there a simpler way to target multiple cities for local SEO reasons like geo targeted terms without having to think of a completely new way to write out the same thing for each city service page, as this is very time consuming on my end. Main questions? Will making template service pages, changing the city name to target different geographic locations and putting a canonical tag for the new pages created, and referring back to the main city page going to be effective in terms of me wanting to rank for multiple cities. Will doing this tell google my content is thin or be considered a duplicate? Will this hurt my rankings? Thanks!
Technical SEO | | Ideas-Money-Art0 -
Content development for improving organic SEO Questions
I have a real estate web site at www.nhfinehomes.com. The site is build on a Windows Server running ASP.NET with a basic, home grown CMS. In my Google Analytic under Traffic Sources, Search Engine Optimization, Queries I see a list of over 1,000 keyword phrases that my site had ranked for. In an effort to improve these rankings, I am under the assumption that if I create original content for these terms and add it to my site, it could help improve ranks? First question, is this a valid strategy worth pursuing assuming the content is good, I use good internal linking practices and make sure not to 'orphan' these pages? Second, because my current site technology doesn't really make it easy to add content, I have figured out how to install a WordPress blog at www.nhfinehomes.com/blog (currently not up since the MySQL DB died for some reason) So am I at any disadvantage adding all this content to /blog vs. right on the primary CMS? Obviously the WordPress template is not the same as the main pages so primary nav and footer nav differ. Of course I can cross link between the two but wanted to make sure this was not going to penalize the value of this content. Third, if there is anyone with Windows/.NET optimization experience at the server and site architecture level and can help me give my site a once over, I would be most appreciative. I'm not the programmer and rely and a friend to help with that and his SEO experience is very limited. Thanks in advance.
Technical SEO | | LinkMoser0 -
Site command / Footprint Question
Hi All, I am looking for websites with keywords in the domain and I am using: inurl:keyword/s The results that come back include sub-pages and not only domains with the keywords in the root domain. example of what i mean: www.website.com/keyword/ What I want displayed only: www.keyword/s.com Does anyone know of a site command i can use to display URL's with keywords in the root domain only? Thanks in Advance Greg
Technical SEO | | AndreVanKets0 -
Google Change of Address with Questionable Backlink Profile
We have a .com domain where we are 301-ing the .co.uk site into it before shutting it down - the client no longer has an office in the UK and wants to focus on the .com. The .com is a nice domain with good trust indicators. I've just redesigned the site, added a wad of healthy structured markup, had the duplicate content mostly rewritten - still finishing off this job but I think we got most of it with Copyscape. The site has not so many backlinks, but we're working on this too and the ones it does have are natural, varied and from trustworthy sites. We also have a little feature on the redesign coming up in .Net magazine early next year, so that will help. The .co.uk on the other hand has a fair few backlinks - 1489 showing in Open Site Explorer - and I spent a good amount of time matching the .co.uk pages to similar content on the .com so that the redirects would hopefully pass some pagerank. However, approximately a year later, we are struggling to grow organic traffic to the .com site. It feels like we are driving with the handbrake on. I went and did some research into the backlink profile of the .co.uk, and it is mostly made up of article submissions, a few on 'quality' (not in my opinion) article sites such as ezine, and the majority on godawful and broken spammy article sites and old blogs bought for seo purposes. So my question is, in light of the fact that the SEO company that 'built' these shoddy links will not reply to my questions as to whether they received a penalty notification or noticed a Penguin penalty, and the fact that they have also deleted the Google Analytics profiles for the site, how should I proceed? **To my mind I have 3 options. ** 1. Ignore the bad majority in the .co.uk backlink profile, keep up the change of address and 301's, and hope that we can just drown out the shoddy links by building new quality ones - to the .com. Hopefully the crufty links will fade into insignificance over time.. I'm not too keen on this course of action. 2. Use the disavow tool for every suspect link pointing to the .co.uk site (no way I will be able to get the links removed manually) - and the advice I've seen also suggests submitting a reinclusion request afterwards- but this seems pointless considering we are just 301-ing to the new (.com) site. 3. Disassociate ourselves completely from the .co.uk site - forget about the few quality links to it and cut our losses. Remove the change of address request in GWT and possibly remove the site altogether and return 410 headers for it just to force the issue. Clean slate in the post. What say you mozzers? Please help, working myself blue in the face to fix the organic traffic issues for this client and not getting very far as yet.
Technical SEO | | LukeHardiman0 -
Another http vs https Question?
Is it better to keep the Transaction/ Payment pages on a commercial website as the only secure ones (https) and remainder of website as http? Or is it better to have all the commercial website as secure (https)?
Technical SEO | | sherohass0 -
New Website and Domain Question
Hi all, I am launching a new website around the end of October and I have purchased a great domain to use for it. My question is should I put some kind of holding page up to try and start building up some domain authority in preperation for launch? Or maybe a blog at www.domain.com/blog and then keep all the blog content at the same location when the full site goes up? Or is it best to wait and just launch the site when the first version is complete? Thanks, Ben
Technical SEO | | BenInder0 -
Redirect questions
Hi! A client of mine have created a new site with a new URL structure which they launched the other day. They have done a 301 redirect on all pages on the old site to the start page on the new site. E.g:
Technical SEO | | lojdqvist
www.olddomain.com/subfolder1/index.html -> www.newdomain.com
www.olddomain.com/subfolder2/index.html -> www.newdomain.com I'm thinking of fixing this now so the redirect instead looks someting like this:
www.olddomain.com/subfolder1/index.html -> www.newdomain.com/newsubfolder1/index.html
www.olddomain.com/subfolder1/index.html -> www.newdomain.com/newsubfolder1/index.html Two questions: 1. Is it worth doing the latter kind of redirect in all cases (after all, it involves quite a lot more work compared to the first solution)? or do you recommend the first solution for all redirect projects?
2. Now that they have already done the first solution, is it at all worth amending this to the latter or is everything spoiled now that they have already gone ahead with the first solution? Many thanks in advance!0 -
Using hyphenated sub-domains or non-hyphenated sub-domains? What is the question! I Any takers?
For our corporate business level domain, we are exploring using a hyphenated sub-domain foir a project. Something like www.go-figure.extreme.com I thought from a user perspective it seems cluttered. The domain length might also be an issue with the new Algorithm big G has launched in recent past. I know with past experience, hyphenated domains usually take longer to index, as they are used by spammers more frequently and can take longer to get out of the supplementary index. Our company site has over 90 million viewers / year, so our brand is well established and traffic isn't an issue. This is for a corporate level project and I didn't have the answer! Will this work? anyone have any experience testing this. Any thoughts will help! Thanks, Rob
Technical SEO | | RobMay0