Question on 301s
-
Hi Everyone,
I have a questions on 301 redirects, i hope someone can give me some help on this.
There was some 301 redirects made on some of the URLs at the beginning of the year, however we are now re-structuring the whole website, which means the URLs which had been given a 301 redirect are now getting another 301.
The question is, should i delete the first 301 redirect from the htaccess file?
Kind Regards
-
Ryan your analogy is fantastic. I totally understand this now and it really makes sense to do it this way.
Thanks for being patient with me
Again thanks all for your feedback on this.
Kind Regards
-
Every URL which is no longer active would require a 301 redirect to the proper page. In the situation you describe:
/a should redirect to /abc
/ab should redirect to /abc
I recognize this seems confusing so forget it's a website for a moment. Think of it as mail after you move.
You lived at 100 Main Street. That is where you received your mail. Now you move to 200 Elm Street. You put in a forward order with the post office (a real world equivalent to a 301 redirect). Now any mail addressed to 100 Main Street will be received at 200 Elm Street.
Now you move again to 300 Wall Street. You would put in another forwarding order so your mail from 200 Elm Street gets delivered to your new address. This solution is fine BUT, your mail from 100 Main Street would be delayed. First it would get forwarded to the 200 Elm Street post office, who would then have to forward it to 300 Wall Street. This process is inefficient (in seo terms, you lose link juice).
You want to change your 100 Main Street forward order to direct your mail to the 300 Wall Street address. Now all of your mail is taken to the proper location in a single hop.
I hope this analogy helps!
-
What happens to the URL
If there are external backlinks going to the URL, are these not going to get lost?
Because as we have mentioned on these 301s, there has been 3 URLs in question.
Hope that makes sense.
-
In the simplest terms, the old page should always be directed to the new page. Think of it as a non-stop flight.
-
Hi Ryan,
Thanks for your feedback, however I am getting a little lost
So what your are saying if I understand is, the 301 should be this:
example.com/a is redirected to example.com/abc
Kind Regards
-
The only thing that concerns me is what CafePress had said "Google stops crawling a link after the 5th redirect or so."
You can offer 100 links on a page. All the links can be to "seomoz.org" and they will all be crawled even though the real URL is "www.seomoz.org" and all 100 links will get redirected.
What CafePress referred to is redirects for a single URL.
www.example.com/a redirects to /ab which redirects to /abc and so forth. A crawler will only follow a single URL so far through a chain of redirects before the PR is completely gone and it stops.
Therefore the preferred solution is to redirect any old or broken URLs to their new URL in a single redirect. I'll share an example based on your site:
Very old URL: example.com/a. It is redirected to example.com/ab
Old URL: example.com/ab. It is redirected to example.com/abc
You could leave these two redirects in place, as-is, and they will work, but it is not recommended. The reason is any traffic to /a will have a double re-direct. First the traffic will go to /ab then to the final destination of /abc. This double redirect is an unnecessary delay, it adds extra points of vulnerability and is a waste of SEO link juice. The preferred solution would be to modify the /a redirect to point to the /abc page directly.
I hope that makes sense.
-
Also, if a page is indexed, which is highly likely (due to XML sitemaps, Google Analytics, Google Toolbar etc), then just removing the 301 redirect (links or no links) means that when this page disappears due to the site changes then you will have an indexed page resulting in a 404 error.
I maintain that you should have single hop 301 redirects on all of the pages that will not be there or will have been moved due to the site updated.
I also agree with what Ryan Kent says about links - you may have some links that have been discovered but not yet recognized pr picked up. If there is a chance that the content has been indexed then it should have an appropriate redirect.
-
Hi Ryan,
The only thing that concerns me is what CafePress had said "Google stops crawling a link after the 5th redirect or so."
I have another issue regarding the 301 re-directs:
We have:
/abcd http://www.example.com/abcde this is actually a 301 on a product page, however we have the same product in a shop page /shop/abcd which we have decided to do away with the shop directory, is it best practice to also do a 301 from the /shop/abcd to /abcde?
Hope that makes sense.
Kind Regards
-
I don't agree with the recommendation to simply delete the 301 due to no visible links. There are two reasons why:
1. It is more work for you to go and research the links to each page
2. There can always be links you are not aware of such as bookmarks, e-mail links, links which don't show up for various reasons, etc.
Just simply modify the 301 to point to the correct URL and you are all set.
-
Thanks for the fantastic feedback.
An example of what has happened on the .htaccess:
/abc http://www.example.com/abcd - This is the 301 that was made in March this year.
/abcd http://www.example.com/abcde - This is the new 301
If i notice that there are no links going to /abc using Open Site Explorer should i just delete this 301?
Kind Regards
-
I would change the original 301 redirect to the new location.
I would then add an additional 301 redirect to the secondary page (the old redirect) to the new location.
So you will have your original URL and the older redirected URL both 301 redirected to where the content now resides. This way you only have one hop on the 301 redirects and you have both old URLs pointing to the new one.
-
should i delete the first 301 redirect from the htaccess file?
The best results would be achieved if each URL had a single 301 redirect to the target page. To that end, yes, you should delete the old 301 redirect and create a new one.
-
+1
Totally forgot about mentioning the inbound links part. Thanks for picking it up, Rick!
-
Hey Gary,
I partially agree with Cafe. However, I wouldn't remove any redirects for URLs which may have backlinks. Maybe it would be a good idea to figure out if any of the redirects which you are removing are from URLs that have earned links? An Open Site Explorer link export would help you figure out if any of those URLs still have value.
-
Hi Gary,
Yes, it is always a good idea to cut down the number of 301 redirects (or any redirects in general) because if I remember correctly, Google stops crawling a link after the 5th redirect or so. You also lose another 10% link juice for each additional redirect.
Lastly, don't forget to 301 redirect the URLs from the beginning of the year to the new re-structured website.
Hope that helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
301 Redirect Question
I am working with a website and I ran a Screaming Frog and noticed there are 4,600 301's on the website (www.srishoes.com). It seems like the issue is between the www. and without it and they aren't working together. Is this something that the website provider should update and what type of impact might this have on the site? Thanks!
Technical SEO | | ReunionMarketing
Matt0 -
HTTPS & 301s
Hi We have like most set up a redirect from HTTP to HTTPS. We also changed our website and set up redirects from .ASP pages to PHP pages We are now seeing 2 redirects in place for the whole of the website.
Technical SEO | | Direct_Ram
http.www.domain.com > https.www.domain.com (1) >> oldwebpage.asp >> new webpage.php (2) The question is: Is there anyway of making the redirect 1 and not 2? thanks
Enver0 -
One more redirect question
If there are two URLs like below: example.com/toys/batman-toys
Technical SEO | | IceIcebaby
example.com/birthday/batman-toys Both have the exact same everything, except URL key. The first example ranks for all KWs and search terms in the SEs. Does having the second page hurt my ranking potential for the first page? Should I redirect the 2nd page to the first or just leave it? As always, thanks for your help.0 -
Canonical Expert question!
Hello, I am looking for some help here with an estate agent property web site. I recently finished the MoZ crawling report and noticed that MoZ sees some pages as duplicate, mainly from pages which list properties as page 1,2,3 etc. Here is an example: http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=2
Technical SEO | | artdivision
http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=3 etc etc Now I know that the best practise says I should set a canonical url to this page:
http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=all but here is where my problem is. http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=1 contains good written content (around 750 words) before the listed properties are displayed while the "page=all" page do not have that content, only the properties listed. Also http://www.xxxxxxxxx.com/property-for-rent/london/houses?page=1 is similar with the originally designed landing page http://www.xxxxxxxxx.com/property-for-rent/london/houses I would like yoru advise as to what is the best way to can url this and sort the problem. My original thoughts were to can=url to this page http://www.xxxxxxxxx.com/property-for-rent/london/houses instead of the "page=all" version but your opinion will be highly appreciated.0 -
Another http vs https Question?
Is it better to keep the Transaction/ Payment pages on a commercial website as the only secure ones (https) and remainder of website as http? Or is it better to have all the commercial website as secure (https)?
Technical SEO | | sherohass0 -
Craw Diagnostics Questions
SEO Moz is reporting that I have 50+ pages with a duplicate content issue based on this URL: http://www. f r e d aldous.co.uk/art-shop/art-supplies/art-canvas.html?manufacturer=178 But I have included this tag in the source: rel="canonical" href="http://www.f r e daldous.co.uk/art-shop/art-supplies/art-canvas.html"/> (I have purposefully added white space to the URLs in this message as I'm not sure about the rules for posting links here) I though this "canonical" tag prevented the duplicate content being indexed? is the reporting by SEOMoz wrong or being over cautious?
Technical SEO | | niallfred0 -
Redirect question
I would like to redirect http://example.com/index.html to http://www.example.com/ Is the code below correct ? RewriteEngine on RewriteCond %{HTTP_HOST}^example.comRewriteRule (.*) http://www.example.com/$1 [R=301,L] RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /index.html\ HTTP/ RewriteRule ^index.html$ http://www.example.com/ [R=301,L]
Technical SEO | | seoug_20050 -
Robots.txt questions...
All, My site is rather complicated, but I will try to break down my question as simply as possible. I have a robots.txt document in the root level of my site to disallow robot access to /_system/, my CMS. This looks like this: # /robots.txt file for http://webcrawler.com/
Technical SEO | | Horizon
# mail webmaster@webcrawler.com for constructive criticism **User-agent: ***
Disallow: /_system/ I have another robots.txt file in another level down, which is my holiday database - www.mysite.com/holiday-database/ - this is to disallow access to /holiday-database/ControlPanel/, my database CMS. This looks like this: **User-agent: ***
Disallow: /ControlPanel/ Am I correct in thinking that this file must also be in the root level, and not in the /holiday-database/ level? If so, should my new robots.txt file look like this: # /robots.txt file for http://webcrawler.com/
# mail webmaster@webcrawler.com for constructive criticism **User-agent: ***
Disallow: /_system/
Disallow: /holiday-database/ControlPanel/ Or, like this: # /robots.txt file for http://webcrawler.com/
# mail webmaster@webcrawler.com for constructive criticism **User-agent: ***
Disallow: /_system/
Disallow: /ControlPanel/ Thanks in advance. Matt0