How to safely reduce the number of 301 redirects / should we be adding so many?
-
Hi All,
We lost a lot of good rankings over the weekend with no obvious cause. Our top keyword went from p3 to p12, for example.
Site speed is pretty bad (slower than 92% of sites!) but it has always been pretty bad. I'm on to the dev team to try and crunch this (beyond image optimisation) but I know that something I can effect is the number of 301 redirects we have in place.
We have hundreds of 301s because we've been, perhaps incorrectly, adding one every time we find a new crawl error in GWT and it isn't because of a broken link on our site or on an external site where we can't track down the webmaster to fix the link. Is this bad practice, and should we just ignore 404s caused by external broken URLs?
If we wanted to reduce these numbers, should we think about removing ones that are only in place due to external broken URLs?
Any other tips for safely reducing the number of 301s?
Thanks, all!
Chris
-
Really helpful. Thanks very much, Sha. Much appreciated!
-
Hi again Chris,
OK, well that makes it interesting!
First, if the list of 301's in your .htaccess file numbers in the hundreds, then there is definitely cause for concern about the effect it might be having on your load times.
The .htaccess file is read from top to bottom until a rule is matched. The first one that is matched will be used and no rule after that will be checked. Obviously, if there are hundreds of rules to check, this can cause processing bottlenecks.
Now there are some things you might be able to do to alleviate the problem IF you have the right conditions.
-
If your rebuild involved the relocation of an entire directory or directories which still contain the same pages as before, you can write a 301 Redirect for all pages in a directory with a single line of code
-
If there is a database behind your site and you have some means of matching pages from the old URL to the new (for example, is there a unique product ID for each page?) then you can use database lookups to write the redirects on the fly. This will eliminate the processing bottlenecks, but can only work if you have a reliable means of matching the pages.
-
If you have only one competition or offer at a time, then you could use database lookups to match any page with say, "competition" in the URL and 301 it to the current competition page (and the same with "offer").
-
For unknown links from external sites you can use a "catch-all" 301 to catch any other page that returns a 404 and send it to a single page (you put this rule at the very end of your .htaccess so it is the last to be matched). You could send these all to home, a generally relevant landing or category page, or a specially designed 404 page. The most important thing is that the page you redirect them to is as relevant as possible, or provides options that may keep the visitor on your site (search, menus etc). I prefer not to send these "random" 404's to home as they are likely to bring an increase in bounce rate. Since bounce is now acknowledged as a ranking factor, I prefer to keep any traffic with a higher probability of bouncing away from pages with the highest Page Authority (PA).
Hope that helps,
Sha
-
-
Yeah, that seems to be the consensus - thanks, Aaron!
Back to the drawing board diagnosing this drop - avg rank over our top 200 keywords has gone from 80-something to 120-something. Not good at all!
Chris
-
Hi Sha,
A combination of things, really. We rebuilt and got rid of a load of legacy pages so there's a few for that reason. We also have a lot of time-limited pages like competitions and special offers that require specific landing pages. The majority are caused by broken links on external sites though - links that never existed, or that have been crawled and added to an autogenerated page incorrectly. The usual nonsense.
So, in brief, there isn't really one cause!
Many thanks for your reply.
Chris
-
In general, 301 aren't bad unless you have a whole string of them. For instance if a 301 redirected to another 301, then another, etc... If 301's are shallow they usually do not present a problem.
-
Hi BaseKit,
What is the reason for there being so many 404's?
Did you move your site or rebuild your site structure? Do you have a lot of pages that are removed after a short time?
The answer to these questions will help to know what is the best approach for your situation.
Sha
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What to do with old content after 301 redirect
I'm going through all our blog and FAQ pages to see which ones are performing well and which ones are competing with one another. Basically doing an SEO content clean up. Is there any SEO benefit to keeping the page published vs trashing it after you apply a 301 redirect to a better performing page?
Technical SEO | | LindsayE0 -
URL structuring / redirect question
Hi there, I have a URL structuring / redirect question. I have many pages on my site but I set each page up to fall under one of two folders as I serve two unique markets and want each side to be indexed properly. I have SIDE A: www.domain/FOLDER-A.com and SIDE B: www.domain/FOLDER-B. The problem is that I have a page for www.domain.com and www.domain/FOLDER-A/page1.com but I do NOT have a page for www.domain/FOLDER-A. The reason for this is that I've opted to make what would be www.domain/FOLDER-A be www.domain.com and act the primary landing page the site. As a result, there is no page located at www.domain/FOLDER-A. My WordPress template (Divi by Elegant Themes) forced me to create a blank page to be able to build off the FOLDER-A framework. My question is that given I am forced to have this blank page, do I leave it be or create a 302 or 307 redirect to www.domain.com? I fear using a 301 redirect given I may want to utilize this page for content at some point in the future. This isn't the easiest post to follow so please let me know if I need to restate the question. Many thanks in advance!
Technical SEO | | KurtWSEO0 -
Site Penalized - 301 Redirect Question
Hello, We have a website that was penalized roughly two years by Google for "Unnatural Links"... We are experiencing a lot of problems with this site, completely unrelated to the penalty or SERPS, and we're debating doing a 301 Re-direct to another site we own that is totally clean and has no "Unnatural Links". If we do a 301 from the penalized site to our alternative website, will there be any cross-contamination? Will the penalty carry over to our other site? Please let me know what you guys think. Thanks
Technical SEO | | Prime850 -
Does any one have experience with SEO and .NET using 301 redirects?
A while ago I altered some of the URL's of my website. Google now thinks that I have two duplicate pages (duplicate content), I have asked my third party web developers (Who use .NET and a custom built CMS system) to simply 301 redirect the old URL to the other. However, my web developers say the following: "Solving the problems by 301 permanent re directs are out of the question as this would create infinite loops. Likely to bring down our server." They also wont do a canonical, as they say there is only one page (but two URLs) Firstly, has any one heard of this before and do they think this is true? Also, does anyone have an alternative method of getting rid of the old URL? Any thoughts would be much appreciated.
Technical SEO | | CoGri0 -
How long after google crawl do you need 301 redirects
We have just added 301's when we moved our site. Google has done a crawl & spat back a few errors. How long do I need to keep those 301's in place? I may need to change some. Thanks
Technical SEO | | Paul_MC0 -
Site blocked by robots.txt and 301 redirected still in SERPs
I have a vanity URL domain that 301 redirects to my main site. That domain does have a robots.txt to disallow the entire site as well. However, for a branded enough search that vanity domain still shows up in SERPs and has the new Google message of: A description for this result is not available because of this site's robots.txt I get why the message is there - that's not my , my question is shouldn't a 301 redirect trump this domain showing in SERPs, ever? Client isn't happy about it showing at all. How can I get the vanity domain out of the SERPs? THANKS in advance!
Technical SEO | | VMLYRDiscoverability0 -
Identifying a 301-redirect problem?
I was looking at the Search Engine Optimization reports for one of my clients in Google Analytics, and I saw that their two biggest landing pages are www.website.com and http://website.com. Does this mean that Google is serving both the 'www' and 'non-www' versions of the website, and thus harming the website's overall ranking? Thanks for any input!
Technical SEO | | williammarlow0 -
302 to 301 redirect
Our site has quite a few 302 redirects that really ought to be 301's. Our IT department is really busy so the question is, given that the 302's have probably been in place for years, is it worth changing them to 301's now? Thanks
Technical SEO | | Houses0