Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is there a limit to Internal Redirect?
-
I know Google says there is no limit to it but I have seen on many websites that too many 301 redirects can be a problem and might negatively affect your rankings in SERPs.
I wanted to know especially from people who worked on large ecommerce site. How do they manage internal redirect from one URL to other and how many according to you are too many. I mean if you get a website that contain 300 plus 301 redirections within the website, how will you deal with that?
Please let me know if the question is not clear.
-
Right. Chain redirects = bad.
However, in the same video of Matt Cutts, he does say that the overall amount doesn't matter, and that's what I was talking about in first part of my previous answer.
Now, let's crunch some numbers to show you that the number of no-chain redirects doesn't matter.
-
Assume that we are in perfect world, so all given manufacturer given numbers actually right and all operations per second are actually operations per second

-
Lets say that standard hosting server is 2GHz power = 2*10^9 computations per second
-
Since all htaccess work/computations are strictly on a server side (bots/browsers just send request to server for response if page should be redirected), the only time which can slow down the request is server response time.
-
Match computations are always considered low computation power processes.
-
so, let's say you have htacces with 1 000 000 redirect rules, server keeps it in memory to do match computations when bots make requests, it means that 2GHz server has to have 2000 requests per second to just START struggling.
So, do you have 2000 requests per second to your website and 1 million redirect rules?

P.S. All number above are very rough approximations
P.P.S. If you really wanna see if your server is/ would struggle - login into web host manager, go to server status and info, look and see how much of your server power is usually being used. Usually that number is lower than 6-7% at 90% of the time.
Hope this clarify some things

-
-
I am going to say what I do and how I think that it works. I am not saying that this is correct or best practice.
When I abandon a URL I do not place the redirect in the .htaccess file in the root directory. Instead, I place an .htaccess file in the folder where the URL was saved. That limits the size of my .htaccess file in the root directory. I believe that reduces the amount of work that your server must do, it does not need to examine a very large .htaccess file.
If you use many folders to categorize your content then you will have small .htaccess files that are easier to manage. From time to time you will be able to redirect entire folders instead of individual files when you abandon a product line or a category of content.
That's what I do.
-
I think you get me wrong, you are talking probably talking about chain redirects as in from a to b and than b to c and may be d. For this Matt himself said in one of his Webmaster videos that Google might not crawl the link after 2 or may be 3 stages.
I am more concern about redirects in total because redirect increase the page load time and page load time is a factor in Google rankings which make me think again before go ahead and set 1000+ redirection (for example)!
But thanks for your reply!
-
I've not seen any instances of a limit to how many redirects you can have pointing to your website. I have some clients who have thousands of redirects in place (lots of old pages being moved to a new version of that product). Those sites haven't had any issues with rankings at all. In fact, many of the links pointing to the sites still reference the URLs that are redirected and those pages that are redirected to are ranking perfectly fine.
The biggest limit I've seen is on chaining. I've seen issues where chained redirects simply aren't followed. However, if you can keep it to a 1 step redirect, or 2, then things should be okay. It doesn't sound like that is what you are asking about though. More from Matt Cutts on this:
http://www.searchenginejournal.com/matt-cutts-discusses-301-permanent-redirects-limits-on-websites/46611/In terms of managing those redirects, you can't usually keep this many on an htaccess file without going a little bit nuts (or risking some future dev deleting those in an effort to clean up the htaccess file - ug). If you are using WordPress, the 301 redirects plugin works quite well: https://wordpress.org/plugins/301-redirects/
Unfortunately, I've also run into sites that aren't in a CMS where you can use a plugin. In those cases, I usually put these redirects in a database table. On the 404 file, I then have the code check the would-be error URL to see if we need to redirect that URL somewhere else. If a redirect is place, it redirects instead of throwing the 404 error. If no redirect is in place, the code then throws a 404 error.
Hope that helps.
-
Hello, my friend.
Well, whenever people says "don't have too many redirects", it doesn't mean not to have too many redirects in total count, for example, if you have old page
a.php redirected to b.php,
and old page c.php redirected to d.php
and so on - there is no any problem. However, what they mean is not to have consecutive redirects - eg.:a.php redirects to b.php, which redirects to c.php, which redirects to d.php, instead of a.php redirecting to d.php straight forward.
Hope this helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do you fix redirect chains and temporary redirects?
Hi, I have a lot of issues popping up with temporary redirects and redirect chains. I'm still confused as to what exactly redirect chains are and I don't know how to find where the "chains" are or how to fix them. I'm having two issues mainly:1. Temporary RedirectsI have around 100 pages on our www.twowayradiosfor.com website that are being flagged as temporary redirects. All of them have one thing in common: they are review pages (basically, when a customer clicks on the Review button to review a certain product, they are redirected to a review page for that product).URL Example: https://www.twowayradiosfor.com/reviewhelpful.asp?ProductCode=CLS1410-COMBO&ID=44&yes=noI went into our website and set any URL containing the following as noindex:/review.aspWill that fix the issue? If yes, will I also need to do that for any URL containing /reviewhelpful.asp?2. Redirect ChainsIt seems like basically every product page on my website has this issue (over 100 pages). Here's an example of one:https://www.twowayradiosfor.com/Motorola-CLS1110-p/cls1110.htmI don't see any broken links on this page or links that redirect to another page that redirects, etc. What is causing this? Is it something on my header bar that is redirecting (since that header bar appears on every page, maybe that is why this issue shows up on a lot of pages)?I am new to Moz and still trying to figure this stuff out. I really appreciate any help. Thanks, Sawyer
Technical SEO | | AllChargedUp0 -
301 redirects delay in picking up
Hi I have been involved in the redesign/development of a website which has up until now had a lot of international traffic. On day of migration I uploaded all the 301 redirects to the website (wordpress) using Simple 301 redirect plugin. I tested a number of them and they appeared to be working. I also submitted the new sitemaps to Search Console. Since migration international traffic - particularly from countries such as india, Phillipines, Sri Lanka etc have significantly dropped off whereas the local traffic and some of the international traffic such as USA has remained fairly consistent. Looking at Analytics and entrances recently it appears as though search results are/were showing a number of pages with 404's (one in particular which received significant traffic and for which I had created a 301 redirection) - I have checked this page using the old url and it re-directs correctly for me and today asked a colleague in India to also check - he is getting the redirection fine. Does Google.in take a significantly longer time to pick these up in search results? Or am I missing something?
Technical SEO | | musthavemarketing0 -
Robots txt. in page with 301 redirect
We currently have a a series of help pages that we would like to disallow from our robots txt. The thing is that these help pages are located in our old website, which now has a 301 redirect to current site. Which is the proper way to go around? 1- Add the pages we want to disallow to the robots.txt of the new website? 2- Break the redirect momentarily and add the pages to the robots.txt of the old one? Thanks
Technical SEO | | Kilgray0 -
Redirect URLS with 301 twice
Hello, I had asked my client to ask her web developer to move to a more simplified URL structure. There was a folder called "home" after the root which served no purpose. I asked for the URLs to be redirected using 301 to the new URLs which did not have this structure. However, the web developer didn't agree and decided to just rename the "home" folder "p". I don't know why he did this. We argued the case and he then created the URL structure we wanted. Initially he had 301 redirected the old URLS (the one with "Home") to his new version (the one with the "p"). When we asked for the more simplified URL after arguing, he just redirected all the "p" URLS to the PAGE NOT FOUND. However, remember, all the original URLs are now being redirected to the PAGE NOT FOUND as a result. The problems I see are these unless he redirects again: The new simplified URLS have to start from scratch to rank 2)We have duplicated content - two URLs with the same content Customers clicking products in the SERPs will currently find that they are being redirect to the 404 page. I understand that redirection has to occur but my questions are these: Is it ok to redirect twice with 301 - so old URL to the "p" version then to final simplified version. Will link juice be lost doing this twice? If he redirects from the original URLS to the final version missing out the "p" version, what should happen to the "p" version - they are currently indexed. Any help would be appreciated. Thanks
Technical SEO | | AL123al0 -
301 Redirects in subfolders
Hi, we're making our site into a static site but I would like to transfer the Google juice. Most of the links and database exist on subfolders though. Could I simply do 301 redirects on the subfolders and retain the value or does it have to be on the full domain?
Technical SEO | | Therealmattyd0 -
301 Redirect with index.asp
I am very new to all of this so forgive the newbie questions I will get better. Ok so after starting a campaign I see that I have many issues including where some pages are being deemed as duplicate content. 1. The report says the http://lucid8.com has duplicate content on 2 other pages 2. When I look at them it shows that http://lucid8.com/index.asp and http://www.lucid8.com are duplicates. 3. Really these are the exactly the same page because the default page that is opened for www.lucid8.com http://www.lucid8.com etc always opens the index.asp page. 4. Now I read that I should do permanent redirects and how to do this via IIS and I tried to do a redirect from index.asp to www.lucid8.com but that does not work because www.lucid8.com is pointing to index.asp and so we end up in a circle. So the question is how do I get rid of these duplicate page references without causing problems. Thanks
Technical SEO | | TroyW0 -
Htaccess redirect with question mark
Hi I have a problem setting up my htaccess for a specific page that has a question mark in the link, and one that has a space in the link and also a question mark. So I would like 2 redirects in my htaccess like that: www.olddomain.com/page.php?page=pagename1 to www.newdomain.com/newpage1.html www.olddomain.com/page.php?page=page name2 to www.newdomain.com/newpage2.html I have tried with something like this but doesn't work: RewriteEngine on RewriteRule ^page.php?page=pagename1 "http://www.newdomain.com/newpage1.html" [R=301,L] RewriteRule ^page.php?page=page name2 "http://www.newdomain.com/newpage2.html" [R=301,L] Could someone tell me what exactly I have to change? Thanks
Technical SEO | | darkanweb0 -
Do search engines treat 307 redirects differently from 302 redirects?
We will need to send our users to an alternate version of our homepage for a few hours for a certain event. The SEO task at hand is to minimize the chance of the special homepage getting crawled and cached in the search engines in place of our normal homepage. (This has happened in the past so the concern is not imaginary.) Among other options, 302 and 307 redirects are being discussed. IE, redirecting www.domain.com to www.domain.com/specialpage. Having used 302s and 301s in the past, I am well aware of how search engines treat them. A 302 effectively says "Hey, Google! Please get rid of the old content on www.domain.com and replace it with the content on /specialpage!" Which is exactly what we don't want. My question is: do the search engines handle 307s any differently? I am hearing that the 307 does NOT result in the content of the second page being cached with the first URL. But I don't see that in the definition below (from w3.org). Then again, why differentiate it from the 302? 307 Temporary Redirect The requested resource resides temporarily under a different URI. Since the redirection MAY be altered on occasion, the client SHOULD continue to use the Request-URI for future requests. This response is only cacheable if indicated by a Cache-Control or Expires header field. The temporary URI SHOULD be given by the Location field in the response. Unless the request method was HEAD, the entity of the response SHOULD contain a short hypertext note with a hyperlink to the new URI(s) , since many pre-HTTP/1.1 user agents do not understand the 307 status. Therefore, the note SHOULD contain the information necessary for a user to repeat the original request on the new URI. If the 307 status code is received in response to a request other than GET or HEAD, the user agent MUST NOT automatically redirect the request unless it can be confirmed by the user, since this might change the conditions under which the request was issued.
Technical SEO | | CarsProduction0