Removing old URLs from Google
-
We rebuilt a site about a year ago on a new platform however Google is still indexing URL's from the old site that we have no control over. We had hoped that time would have 'cleaned' these out but they are still being flagged in HTML improvements in GWT. Is there anything we can do to effect these 'external' dropping out of the indexing given that they are still being picked up after a year.
-
I had kind of a 'derp moment'. If you control the same domain, take a look at your historic traffic (if possible) and definitely do a link audit on the error pages. You can 301 the old pages to their newer, relevant, counterparts.
I would take that tact before removal.
-
Thanks for the helpful responses and yes Ben we are based in old Jersey...
-
Do you mean that it's the same domain and the old URLs are still floating around? If so, you can try removing them via GWT. Just be very careful which URLs you list for removal. Make a list and check it three times.
If these are 404s pointing to your site from external domains, the best you can do is contact the administrator at the site and request they point to a relevant page with a 200 response code.
-
Are you able to access and edit your .HTACCESS file? It should be in the root directory of your domain. If so you should be able to create 301 redirects pointing one URL permanently to another. If not things may be more difficult.
P.S. Are you guys really based in Jersey? It's rather rare to meet a fellow Channel Islander here!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Changing url (permalink) structure of website??
Currently I'm working on SEO of one website www.mocomi.com. I want to change url (permalink) structure of entire website which has more than 5000 pages. Currently website have structure of http://mocomi.com/tenali-raman-the-kings-condition/ Which I want to change it to http://mocomi.com/fun/stories/tenali-raman/tenali-raman-the-kings-condition/ Likewise I want to change entire website permalink url structure to make site architecture more SEO friendly. Which means I'am going to add only categories & subcategories before actual link. Kindly guide with following questions which I need to move forward with this step. How much is it worth to change URL structure? Checklist or factors I need to consider while making this decision? Is it a good practice to change URL's of entire website at once or Should I change it in Parts? How much time it takes google to rank those urls again? Which are the best practices to do so?
On-Page Optimization | | Mocomi1 -
Removing non www and index.php
Hi, I'm green when it comes to altering the htaccess file to remove non www and index.php. I think I've managed to redirect the urls to www however not sure if I've managed to remove the index.php. I'm pasting the contents of the htaccess file here maybe someone can identify if I have unwanted lines of code and if it is up to standard (there are a lot of comments in #) not sure if needed but I've left them as I don't want to screw up anything. Thanks 🙂 @package Joomla @copyright Copyright (C) 2005 - 2016 Open Source Matters. All rights reserved. @license GNU General Public License version 2 or later; see LICENSE.txt READ THIS COMPLETELY IF YOU CHOOSE TO USE THIS FILE! The line 'Options +FollowSymLinks' may cause problems with some server configurations. It is required for the use of mod_rewrite, but it may have already been set by your server administrator in a way that disallows changing it in this .htaccess file. If using it causes your site to produce an error, comment it out (add # to the beginning of the line), reload your site in your browser and test your sef urls. If they work, then it has been set by your server administrator and you do not need to set it here. No directory listings IndexIgnore * Can be commented out if causes errors, see notes above. Options +FollowSymlinks
On-Page Optimization | | KeithBugeja
Options -Indexes Mod_rewrite in use. RewriteEngine On
RewriteCond %{REQUEST_URI} ^/index.php/
RewriteRule ^index.php/(.*) /$1 [R,L] Begin - Rewrite rules to block out some common exploits. If you experience problems on your site then comment out the operations listed below by adding a # to the beginning of the line. This attempts to block the most common type of exploit attempts on Joomla! Block any script trying to base64_encode data within the URL. RewriteCond %{QUERY_STRING} base64_encode[^(]([^)]) [OR] Block any script that includes a0 -
Google is indexing urls with parameters despite canonical
Hello Moz, Google is indexing lots of urls despite the canonical in my site. Those urls are linked all over the site with parameters like ?, and looks like Google is indexing them despite de canonical. Is Google deciding to index those urls because they are linked all over the site? The canonical tag is well implemented.
On-Page Optimization | | Red_educativa0 -
Getting Google to see major changes?
I am making some big changes to my site based on Moz's advice. How can I get google to see all of them? I submitted them to Google Webmasters. Is that the best thing to do? How long until Google updates all my pages?
On-Page Optimization | | dealblogger0 -
What´s the penalization extent applied by Google?
Hi! I still don´t get this web site penalization applied by Google due to duplicate content. My site has many of pages that were among the first positions for top keywords (A Photoshop web site). Those pages were linked by sites like LifeHacker, BoingBoing, Microsiervos, SmashingMagazine, John Nack, and many other well known blogs. After mid February 2012 everything went down the drain. I lost half of my traffic and my well ranked pages are now almost nowhere to be found. I have plenty of ads in some pages of my site, and duplicate content (amazon product description only) in other pages of my site. So, the good quality pages my site has, are no longer considered as good quality just because I have some duplicate content or ad filled pages? I´m not complaining. I´m trying to understand this. Google needs to serve good information to their visitors. But since they found some trash in my site, they decide to remove both the trash and the good information from the search engine? That doesn´t sound logical to me. Why don´t they just remove the trash and leave the good content? Of course, I understand that information is added everyday and some may come up with something better than mine, but dropping 40 or more places in the ranking sounds more like a penalty to me. Again, I´m not complaining (although it sounds like I am!), just want to understand the reasons behind this. Thanks, Enrique
On-Page Optimization | | enriquef0 -
Meta refresh - nojavascript url
seomox is telling me that I am getting a page that is not being indexed or crawled and since the crawl status code is 200 and there are no robots the meta-refresh url must be the problem. the meta refresh url is different than the on page report card url as it's the nojavascript url which my developer says should be ok. see his comments below. The is redirecting to http://mastermindtoys.com/store/nojavascript.html only in case if the JavaScript is disabled in the client browser. This is the right way to do it, I don’t understand why this might be a problem, otherwise MM has to implement Noscript pages that have a real content. I didn’t get what’s wrong about accessibility. The code 200 means it is accessible, and yes there is nothing to access if JavaScript is disabled on browser. I think there are no modern retail sites that would do any sensible business with the scripting disabled in browsers.The H1 is really present 2 times and second occurrence can be removed, though I highly doubt about importance of this change.Regarding duplicates – what URLs are considered duplicates? Can you please send me examples?I am not aware of canonical URL problem for MM site unless we consider old .asp links as duplicate links of the canonical product pages. I would appreciate if SEOMoz gave us an example what they mean.I suspect that the page is not getting indexed as a result of this or I'm just not getting a good score. Which is it?
On-Page Optimization | | mastermindtoys0 -
Should I let Google index tags?
Should I let Google index tags? Positive? Negative Right now Google index every page, including tags... looks like I am risking to get duplicate content errors? If thats true should I just block /tag in robots.txt Also is it better to have as many pages indexed by google or it's should be as lees as possible and specific to the content as much as possible. Cheers
On-Page Optimization | | DiamondJewelryEmpire0