Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Proper 301 in Place but Old Site Still Indexed In Google
- 
					
					
					
					
So i have stumbled across an interesting issue with a new SEO client. They just recently launched a new website and implemented a proper 301 redirect strategy at the page level for the new website domain. What is interesting is that the new website is now indexed in Google BUT the old website domain is also still indexed in Google? I even checked the Google Cached date and it shows the new website with a cache date of today.
The redirect strategy has been in place for about 30 days. Any thoughts or suggestions on how to get the old domain un-indexed in Google and get all authority passed to the new website?
 - 
					
					
					
					
How big is the site in question? How many pages are there to de-index?
What does Google Webmaster Tools tell you about the old domain? Does it show pages being removed from the index over time?
If you do a site:{old domain} query, can you see that the number of results being returned is gradually decreasing?
How have you implemented the redirects?
Have you submitted a change of address request in Webmaster Tools?
On the new website, have you submitted a sitemap fom the old website as well as the new one?
What does the backlink profile on the old domain look like? Can you start to get authoritative links to the old site updated? What about any embedded internal links in your content - have they also been updated?
More guidance from Google here:
 - 
					
					
					
					
It could just be me kchandler, but I've seen it take as long as 8 months for old pages to get purged from Google's index, redirected or not. The redirect and indexing are independent of one another.
 - 
					
					
					
					
Kyle -
Sorry this is so puzzling. The only other thing I could think of is that perhaps the older pages still somehow exist and/or are being served by the server? For example, the .htaccess file might have the /old-page.php redirecting to the /new-page.php... but somehow the old-page.php is still accessible? I'd also look at caching, too? For example, our site, www.CustomerParadigm.com uses varnish for caching, so if we make a change to the site, we need to clear out that page or the change won't be reflected publicly.
Hope this helps?
-- Jeff
 - 
					
					
					
					
Hi Jeff, thank you for the quick response, it is truly appreciated
Unfortunately i am not able to publicly release their URL in forums due to part of our contract. However i can provide some feedback to your ideas.
- Different web servers - the website is the same and on the same hosting platform, they just updated their branding and along with that their domain name
 - WWW. vs non-WWW. - I did a quick check and it looks like both versions of the old domain properly 301 redirect no matter what the subdomain. I am checking that both with my Chrome developer tools as well s checkmyheaders.com.
 - Robots.txt on old server - as it related to my first bullet, it is technically the same website and server the the robots.txt is the same for the new website just reflecting the new domain.
 
Are there any other things that i could look at for a sanity check? I have never seen a website not get de-indexed after a 301 redirect. Do you think i would need to submit something to Google Webmaster Tools for the old URLs/domains?
Regards, Kyle
 - 
					
					
					
					
Without seeing the new and old sites, my first impression is that this might have been caused by having the older site on a different server; the newer site might be on a newer, different Web server. If this is the case, and the older server is still online, I'd check your DNS zone files to make sure that the older site isn't somehow still accessible? I've seen cases where there's two A records for the www. version of a domain; not ideal, but it can cause issues. I'd also set the robot.txt file on the older server / older site to no-index / no-follow.
Hope this helps?
Thanks,
- Jeff
 
 
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
- 
		
		
Moz Tools
Chat with the community about the Moz tools.
 
- 
		
		
SEO Tactics
Discuss the SEO process with fellow marketers
 
- 
		
		
Community
Discuss industry events, jobs, and news!
 
- 
		
		
Digital Marketing
Chat about tactics outside of SEO
 
- 
		
		
Research & Trends
Dive into research and trends in the search industry.
 
- 
		
		
Support
Connect on product support and feature requests.
 
Related Questions
- 
		
		
		
		
		
		
Trying to get Google to stop indexing an old site!
Howdy, I have a small dilemma. We built a new site for a client, but the old site is still ranking/indexed and we can't seem to get rid of it. We setup a 301 from the old site to the new one, as we have done many times before, but even though the old site is no longer live and the hosting package has been cancelled, the old site is still indexed. (The new site is at a completely different host.) We never had access to the old site, so we weren't able to request URL removal through GSC. Any guidance on how to get rid of the old site would be very appreciated. BTW, it's been about 60 days since we took these steps. Thanks, Kirk
Intermediate & Advanced SEO | | kbates0 - 
		
		
		
		
		
		
Google Indexing Request - Typical Time to Complete?
In Google Search Console, when you request the (re) indexing of a fetched page, what's the average amount of time it takes to re-index and does it vary that much from site to site or are manual re-index request put in a queue and served on a first come - first serve basis despite the site characteristics like domain/page authority?
Intermediate & Advanced SEO | | SEO18050 - 
		
		
		
		
		
		
Google Ignoring Canonical Tag for Hundreds of Sites
Bazaar Voice provides a pretty easy-to-use product review solution for websites (especially sites on Magento): https://www.magentocommerce.com/magento-connect/bazaarvoice-conversations-1.html If your product has over a certain number of reviews/questions, the plugin cuts off the number of reviews/questions that appear on the page. To see the reviews/questions that are cut off, you have to click the plugin's next or back function. The next/back buttons' URLs have a parameter of "bvstate....." I have noticed Google is indexing this "bvstate..." URL for hundreds of sites, even with the proper rel canonical tag in place. Here is an example with Microsoft: http://webcache.googleusercontent.com/search?q=cache:zcxT7MRHHREJ:www.microsoftstore.com/store/msusa/en_US/pdp/Surface-Book/productID.325716000%3Fbvstate%3Dpg:8/ct:r+&cd=2&hl=en&ct=clnk&gl=us My website is seeing hundreds of these "bvstate" urls being indexed even though we have a proper rel canonical tag in place. It seems that Google is ignoring the canonical tag. In Webmaster Console, the main source of my duplicate titles/metas in the HTML improvements section is the "bvstate" URLs. I don't necessarily want to block "bvstate" in the robots.txt as it will prohibit Google from seeing the reviews that were cutoff. Same response for prohibiting Google from crawling "bvstate" in Paramters section of Webmaster Console. Should I just keep my fingers crossed that Google honors the rel canonical tag? Home Depot is another site that has this same issue: http://webcache.googleusercontent.com/search?q=cache:k0MBLFcu2PoJ:www.homedepot.com/p/DUROCK-Next-Gen-1-2-in-x-3-ft-x-5-ft-Cement-Board-172965/202263276%23!bvstate%3Dct:r/pg:2/st:p/id:202263276+&cd=1&hl=en&ct=clnk&gl=us
Intermediate & Advanced SEO | | redgatst1 - 
		
		
		
		
		
		
Where is the best place to put a sitemap for a site with local content?
I have a simple site that has cities as subdirectories (so URL is root/cityname). All of my content is localized for the city. My "root" page simply links to other cities. I very specifically want to rank for "topic" pages for each city and I'm trying to figure out where to put the sitemap so Google crawls everything most efficiently. I'm debating the following options, which one is better? Put the sitemap on the footer of "root" and link to all popular pages across cities. The advantage here is obviously that the links are one less click away from root. Put the sitemap on the footer of "city root" (e.g. root/cityname) and include all topics for that city. This is how Yelp does it. The advantage here is that the content is "localized" but the disadvantage is it's further away from the root. Put the sitemap on the footer of "city root" and include all topics across all cities. That way wherever Google comes into the site they'll be close to all topics I want to rank for. Thoughts? Thanks!
Intermediate & Advanced SEO | | jcgoodrich0 - 
		
		
		
		
		
		
Google places keyword variations
Hi all, I have a site that is ranking #1 in Google Places for its main <city><keyword>search... but it does not rank for any of its basic keyword variations, which I find very confusing.</keyword></city> ie (just an example) Chicago Caterer (ranked #1 in google places)
Intermediate & Advanced SEO | | x2264983x
Chicago Caterers (not ranked in google places)
Chicago Catering (not ranked in google places)
Chicago Catering Company (not ranked in google places)
Chicago Catering Companies (etc..) How can I secure a google places ranking for these simple keyword variations? Do I build links to the google plus page using that anchor text? Do I get citations that contain that keyword somewhere on the page? Do I optimize for these keyword variations on the actual website itself? (not the places listing). Obviously I don't stuff these keywords into the google places listing. Any help would be much appreciated!0 - 
		
		
		
		
		
		
How to properly link network of microsites and main sites?
Law firm has a main brand site (lawfirmname.com) with lots of content focusing on personal injury related areas of law. They also do other unrelated areas of law such as bankruptcy and divorce. They have a separate website for bankruptcy and a separate one for divorce. These websites have good quality content, a backlinking campaign, and are fairly large websites, with landing pages for different cities. They also have created local microsites in the areas of bankruptcy and divorce that target specific smaller cities that the main bankruptcy site and divorce site do not target well. These microsites have a good deal of original content and the content is mostly specific to the city the website is about, and virtually no backlinks. There are about 15 microsites for cities in bankruptcy and 10 in divorce and they rank pretty well for these city specific local searches. None of these sites are linked at all, and all 28 of the sites are under the same hosting account (all are subdomains of root domain of hosting account). Question, should I link these sites together at all and if so how? I considered making a simple and general page on the lawfirmname.com personal injury site for bankruptcy and divorce (lawfirmname.com/bankruptcy and lawfirmname.com/divorce) and then saying on the page something to the effect of "for more information on bankruptcy go to our main bankruptcy site at ....." and putting the link to the main bankruptcy site. Same for divorce. This way users can go to lawfirmname.com site and find Other Practice Areas, go to bankruptcy page, and link to main bankruptcy site. Is this the best way to link to these two main sites for bankruptcy and divorce or should I be linking upward? Secondly, should I link the city specific microsites to any of the other sites or leave them completely separate? Thirdly, should all of these sites be hosted on the same account or is this something that should be changed? I was considering not linking the city specific sites at all, but if I did this I didn't know if I should create different hosting accounts for them (which could be expensive). The sites work well in themselves without being linked, but wanted to try to network them in some way if possible without getting penalized or causing any issues with the search engines. Any help would be appreciated on how to network and host all of these websites.
Intermediate & Advanced SEO | | broca777110 - 
		
		
		
		
		
		
How to stop Google crawling after 301 redirect?
I have removed all pages from my old website and set 301 redirect to new website. But, I have verified old website with Google webmaster tools' HTML verification file which enable me to track all data and existence of pages in Google search for my old website. I was assumed that, Google will stop crawling and DE-indexed all pages after 301 redirect. Because, I have set 301 redirect before 3 months. Now, I'm able to see Google bot activity on my website with help of Google webmaster tools. You can find out attachment to know more about it. How can it possible & How Google can crawl removed pages? You can see following image to know more about it. First & Second
Intermediate & Advanced SEO | | CommercePundit0 - 
		
		
		
		
		
		
Old Redirecting Website Still Showing In SERPs
I have a client, a plumber, who bought another plumbing company (and that company's domain) at one point. This other company was very old and has a lot of name recognition so they created a dedicated page to this other company within their main website, and redirected the other company's old domain to that page. This has worked fine, in that this page on the main site is now #1 when you search for the other old company's name. But for some reason the old domain comes up #2 (despite the fact that it's redirecting). Now, I could understand if the redirect had only been set up recently, but I'm reasonably sure this happened about a year ago. Could it be due to the fact that there are many sites out there still linking to that old domain? Thanks in advance!
Intermediate & Advanced SEO | | VTDesignWorks1