URL no longer ranks after redesign, everything was done by the book. Can someone plz help?
-
Hey Mozzers,
After a redesign of our site we have a huge problem with 1 particular URL. We can't understand what is wrong with it. Our site was redesigned about 3 weeks ago everything was done to make sure the process was smooth. Content on the new url was exactly the same as before, title tags and meta descriptions were copy pasted, 301 redirects were set up. Our site has 47 pages and all other pages are still ranking the same way as they used to before the update but this one url has just dropped from serps or once a while is appearing on 2nd, 3rd pages.
URL in question is: https://dcacar.com/dulles-limo-service/
We are a MOZ pro client and have the data on where this url was ranking prior to the update. Keywords it ranked for #1position on page one- some examples ( car service to dulles, car service to iad, dulles airport limo service) and many more. The only thing that has changed was the url the old url was https://www.dcacar.com/dulles-limo-service.html
After the new design we got rid of the ".html" on all our urls, the funny thing is that all our 46 urls were updated the same way but only this url is experiencing this problem. Google search console shows that url is indexed and on Google but sometimes it does not show up on serps.
Can anyone please help? 20% of our business comes from this url and we are loosing real money everyday and I can't figure out was is wrong with it.
Thanks all,
Davit
-
This echoes what Thomas said but I'll weigh in anyway. This is your old URL: https://www.dcacar.com/dulles-limo-service.html. You say it has been redirected to the new page with a 301, which is only partially accurate. You need to install this Chrome extension which will allow you to inspect redirects and chains: https://chrome.google.com/webstore/detail/redirect-path/aomidfkchockcldhbkggjokdkkebmdll?hl=en. Visit the old URL (http://www.dcacar.com/dulles-limo-service.html) and you will see that it redirects no less than four times (instead of just once). Google probably didn't follow your redirect all the way through. This is your page's redirect path: https://www.dcacar.com/dulles-limo-service.html (301) -> https://dcacar.com/dulles-limo-service.html (301) -> https://dcacar.com/dulles-limo-service.html/ (301) -> https://dcacar.com/dulles-limo-service (301) -> https://dcacar.com/dulles-limo-service/ (200) . Redirects should be A to B, simple. They shouldn't chain up like this
-
Hi Thomas, we are working on fixing the issues you mentioned right now. I will update you if it works. Thanks for your detailed analysis. It means the world to us.
-
You are facing problem due to redirection rule. You are saying you are redirecting as 301 but actually status code is 302. So you need to solve this issue first. http://prntscr.com/q9fbah
-
Hi Davit, I found the problem fix your servers redirect chains
Google recommends to not have more than 5 hops in a redirect chain. We found 6 hops. This could mean slower or delayed crawling as well as potential SEO problems caused by this.
Look https://i.imgur.com/w40qS3b.jpg
-
Your .html page
-
Fix your nginx config file.
-
https://cobwwweb.com/remove-html-extension-and-trailing-slash-in-nginx-config
Then this hits a 302 (will not pass Link equity)https://dcacar.com/dulles-limo-service.html/
| 302 | 302 redirect to:https://dcacar.com/dulles-limo-service |
https://stackoverflow.com/questions/5730092/how-to-remove-html-from-url
Might be faster to use https://support.cloudflare.com/hc/en-us/articles/224509547-Recommended-Page-Rules-to-Consider
<code>RewriteCond%{REQUEST_FILENAME}!-f RewriteCond%{REQUEST_FILENAME}!-d</code>
checks that if the specified file or directory respectively doesn't exist, then the rewrite rule proceeds:
`RewriteRule^(.*)\.html$ /$1 [L,R=301]` [https://www.aleydasolis.com/htaccess-redirects-generator/page-with-to-without-extension/](https://www.aleydasolis.com/htaccess-redirects-generator/page-with-to-without-extension/)
<ifmodule mod_rewrite.c="">RewriteEngine on
RewriteRule ^/?(.*).(html)$ /$1 [R=301,L]</ifmodule><ifmodule mod_rewrite.c="">RewriteEngine on
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule ^(.*)/$ /$1 [L,R=301]</ifmodule>Remove HTML Extension
There are a few different ways to go about removing the
.html
extension. I've found the following to work just fine:<code>server { rewrite ^(/.*)\.html(\?.*)?$ $1$2 permanent; }</code>
Then we have to make sure Nginx knows what files to look for, and for that we use the
try_files
directive. We'll look for a file with the current$uri
and an.html
extension, and if no file exists, we check for a directory with that name and serve the index. Otherwise, we render a404
error.<code>server { rewrite ^(/.*)\.html(\?.*)?$ $1$2 permanent; index index.html; try_files $uri.html $uri/ $uri =404; }</code>
Remove Trailing Slashes
After that first step, we have a URL without a
.html
extension. However, if the file of interest was anindex.html
file, it could still be accessed via the name of the parent directory with a trailing slash.For me, for example, I was getting a URL at http://cobwwweb.com/page/2/, but I didn't want the trailing slash.
We have to remove the trailing slash after we have removed the
.html
extension.The gotcha here is that we have to alter the
try_files
directive to look for anindex.html
file first.<code>server { rewrite ^(/.*)\.html(\?.*)?$ $1$2 permanent; rewrite ^/(.*)/$ /$1 permanent; index index.html; try_files $uri/index.html $uri.html $uri/ $uri =404; }</code>
Pulling It Together
To pull it all together, I'll share a slightly-altered version of my config for this site.
<code>server { listen 80; server_name cobwwweb.com *.cobwwweb.com; rewrite ^(/.*)\.html(\?.*)?$ $1$2 permanent; rewrite ^/(.*)/$ /$1 permanent; root /path/to/project/root; index index.html; try_files $uri/index.html $uri.html $uri/ $uri =404; error_page 404 /404.html; error_page 500 502 503 504 /500.html;</code>
Hey, regular site audit we'll find out more these problems. do you want one?
Hope this helps,
Let me know if not,
Tom
-
-
Hi Davit,You need to link internally from your current site
-
The page is an orphan according to Moz and only has when the mean pointing to it according to link research tools
-
You are #5 in the SERPS as on now I searched from New England
-
Google is not indexing the full page.
-
You change from Wordpress to?
-
https://web.archive.org/web/20170928065156/http://www.dcacar.com/dulles-limo-service.html
-
LRT Power3
-
3
-
LRT Trust1
-
Backlinks 2M
-
Ref Domains 510
-
PAGE
-
LRT Power3
-
6
-
LRT Trust2
-
Backlinks 2
-
Ref Domains 1
-
Links
-
I will run a test and post if that's ok?
-
Have you checked Google search console will you try to make sure that it crawls it again. It appears that it's not very high priority page considering the last time crawled by Google was November 20
Hope this helps,Tom
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Discrepancy in keyword ranking from webmasters and actual ranking.
I have been tracking ranks of some keywords important to my business since the last 2 months. Recently I have observed that, for one of my keywords, google webmasters is giving the avg position as 8 but when i search in google it comes in the 6th page. I know that webmasters tools gives the average position but i do not think there will be such big difference in the ranks. Please help.Thanks.
Intermediate & Advanced SEO | | seomoz12320 -
Alternative Markup Challenge. Can anyone help?
I have a challenge around alternative markup. We currently operate a single domain with geo-targeted folders and alternative markup implemented. We are now now looking to expand this out to non-English content. Current Implementation; All generic English language content hosted on the main domain, with x5 English language content variations (locales) available under a folder structure (.com/en-us/ etc.). Alternative markup is in place for all locales within the HTML, implemented automatically by developers via the CMS. Locale folders geo-targeted via GWT and Bing WT. Planned Launch; Introduction of 5 new non-English locale folders (e.g. /de-de/ etc.), targeted to their respective country and language. Content language will be mixed, with around 1/10 of pages translated and the other 9/10 of pages (business listings) having their body content remain in English, with headers / footers translated. Locale folders will be geo-targeted via GWT and Bing WT. Folder and markup usage TBC. Options; Folders; Implement folder structure /de/, attempting to indicate country but not language (issue; usually a single identifier indicates language, not country?). Implement /de-de/ folder structure to match the English locales and maintain correct country targeting (issue; some content is not in language). Alternative markup; Do not make use of markup at all. Implement CMS based automated markup on all English and non-English content throughout the locale (e.g. /de-de/), but exclude English language versions (e.g. /en-gb/). Attempt manually implementing markup to bridge the English and non-English locales, potentially creating future issues with new content going live and content being removed. A heavy risk. Current approach is webmaster tools targeting, a /de-de/ folder structure and automated implementation of markup. This means English language URLs will have markup and non-English language URLs will have markup, but they will not match up (e.g. English pages will never have markup for non-English language content). If you minds haven't melted, what's your thoughts? Any help is much appreciated.
Intermediate & Advanced SEO | | HelloAlba0 -
Does having shorter URLs help with rankings?
Hello here.I own an e-commerce website (virtualsheetmusic.com), and some of our most important category pages have pretty long URLs. Here is an example: http://www.virtualsheetmusic.com/downloads/Indici/Violin.html I am evaluating the possibility to shorten URLs like the above to something like: http://www.virtualsheetmusic.com/violin/ But since it is going to pretty hard and time consuming (considering the custom system we have in place on that site), I am trying to find out if it really matters and worth doing it from a SEO stand point. I am aware that from a user prospective shorter URLs are preferable, and we plan to pursue a better URL architecture on our website in the near future just for that, but this question, at the moment, should be strictly related to SEO. Any thoughts on this topic are very welcome!
Intermediate & Advanced SEO | | fablau0 -
Help with htaccess
I just setup a WP install in a subfolder: domain.com/development/ However, there is an existing htaccess file in the root which contains the following: RewriteRule ^([A-Za-z_0-9-]+)$ /index.php?page=$1 [QSA]
Intermediate & Advanced SEO | | SCW
RewriteRule ^([A-Za-z_0-9-]+)/$ /index.php?page=$1 [QSA]
RewriteRule ^([A-Za-z_0-9-]+)/([a-z]+)$ /index.php?page=$1&comp=$2 [QSA]
RewriteRule ^([A-Za-z_0-9-]+)/([a-z]+)/$ /index.php?page=$1&comp=$2 [QSA] I need to leave the rules as-is due to the nature of CMS (not WP) under the root domain. Is it possible to include an exception or condition which allows URL requests containing /development/ to resolve to that folder? I tried to add: RewriteRule ^development/([A-Za-z_0-9-]+)$ /development/index.php?page=$1 [QSA] but this seems to send it in a loop back to the root. Thanks!!!0 -
Can URLs blocked with robots.txt hurt your site?
We have about 20 testing environments blocked by robots.txt, and these environments contain duplicates of our indexed content. These environments are all blocked by robots.txt, and appearing in google's index as blocked by robots.txt--can they still count against us or hurt us? I know the best practice to permanently remove these would be to use the noindex tag, but I'm wondering if we leave them they way they are if they can still hurt us.
Intermediate & Advanced SEO | | nicole.healthline0 -
How can Google index a page that it can't crawl completely?
I recently posted a question regarding a product page that appeared to have no content. [http://www.seomoz.org/q/why-is-ose-showing-now-data-for-this-url] What puzzles me is that this page got indexed anyway. Was it indexed based on Google knowing that there was once content on the page? Was it indexed based on the trust level of our root domain? What are your thoughts? I'm asking not only because I don't know the answer, but because I know the argument is going to be made that if Google indexed the page then it must have been crawlable...therefore we didn't really have a crawlability problem. Why Google index a page it can't crawl?
Intermediate & Advanced SEO | | danatanseo0 -
My site rank is not consistent. Once it at first page , then for the next week it is not found in top 100 position. Again two/ three weeks later it ranked automatically without any work. Why this is happening?
Here's the following are available in my site: robot.txt file is included sitemap available Natural link building going on. in a week total 100 links we are creating. 30 social bookmarks, 30 directory submission, 20 blog comments, 20 forum links All the blog and forum links are from relevant sources. Please help me ..
Intermediate & Advanced SEO | | coldfireinc0 -
How can i get high ranking on yahoo?
My search terms are highly visible on Google but is hardly visible on Yahoo. This has been the case for a few months. What can I do to boost my ranking on Yahoo? Thanks!
Intermediate & Advanced SEO | | csknight0