Sitemap issue? 404's & 500's are regenerating?
-
I am using the WordPress SEO plugin by Yoast to generate a sitemap on http://www.atozqualityfencing.com. Last month, I had an associate create redirects for over 200 404 errors. She did this via the .htaccess file. Today, there are the same amount of 404s along with a number of 503 errors. This new Wordpress website was constructed on a subdirectory and made live by simply entering some code into the .htaccess file in order to direct browsers to the content we wanted live. In other words, the content actually resides in a subdirectory titled "newsite" but is shown live on the main url.
Can you tell me why we are having these 404 & 503 errors? I have no idea where to begin looking.
-
You likely have a .htaccess issue causing a rewrite error. You may want to examine or replace your .htaccess with a default. Also, I've seen some plugins cause this error.
What is happening is this:
http://www.atozqualityfencing.com/newsite
is sent to:
http://www.atozqualityfencing.com/newsite/
Note the trailing slash.
But that page is returning a 404 error.
If I go to
http://www.atozqualityfencing.com/newsite/index.php it redirects to
http://www.atozqualityfencing.com/newsite/
So there is likely something wrong in the redirect rules. I would try disabling all plugins. If that fails, compare the current htaccess to a default one and remove any modifications.
.
-
Wondering if anyone else out there has some insight as to whether the information in my previous post seems to be correct.
-
Oye, Jeff - this is a little bit over my head so bear with me as I work it through.
I went to redbot.org and entered the url of where the main website is actually living (http://www.atozqualityfencing.com/newsite). I received this information:
HTTP/1.1 301 Moved Permanently Date: Sun, 24 Aug 2014 14:56:10 GMT Server: Apache Location: [http://www.atozqualityfencing.com/newsite/](https://redbot.org/?uri=http://www.atozqualityfencing.com/newsite/&req_hdr=Referer%3Ahttp://www.atozqualityfencing.com/newsite) Cache-Control: max-age=3600 Expires: Sun, 24 Aug 2014 15:56:10 GMT Content-Length: 326 Keep-Alive: timeout=5, max=100 Connection: Keep-Alive Content-Type: text/html; charset=iso-8859-1 When I clicked on the url listed under Location above, I receive the following information:
HTTP/1.1 404 Not Found Date: Sun, 24 Aug 2014 14:59:59 GMT Server: Apache X-Pingback: http://www.atozqualityfencing.com/newsite/xmlrpc.php Expires: Wed, 11 Jan 1984 05:00:00 GMT Cache-Control: no-cache, must-revalidate, max-age=0 Pragma: no-cache Vary: Accept-Encoding,User-Agent Keep-Alive: timeout=5, max=100 Connection: Keep-Alive Transfer-Encoding: chunked Content-Type: text/html; charset=UTF-8
This has me confused and I wondering if the method used for making the revised website is either not good or is missing something. Here are the articles that were followed for "moving" the newsite redesign to the live url. ``` [http://codex.wordpress.org/Giving_WordPress_Its_Own_Directory](http://codex.wordpress.org/Giving_WordPress_Its_Own_Directory) [http://codex.wordpress.org/Moving_WordPress#When_Your_Domain_Name_or_URLs_Change](http://codex.wordpress.org/Moving_WordPress#When_Your_Domain_Name_or_URLs_Change) ``` Can you provide any further assistance? Thanks, Janet ```
-
A 503 error is a service unavailable error. I have seen situations where redirects are incorrect and loop. Depending on the hosting setup, this can trigger various HTTP error codes.
The best way to debug this is by looking at your Apache access logs. Scan your logs for the 503 errors. Pay attention to the URL being requested as well as the referring URL.
Very likely, there's some looping process and in cases when Apache runs on FastCGI, you can get a 503 error due to too many processes being triggered.
Also, due to how WP handles 404's, I've seen many plugins mask underlying causes. So if you have any plugins that impact error handling, you may need to remove those while debugging.
You can also use http://www.redbot.org/ to check the headers for any page that should be redirected. That tool should return a Location header with a URL. Visit that Location URL in your browser and make sure it resolves.
The goal here is to try to replicate the behavior. Once you can replicate the behavior, dig into your redirect/rewrite rules and examine the logic to determine why you are seeing the loops or failures.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Another company's website indexing for my site
Hi, I am looking at all the pages which Google are indexing for my website and have come across pages of another company's website. I have contacted them through their online form and Facebook page asking for them to remove their listings for us, but to no avail so far. Is there a way I can do this myself?
Technical SEO | | British-Car-Registrations0 -
Generation 'child' sitemaps?
First off, am I correct in thinking that a 'child' sitemap is a sitemap of a subfolder and everything that sits under it, i.e. www.example.com/example If so, can someone give me a good recommendation for generation a free child sitemap please? Many thanks, Rhys
Technical SEO | | SwanseaMedicine0 -
Sitemap
Hi, I have generated a dynamic sitemap and submit it in search console, but there is a huge gap between the number of submitted pages and the number of indexed pages. 143,206 URLs submitted 2,151 URLs indexedwhy we have this gap and what should I do to reduce it?
Technical SEO | | Digikala0 -
Consistent top 10 in G image search - but a different 'stolen' version every time!
I have a photo that was uploaded back in 2005. It is an aerial shot and has received a fair bit of traffic over the years. I'm pretty sure it was ranked #1 in Google Images for the town name for a while. Now, however, it never ranks. Well actually it does. But every single time it is a version on a different website that is being used without permission.
Technical SEO | | Cornwall
And I'm not talking about one website. Every time I fill out a DMCA and have the image removed only to see a completely different website featuring in the top 10. This has happened 5 times so far and I'm just about to fill out another DMCA request. What is going on? Surely Google in its infinite wisdom is smart enough to check the timestamp or date cues on page to figure out which is the original. These other sites are often complete unknowns compared to my site which is a 12yr old authority site on the subject.
Don't get it!0 -
MSNbot Issues
We found msnbot is doing lots of request at same time to one URL, even considering we have caching, it triggers many requests at same time so caching does not help at the moment: For sure we can use mutex to make sure URL waits for cache to generate, but we are looking for solution for MSN boot. 123.253.27.53 [11/Dec/2012:14:15:10 -0600] "GET //Fun-Stuff HTTP/1.1" 200 0 "-" "msnbot/2.0b (+http://search.msn.com/msnbot.htm)" 1.253.27.53 [11/Dec/2012:14:15:10 -0600] "GET //Type-of-Resource/Fun-Stuff HTTP/1.1" 200 0 "-" "msnbot/2.0b (+http://search.msn.com/msnbot.htm)" 1.253.27.53 [11/Dec/2012:14:15:10 -0600] "GET /Browse//Fun-Stuff HTTP/1.1" 200 6708 "-" "msnbot/2.0b (+http://search.msn.com/msnbot.htm)" We found the following solution: http://www.bing.com/community/site_blogs/b/webmaster/archive/2009/08/10/crawl-delay-and-the-bing-crawler-msnbot.aspx Bing offers webmasters the ability to slow down the crawl rate to accommodate web server load issues. User-Agent: * Crawl-Delay: 10 Need to know if it’s safe to apply that. OR any other advices. PS: MSNBot gets so bad at times that it could trigger a DOS attack – alone! (http://www.semwisdom.com/blog/msnbot-stupid-plain-evil#axzz2EqmJM3er).
Technical SEO | | tpt.com0 -
Www & non-www URL Issue
Hi is there any tools out there which can enable me to find pages which both serve www and non-www versions of each URL
Technical SEO | | monster990 -
Ignore url parameters without the 'parameter=' ?
We are working on an ecommerce site that sorts out the products by color and size but doesn't use the sortby= but uses sortby/. Can we tell Google to ignore the sortby/ parameter in Webmaster Tools even though it is not followed by an = sign? For example: www.mysite.com/shirts/tshirts/shopby/size-m www.mysite.com/shirts/tshirts/shopby/color-black Can we tell WMT to ignore the 'shopby/' parameter so that only the tshirts page will be indexed? Or does the shopby have to be set up as 'shopby=' ? Thanks!
Technical SEO | | Hakkasan0 -
Should I have a 'more' button for links?
I have a website that has a page for each town. rather than listing all the towns with a link to each, I want to show only the most popular towns and have a 'more' button that shows all of them when you click it. I know that the search engine can always see the full list of links and even though the visitor can't this doesn't go against Google guidelines because there is no deception involved, the more button is quite clear. However, my colleague is concerned that this is 'making life hard' for the search engines and so the pages are less likely to be indexed. I disagree. Is he right to worry about this??
Technical SEO | | mascotmike0