How can I Improve Loading Speed? - Parker Dubrule Lawyers
-
Parker Dubrule Lawyers' website at parkerdubrulelawyers.com seemed to be loading quite slow this morning (>5 Seconds). I added a lazyload plugin, minified JS and CSS, and ensured that the images were optimized---all of this seemed to help and brough it down to under 2 seconds. We are looking at more reliable hosting options for our clients---ones that are inherently faster possibly without these plugins being added to the mix. Does anyone have insight on a safe, secure, and fast hosting/server option to enhance the experience from the get go? All of the websites that we build are in Worpress.
Your help is much appreciated! Thanks!
-
Thank you for such a detailed response!
-
Thank you for your insight!
-
Hosting & DNS
It looks like the DNS response is all over the place. Sometimes it's acceptable... at ~100ms, other times not so much. A better DNS provider would be worth looking into. Amazon Route 53 or Dyn are pretty good options.
For shared hosting, I can second SiteGround. It's a solid host for lower budgets. DigitalOcean is a very solid and inexpensive VPS, but there will be less hand holding. I plan on migrating to DO in the next week or so. My current host just removed sudo privileges from their VPS accounts. I know, right!?!
Sweet, sweet PHP 7 and Redis - here I come.
Things You Can Fix Immediately
Run the site through Pagespeed Insights. Make a punch list and go from there. There's also a download link for 'optimized' resources. Usually I only take images from that. More on that later.
One of the big ones is 'Remove Render Blocking JavaScript. The quick fix is moving Web Font Loader script. and the GA script to the footer. You're halfway there, in a lot of instances.
Images
A couple of slider images are still over 200KB. If there's anything you can do to reduce that, do so. The Pagespeed Insights tool states that the home page has a couple images that could be compressed further. Even though the savings are minuscule, it adds up over requests.Fonts
One of the better performance increases can be had with fonts. Again, move your Web Font Loader script to the footer. Consider using fewer character sets. Do you really need greek-ext, cyrillic or vietnamese character sets? If not, remove them.
Another fun one is using the preconnect tag. Here's a practical guide to web font performance from an author at CSS-Tricks. Just make sure to use fonts.gstatic, instead of fonts.typonine in the code snippet. Here's a fairly detailed reason why you want to use preconnect, from Ilya Grigorik. (Seriously, follow that guy if you're not already doing so.)
CSS, HTML & JavaScript
The site appears to have unminified CSS inlined in the head of the document. That's a lot of CSS, and it probably isn't all critical path. In order to render a page as fast as possible, you need to display visible content first. Drop your various style sheets into the critical path CSS generator.
You can input your newly generated critical path CSS into Autoptimize. It's a very handy plugin that minifies HTML, CSS and JavaScript. It will also combine your CSS & JavaScript to reduce requests. Another handy feature includes, you guessed it, Inline Critical Path CSS.
You will likely have to remove BWP Minify, as Autoptimize handles most - if not all - of those functions. More is not better, in this instance. In fact, you should disable any caching plugin options which handle minification.
Gzip Compression & Cache Expiration
It looks like cache expiration settings aren't setup for some basic MIME types (CSS, JPEG, etc.) Consider setting up a caching plugin, such as Super Cache or Total Cache. Failing that, this is one of the better htaccess settings repos.
***Edit: One of the issues involves query strings in static resources. Here's a good resource, with a few options to handle that.
As always make these changes in a test environment. And best of luck. You'll probably be happier, with a lot of projects.
-
I use iClickAndHost, SiteGround as shared hosting. I also use Amazon EC2, Linode, DigitalOcean, Vultr as VPS. For CDN - Amazon CloudFront and S3. Everything works perfect.
But you should diagnose your hosting issues before considering switching them. Can be something temporary - DDoS, hardware failure, network overload, etc.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Do you think have to re-submit my site to search engines after I made improvements?
Some time ago I started to do SEO for a one-page website and didn't get any positive result: no traffic, no filled in online booking form (yet another, multiple page website offering the same service yielded in multiple filled-in "schedule an appointment" forms). I found out my one-page website was considered to be "keyword-spamming" and converted it to a multiple page one. Its domain authority went up, but it doesn't still bring any traffic. I am thinking maybe I have to let the search engines know that it has been updated so they stop penalizing it? Do you think it might help and if yes, what exactly I should do? Will be thankful very much for any suggestion!
Technical SEO | | kirupa0 -
Can Googlebot crawl the content on this page?
Hi all, I've read the posts in Google about Ajax and javascript (https://support.google.com/webmasters/answer/174992?hl=en) and also this post: http://moz.com/ugc/can-google-really-access-content-in-javascript-really. I am trying to evaluate if the content on this page, http://www.vwarcher.com/CustomerReviews, is crawlable by Googlebot? It appears not to be. I perused the sitemap and don't see any ugly Ajax URLs included as Google suggests doing. Also, the page is definitely indexed, but appears the content is only indexed via its original source (Yahoo!, Citysearch, Google+, etc.). I understand why they are using this dynamic content, because it looks nice to an end-user and requires little to no maintenance. But, is it providing them any SEO benefit? It appears to me that it would be far better to take these reviews and simply build them into HTML. Thoughts?
Technical SEO | | danatanseo0 -
Can spiders crawl jQuery Fancy Box scripts
Hi Everyone - I'm not a technical person at all. I have some content that will be hidden until a user clicks "learn more" where upon it will be displayed via jQuery Fancy Box script. The content behind the learn more javascript is important and I need it to be crawled by search engine spiders. Does anyone know if there will be a problem with this script?
Technical SEO | | Santaur0 -
Can hreflang replace canonicalisation ?
Hi Im working with a site that has ALOT of duplicate content and have recommended developer fix via correct use of Canonicalisation i.e the canonical tag. However a US version (of this UK site) is about to be developed on a subfolder (domain.com/uk/ & domain.com/US/ etc so also looking into adopting the hreflang attribute on these. Upon reading up about the hreflang attribute i see that it performs a degree of canonicalisation too. Does that mean that developing the international versions with hreflang means there's no need to apply canonicalistion tags to deal with the dupe content, since will deal with the original dupe content problems as well as the new country related dupe content, via the hreflang ? I also understand that hreflang and canonicalisation can conflict/clash on different language versions of international subfolders etc as per: http://www.youtube.com/watch?v=Igbrm1z_7Hk In this instance we are only looking at US/UK versions but very likely will want to expand into non english countries too in the future like France for example. So given both the above points if you are using hreflang is it advisable (or even best) to totally avoid the canonical tag ? I would be surprised if the answers yes, since whilst makes logical sense given the above (if the above statements are correct), that seems strange given how important and standard best practice canonical usage seems to be these days. What best ? Use the Hreflang alone, or the Canonical tag alone or both ? What does everyone else do in similar situation ? All Best Dan
Technical SEO | | Dan-Lawrence0 -
How does Progressive Loading, aka what Facebook does, impact proper search indexation?
My client is planning on integrating progressive loading into their main product level pages (those pages most important to conversions and revenue). I am not skilled on "progressive laoding" but was told this is what Facebook does. Currently, the site's pages are tabbed and use Ajax. Is there any negative impact by changing this up by including progressive loading? If anyone can help me understand what this is and how it might impact a site from an SEO perspective, please let me know. thanks a ton!! Janet
Technical SEO | | ACNINTERACTIVE1 -
Should i include this group of text to improve ranking
Hi i am worried about my site www.in2town.co.uk because it it a magazine site i am finding it hard to find space to put text to describe the site for the search engines. I have put text near the top of the site which is under the logo and then i have put more text under the first set of headline where i have put a twitter sign and text but it does not look that great. I rank five in google for the word lifestyle magazine and feel i could do much better and i need to rank for other words but because my site is a magazine site, i am struggling. can anyone offer me advice and let me know if the text looks rubbish please.
Technical SEO | | ClaireH-1848860 -
How do i can change my instant preview image?
Since we've put an index.htm file indicating that our webiste is under reconstruction and have removed it few hours after, the instant preview picture still showing the informative message. It has been 40 day since and we've submit our sitemap few times. Is there a way to change it manually?
Technical SEO | | webit400 -
Can someone break down 'page level link metrics' for me?
Sorry for the, again, basic question - can someone define page level link metrics for me?
Technical SEO | | Benj250