How can I Improve Loading Speed? - Parker Dubrule Lawyers
-
Parker Dubrule Lawyers' website at parkerdubrulelawyers.com seemed to be loading quite slow this morning (>5 Seconds). I added a lazyload plugin, minified JS and CSS, and ensured that the images were optimized---all of this seemed to help and brough it down to under 2 seconds. We are looking at more reliable hosting options for our clients---ones that are inherently faster possibly without these plugins being added to the mix. Does anyone have insight on a safe, secure, and fast hosting/server option to enhance the experience from the get go? All of the websites that we build are in Worpress.
Your help is much appreciated! Thanks!
-
Thank you for such a detailed response!
-
Thank you for your insight!
-
Hosting & DNS
It looks like the DNS response is all over the place. Sometimes it's acceptable... at ~100ms, other times not so much. A better DNS provider would be worth looking into. Amazon Route 53 or Dyn are pretty good options.
For shared hosting, I can second SiteGround. It's a solid host for lower budgets. DigitalOcean is a very solid and inexpensive VPS, but there will be less hand holding. I plan on migrating to DO in the next week or so. My current host just removed sudo privileges from their VPS accounts. I know, right!?!
Sweet, sweet PHP 7 and Redis - here I come.
Things You Can Fix Immediately
Run the site through Pagespeed Insights. Make a punch list and go from there. There's also a download link for 'optimized' resources. Usually I only take images from that. More on that later.
One of the big ones is 'Remove Render Blocking JavaScript. The quick fix is moving Web Font Loader script. and the GA script to the footer. You're halfway there, in a lot of instances.
Images
A couple of slider images are still over 200KB. If there's anything you can do to reduce that, do so. The Pagespeed Insights tool states that the home page has a couple images that could be compressed further. Even though the savings are minuscule, it adds up over requests.Fonts
One of the better performance increases can be had with fonts. Again, move your Web Font Loader script to the footer. Consider using fewer character sets. Do you really need greek-ext, cyrillic or vietnamese character sets? If not, remove them.
Another fun one is using the preconnect tag. Here's a practical guide to web font performance from an author at CSS-Tricks. Just make sure to use fonts.gstatic, instead of fonts.typonine in the code snippet. Here's a fairly detailed reason why you want to use preconnect, from Ilya Grigorik. (Seriously, follow that guy if you're not already doing so.)
CSS, HTML & JavaScript
The site appears to have unminified CSS inlined in the head of the document. That's a lot of CSS, and it probably isn't all critical path. In order to render a page as fast as possible, you need to display visible content first. Drop your various style sheets into the critical path CSS generator.
You can input your newly generated critical path CSS into Autoptimize. It's a very handy plugin that minifies HTML, CSS and JavaScript. It will also combine your CSS & JavaScript to reduce requests. Another handy feature includes, you guessed it, Inline Critical Path CSS.
You will likely have to remove BWP Minify, as Autoptimize handles most - if not all - of those functions. More is not better, in this instance. In fact, you should disable any caching plugin options which handle minification.
Gzip Compression & Cache Expiration
It looks like cache expiration settings aren't setup for some basic MIME types (CSS, JPEG, etc.) Consider setting up a caching plugin, such as Super Cache or Total Cache. Failing that, this is one of the better htaccess settings repos.
***Edit: One of the issues involves query strings in static resources. Here's a good resource, with a few options to handle that.
As always make these changes in a test environment. And best of luck. You'll probably be happier, with a lot of projects.
-
I use iClickAndHost, SiteGround as shared hosting. I also use Amazon EC2, Linode, DigitalOcean, Vultr as VPS. For CDN - Amazon CloudFront and S3. Everything works perfect.
But you should diagnose your hosting issues before considering switching them. Can be something temporary - DDoS, hardware failure, network overload, etc.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How Often Can I Change My Meta Titles? (Product Discounts)
Hello, I have an ecommerce website. Due to discounts and sales on most of my products, the discount rates change very often. Some products have %10 discounts for 3 days and then same products might have %50 discounts for another 5 days. I would like to show these product specific discounts on Meta Titles and Descriptions dynamically. My system will automatically update the Meta Titles once the discount on a product changes. My question is, how often can I change the Meta Titles? Is changing them (only discount rates = 1word) too often bad for my SEO? What is Goolge's approach on this? Thank you in advance for all your help. Best,
Technical SEO | | yigitgok
Yigit0 -
Can Anybody Understand This ?
Hey guyz,
Technical SEO | | atakala
These days I'm reading the paperwork from sergey brin and larry which is the first paper of Google.
And I dont get the Ranking part which is: "Google maintains much more information about web documents than typical search engines. Every hitlist includes position, font, and capitalization information. Additionally, we factor in hits from anchor text and the PageRank of the document. Combining all of this information into a rank is difficult. We designed our ranking function so that no particular factor can have too much influence. First, consider the simplest case -- a single word query. In order to rank a document with a single word query, Google looks at that document's hit list for that word. Google considers each hit to be one of several different types (title, anchor, URL, plain text large font, plain text small font, ...), each of which has its own type-weight. The type-weights make up a vector indexed by type. Google counts the number of hits of each type in the hit list. Then every count is converted into a count-weight. Count-weights increase linearly with counts at first but quickly taper off so that more than a certain count will not help. We take the dot product of the vector of count-weights with the vector of type-weights to compute an IR score for the document. Finally, the IR score is combined with PageRank to give a final rank to the document. For a multi-word search, the situation is more complicated. Now multiple hit lists must be scanned through at once so that hits occurring close together in a document are weighted higher than hits occurring far apart. The hits from the multiple hit lists are matched up so that nearby hits are matched together. For every matched set of hits, a proximity is computed. The proximity is based on how far apart the hits are in the document (or anchor) but is classified into 10 different value "bins" ranging from a phrase match to "not even close". Counts are computed not only for every type of hit but for every type and proximity. Every type and proximity pair has a type-prox-weight. The counts are converted into count-weights and we take the dot product of the count-weights and the type-prox-weights to compute an IR score. All of these numbers and matrices can all be displayed with the search results using a special debug mode. These displays have been very helpful in developing the ranking system. "0 -
Can hreflang replace canonicalisation ?
Hi Im working with a site that has ALOT of duplicate content and have recommended developer fix via correct use of Canonicalisation i.e the canonical tag. However a US version (of this UK site) is about to be developed on a subfolder (domain.com/uk/ & domain.com/US/ etc so also looking into adopting the hreflang attribute on these. Upon reading up about the hreflang attribute i see that it performs a degree of canonicalisation too. Does that mean that developing the international versions with hreflang means there's no need to apply canonicalistion tags to deal with the dupe content, since will deal with the original dupe content problems as well as the new country related dupe content, via the hreflang ? I also understand that hreflang and canonicalisation can conflict/clash on different language versions of international subfolders etc as per: http://www.youtube.com/watch?v=Igbrm1z_7Hk In this instance we are only looking at US/UK versions but very likely will want to expand into non english countries too in the future like France for example. So given both the above points if you are using hreflang is it advisable (or even best) to totally avoid the canonical tag ? I would be surprised if the answers yes, since whilst makes logical sense given the above (if the above statements are correct), that seems strange given how important and standard best practice canonical usage seems to be these days. What best ? Use the Hreflang alone, or the Canonical tag alone or both ? What does everyone else do in similar situation ? All Best Dan
Technical SEO | | Dan-Lawrence0 -
Oh no googlebot can not access my robots.txt file
I just receive a n error message from google webmaster Wonder it was something to do with Yoast plugin. Could somebody help me with troubleshooting this? Here's original message Over the last 24 hours, Googlebot encountered 189 errors while attempting to access your robots.txt. To ensure that we didn't crawl any pages listed in that file, we postponed our crawl. Your site's overall robots.txt error rate is 100.0%. Recommended action If the site error rate is 100%: Using a web browser, attempt to access http://www.soobumimphotography.com//robots.txt. If you are able to access it from your browser, then your site may be configured to deny access to googlebot. Check the configuration of your firewall and site to ensure that you are not denying access to googlebot. If your robots.txt is a static page, verify that your web service has proper permissions to access the file. If your robots.txt is dynamically generated, verify that the scripts that generate the robots.txt are properly configured and have permission to run. Check the logs for your website to see if your scripts are failing, and if so attempt to diagnose the cause of the failure. If the site error rate is less than 100%: Using Webmaster Tools, find a day with a high error rate and examine the logs for your web server for that day. Look for errors accessing robots.txt in the logs for that day and fix the causes of those errors. The most likely explanation is that your site is overloaded. Contact your hosting provider and discuss reconfiguring your web server or adding more resources to your website. After you think you've fixed the problem, use Fetch as Google to fetch http://www.soobumimphotography.com//robots.txt to verify that Googlebot can properly access your site.
Technical SEO | | BistosAmerica0 -
How can I have pages with media that changes and avoid duplicate content when the text stays the same?
I want to have a page that describes a specific property and/or product. The top part of the page has media options such as video and photos while the bottom includes the description. I know I can set up the media in tabs and have it separated by javascript, but everything resides on one page so there are no duplicate content issues. Example: http://www.worldclassproperties.com/properties/Woodside BUT what if I need to the photos and the videos to have separate URLs so I can link to them individually? For example, for a real estate site blog, I may want to send visitors to the page of the home tour. I don't want to link them to the version of the page with the photos because I want them to arrive on the video portion. Example: http://www.worldclassproperties.com/properties/Woodside?video=1 Is there any way to get around the problem that would result from the duplicate content of the product/property description? I do not have the resources in the budget to make two unique descriptions for every page.
Technical SEO | | WebsightDesign0 -
Destination URL in SERPs keeps changing and I can't work out why.. Help.
I am befuddled as to why our destination URL in SERPs keeps changing oak furniture was nicely returning http://www.thefurnituremarket.co.uk/oakfurniture.asp then I changed something yesterday I did 2 things. published a link to that on facebook as part of a competition. redirected dynamic pages to the static URL for oak furniture.. Now for oak furniture the SERPs in GG UK is returning our home page as the most relevant landing page.. Any Idea why? I'm leaning to an onpage issue than posting on FB.. Thoughts?
Technical SEO | | robertrRSwalters0 -
Browser Pop Ups - Can it be SEO Friendly and how?
Hi The designer of the company I work for is re-designing Pop Up browsers as well as inline Pop up and Drop down menus. He needs SEO requirements - how can they be SEO friendly? Thanks a lot for your help! SL. Please see below the detail: Browser Pop Ups all include:
Technical SEO | | charsimona
• a browser title,
• a logo and title in the title bar,
• a close window button and
• a call to action (that closes pop-up when clicked). Usage:
Use when you'd like to offer additional information to the
user but, not take the user away from the main page. Inline Pop up and Drop down menus. The inline pop-up & drop down is used to display additional menu options, functionality
or content on the page without dedicating real estate in the page layout. It's a part of the page HTML to retain SEO value and thus does not trigger pop-up blockers. A title bar displays when content of the pop-up or drop down is not in context of
the trigger. When used as a drop down, it is attached to the the bottom of it's trigger and left-aligned (unless it would exceed beyond the browser chrome, then it's right-aligned). When used as a pop-up, it is centered vertically/horizontally in the users browser window.
The inline pop-up/drop down can be triggered differently per instance (e.g. onclick, onhover with delay). It can be closed by: clicking on link/location that triggered the pop-up/drop down (a.k.a. close icon) clicking anywhere outside the pop-up/drop down There are 5 widths to choose from, based on the needs of the content: 196px (3 columns) 266px (4 columns) 406px (6 columns) 546px (8 columns) 658px (10 columns)0 -
What has to be changed to improve rank?
What has to be changed to improve rank? We had "hip hop jewelry" keyword for a while, position 4. All of the sudden it dropped for position 6 and never went back. We did some on page optimization and got couple of links here and there... but so far we still at position 6. Please suggest us what has to be done?
Technical SEO | | DiamondJewelryEmpire0