I need help on web page load time, its very bad!
-
Note: This is KILLING my customer experience.
Here is my webpage: http://www.stbands.com
Here is a speed test that may help you (look at the poor ratings in the upper corner)
http://www.webpagetest.org/result/110628_MW_Y8CQ/1/details/
I have an F on "Cache Static Content" - anyone know how I can fix this?
Also, it is a e-commerce website hosted through core commmerce. I have some access to code but not all of it. Some of it is dynamic. However, if you tell me specific things I can forward it to their very awesome tech department. They are very willing to work with me and are now considering implementing a CDN after I schooled them.
Any help is greatly appreciated. Don't be afraid to get very technical - I may not understand it, but the engineers there will.
-
John - Thanks, i'll start here. I'm not sure why they are set up like this (facepalm)
-
John - Thanks, i'll start here. I'm not sure why they are set up like this (facepalm)
-
You might also want to try to Google speed plugin, http://code.google.com/speed/page-speed/, it identify issues and will give fix suggestions that you can pass to your tech guys/gals.
-
My load time went down by 50%. Non eCommerce though. I would say yes. However I would definitely look into all the variables with eCommerce. Should be great though.
-
My site uses SSL, which means I have to pay $20 / month. I need to make sure it is going to be worth it before I do that. Is it?
-
Sure, there are two areas to address on your server config. One is the cache-control which is returned in the header. I set a longer period of cache for all images and scripts to save users downloading new copies time and again to their cache. ie: max-age=3600, must-revalidate - see more here: http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html
Then I would also set some rules around using 304 status on page furniture and other assets which do not change frequently.
Aside from this, as far as I am aware, you should ensure your stack is optimised. The recent Search Insight session from Google was interesting as in that presentation they talked a lot about the average load and latency times they see, useful to use as a benchmark in tuning your own speed.
Cheers,
Damien
-
A lot of items on your home page are getting served securely... which isn't necessary, and will prevent the browser from caching them properly. For example:
- https://www.stbands.com/javascript/jquery/jquery.min.js
- https://www.stbands.com/css/dynamic-css.php?currentlyActivePageId=1
- https://www.stbands.com/uploads/image/Custom Wristbands(1).jpg
- https://www.stbands.com/images/categories/783.jpg
- https://www.stbands.com/images/categories/785.jpg
- https://www.stbands.com/images/categories/786.jpg
- https://www.stbands.com/images/categories/787.jpg
- https://www.stbands.com/images/categories/788.jpg
Since it's not a secure page, I wouldn't be serving all of these securely. I'd use http:// instead of https://
-
I am a huge fan of Cloudflare!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Why does Google's search results display my home page instead of my target page?
Why does Google's search results display my home page instead of my target page?
Technical SEO | | h.hedayati6712365410 -
Search Console Indexed Page Count vs Site:Search Operator page count
We launched a new site and Google Search Console is showing 39 pages have been indexed. When I perform a Site:myurl.com search I see over 100 pages that appear to be indexed. Which is correct and why is there a discrepancy? Also, Search Console Page Index count started at 39 pages on 5/21 and has not increased even though we have hundreds of pages to index. But I do see more results each week from Site:psglearning.com My site is https://wwww.psglearning.com
Technical SEO | | pdowling0 -
Need Third Party Input. Our Web host blocked all bots including Google and myself because they believe SEO is slowing down their server.
I would like some third party input... partly for my sanity and also for my client. I have a client who runs a large online bookstore. The bookstore runs in Magento and the developers are also apparently the web host. (They actually run the servers.. I do not know if they are sitting under someones desk or are actually in a data center) Their server has been slowed down by local and foreign bots. They are under the impression my SEO services are sending spammer bots to crawl and slow down their site. To fix the problem they disallowed all bots. Everything, Google, Yahoo, Bing. They also banned my access from the site. My clients organic traffic instantly took a HUGE hit. (almost 50% of their traffic is organic and over 50% is Organic + Adwords most everything from Google) Their keyword rankings are taking a quick dive as well. Could someone please verify the following as true to help me illustrate to my client that this is completely unacceptable behavior on part of the host. I believe: 1.) You should never disavow ALL robots from your site as a solution for spam. As a matter of fact most of the bad bots ignore robots.txt anyways. It is a way to limit where Google searches (which is obviously a technique to be used) 2.) On site SEO work as well as link building, etc. is not responsible for foreign bots and scrappers putting a heavy load on the server. 3.) Their behavior will ultimately lead to a massive loss of rankings (already happening) and a huge loss of traffic (already happening) and ultimately since almost half the traffic is organic the client could expect to lose a large sum of revenue from purchases made by organic traffic since it will disappear. Please give your input and thoughts. I really appreciate it!
Technical SEO | | JoshuaLindley1 -
Deleting 30,000 pages all at once - good idea or bad idea?
We have 30,000 pages that we want to get rid of. Each product within our database has it's own page. And these particular 30,000 products are not relevant anymore. They have very little content on them and are basically the same exact page but with a few title changes. We no longer want them weighing down our database so we are going to delete them. My question is - should we get rid of them in smaller batches like 2,000 pages at a time, or is it better to get rid of all them in one fell swoop? Which is least likely to raise a flag to Google? Anyone have any experience with this?
Technical SEO | | Viewpoints0 -
Will updating part of my site help a static web page
Hi, what i am trying to find out is, i have a page on my site http://www.clairehegarty.co.uk/virtual-gastric-band-with-hypnotherapy and i would like to know, once i have got the page to the way i want it, the page will not change, so i would like to know if i update my site and add pages and articles, will the updates help this page with google rankings, or do i have to keep updating this page if i want it to rank high with google. i have seen pages that have never changed but they continue to rank high with google and i would like to know their secret
Technical SEO | | ClaireH-1848860 -
Determining When to Break a Page Into Multiple Pages?
Suppose you have a page on your site that is a couple thousand words long. How would you determine when to split the page into two and are there any SEO advantages to doing this like being more focused on a specific topic. I noticed the Beginner's Guide to SEO is split into several pages, although it would concentrate the link juice if it was all on one page. Suppose you have a lot of comments. Is it better to move comments to a second page at a certain point? Sometimes the comments are not super focused on the topic of the page compared to the main text.
Technical SEO | | ProjectLabs1 -
SEOMoz is indicating I have 40 pages with duplicate content, yet it doesn't list the URL's of the pages???
When I look at the Errors and Warnings on my Campaign Overview, I have a lot of "duplicate content" errors. When I view the errors/warnings SEOMoz indicates the number of pages with duplicate content, yet when I go to view them the subsequent page says no pages were found... Any ideas are greatly welcomed! Thanks Marty K.
Technical SEO | | MartinKlausmeier0 -
So I Get An Interior Page To Rank, But Lower. How does that help?
Hi, So, I have a term that would bounce around #3 to 5 for. I make a page some months ago that is solely targeted to that term. And, voila! Google sees the new page as the best result instead of the home page and the new page ranks, but at #9 or #10. Of course the homepage is a stronger page, but the new page is better targeted to that term. Using the handy SEOMoz toolbar, the homepage has a page authority of 59 and the newer interior page has 32. Both are equally functional to my purpose. Part of me just wants to 301 in the interior page to the homepage and forget about it. It would take forever to get that interior page up to a similar page authority through the magic of links. What do you think I should do. I feel like I'm channeling Wile E. Coyote. Thanks! Best...Mike
Technical SEO | | 945010