Bounce Back or Bounce Through
-
Bounce rate is defined as 'single page visits to a site divided by total visits to the site' as I understand it. It could be argued that a well designed site might vector people on to other sites effectively (I generally use Wikipedia this way for instance). On the other hand a site that bounces people back to where they came from may be genuinely poor. So the questions:
Is the bounce rate really calculated in the stated way by Google?
Is it used, as far as we know, as a metric for the search engine?
What should we do to mitigate the effects of this poor metric?!
thanks,
Mike
-
Actually, bounce rate would be of a concern to search engines, at least for visits that originate from the search engine. The SEs want the users to have a good experience, and if a user clicks on a result and then comes right back to the results page, the SEs may feel that the user did not have a good experience with that result and maybe a different result for that query should be shown.
-
Thanks, yes, it looks from this as if the experts think that Google is doing what we would hope they do and not take account of bounce through. Although of course there may be good reasons for a site not wanting bounce through either (as EGOL notes), it shouldn't be a concern for the search engines
-
As far as I'm aware, Google will use your 'bounce back' rate (where by users return to the search results page straight away) as a search metric as this could indicate whether the site is relevant for that specific search query. This was mentioned in the 2011 SEO Ranking Factors Report.
Hope that helps
-
If search engines are using this data they are certainly only using it for sites competing for the same or similar keywords.
A high bounce rate can be bad or it can be "normal". It would be bad if your site is offensive (and people run away), it can be bad if your site has irrelevant content for the query, it can be bad if your site has thin content, you can probably think of more.
It can be normal if you have a dictionary site and the searcher finds the word, gets the definition and leaves happily.
THE IMPORTANT THING TO DO..... I believe that everyone should be working to reduce their bounce rate and any webmaster should be able to find improvements.
The best way to do it is to have relevant links, obviously placed on every page. For example in the dictionary site your goal should be have linked words within the definition, links to related words adjacent to the definition and links to a few enticing articles along the side.
On an article site you can links within the text to related articles, a "recommended" box of links beside the article and even a few enticing links to "popular" or "related" articles where every one will see them.
Try to reduce your bounce rate by improving your site and making your relevant content visible on every page.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Back link plan discussion
When you have a lot of keywords that you rank for say something like 15,000 or more. How do you develop a good back link plan? I was thinking to first look at the highest volume keywords we already rank for but aren't in the top 1-3 spots. To focus on those few words trying to obtain more high quality back links. But I'm not sure if this is the best plan . What would you do? What are some good consistent back link plans you can use to work on a keyword or lots of keywords? Thanks for the discussion, Chris
Algorithm Updates | | Cfarcher1 -
Do the back-links go wasted when anchor text or context content doesn't match with page content?
Hi Community, I have seen number of back-links where the content in that link is not matching with page content. Like page A linking to page B, but content is not really relevant beside brand name. Like page with "vertigo tiles" linked to page about "vertigo paints" where "vertigo" is brand name. Will these kind of back-links completely get wasted? I have also found some broken links which I'm planning to redirect to existing pages just to reclaim the back-links even though the content relevancy is not much beside brand name. Are these back-links are beneficial or not? Thanks
Algorithm Updates | | vtmoz0 -
Do back-links to non indexed sub-domains / sub-directories considered by Google as website backlinks and pass Pagerank to website?
Hi, If some noindexed links on our website or sub-domain got some backlinks, will that backlinks pass Pagerank / linkjuice to website? Will they be considered as backlinks to website by Google? Here is a statement from Matt cutts for the question. My question is same as below with answer? Eric Enge: Can a NoIndex page accumulate PageRank? Matt Cutts: A NoIndex page can accumulate PageRank, because the links are still followed outwards from a NoIndex page. Thanks
Algorithm Updates | | vtmoz0 -
Domain location is a ranking factor? Back links & website?
If a website trying to rank in US and it has received many back-links from domains hosting from other countries; how it will impact website ranking? Can a website hosted in country will rank well in other country? How much the hosted location matters? Like....domain hosted in Germany but trying to rank in US?
Algorithm Updates | | vtmoz0 -
Moving to https and back to http, would it it hurt?
We have redirected everything on our blog from http to https. Our blog is in a subfolder so that now it looks like this: https://ourdomain/blog; But everything else i.e. our shop continues to be on http at http://ourdomain We are wondering: 1- Does the domain authority for SEO purposes have different values for the http and the https version of a domain? 2- If yes, is there a way to check the difference in authority between the http and the https version? 3- If we do have a higher authority on our http version (as historically we have been mostly on our http), would it make sense to go back to the http for the blog to enjoy that authority too? 4- Would changing our mind and going back to http after a few months of just having moved to https from http send any negative signals to Google? Would Google care if we do a back and forth essentially? Many thanks!
Algorithm Updates | | TVape0 -
Does Google Analytics Adjusted Bounce Rate Lead to Increase in Average Time per Visitor?
Hello, I just recently implemented adjusted bounce rate onto one of the websites that I track via google analytics. (http://searchenginewatch.com/article/2322974/How-to-Implement-Adjusted-Bounce-Rate-ABR-via-Google-Tag-Manager-Tutorial) Since doing so, obviously my bounce rate has gone down significantly, nearly half of what it use to be, but I've also noticed an increase in the average time per visitor. In fact, the increase of average time per visitor began the same day I adjusted the bounce rate. Has this happened to anyone else? Can someone please explain why/how this may occur?
Algorithm Updates | | WebServiceConsulting.com0 -
Bounce Rate a factor?
Hi Guys, I've been SEOing two sites for just over 6 months, making very little headway in the process. All the tactics I have employed work on the many other sites that we do SEO work for. I have written pages of rich, useful content, metas, built links from various sources, re-built both sites on a better platform, alt tags are optimised etc etc. This seems to have brought me some success but nothing to write home about. The only major difference I can see between these two sites and my others is bounce rate. The SEO pages I have produced have a fairly healthy bounce rate (20-30%) which isn't a problem, but, both sites have blogs which draw (a lot of) visitors from various social networks and the bounce rate is through the roof (80-90%) obviously increasing the average for the whole site. Now, I know there has been various discussions around this with no real outcome but I cannot see what else it can be. Am I missing something? I should add that both sites are in competitive sectors but not that competitive to stop me seeing at least top 100 results....
Algorithm Updates | | SEOBirmingham810