Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Does the use of sliders for text-on-page, effects SEO in any way?
-
The concept of using text sliders in an e-commerce site as a solution to placing SEO text above or in between product and high on ages, seems too good to be true.... or is it?
How would a text slider for FAQ or other on-page text done with sliding paragraphs (similar but not this specific code- http://demo.tutorialzine.com/2010/08/dynamic-faq-jquery-yql-google-docs/faq.html) might effect text-on-page SEO. Does Google consider it hidden text?
Would there be any other concerns or best practices with this design concept?
-
Fredrik,
This is very helpful and gives me a clearer understanding as to how to make this work properly. The example was just that, and meant to explain basic functionality. We'll make sure we end up using an index-able HTML based version.
Much thanks for your advise.
ron
-
Hi Ron
As Paul stated there are many ways of doing sliders. Most of the new sliders out there do work with JavaScript but often used already loaded dom elements for the slides. That means that the actual content is in the HTML and the JavaScript is used to animate or style them. This content would then be indexed just as a normal div would.
You can also use http://www.seobrowser.com/, (simple option is free) to see the page as Google would see it. If you then can read your content it should be possible to index it.
One thing to think of is that sliders, as the name implies, often contains more than one slide. If the slider has a heading in it it might be a good thing to make the first heading H1 and secondary sliders H2. This way you can place your most important content in the first slide.
Not sure if you use Jquery but if you do http://jquerytools.org/ offer great power and flexibility. Please note that I am NOT connected to them or work for them. We have just used their scripts on variious of our projects.
I had a quick look at your example and unfortunetely that would have a very hard time getting indexed since content is in the javascript. I would consider putting all content in the HTML and then just hide and show sections using Jquery instead.
Have a great day and good luck
Fredrik
-
Hi Paul,
Thank you so much for the detailed answer. deep down i worried this might be the case.
The truth is that the text in question is pretty much for SEO reasons only. Do you know f a better way, or another kind of script that would serve to have the text indexed?
Ron
-
The answer is that it actually depends very much on exactly what kind of coding is used to accomplish the effect, Ron.
In most cases, this kind of slider effect is accomplished using some variation of JavaScript. While Google has said it is "trying" to have it's crawlers recognize text from scripts, it almost never works that way.
So it won't be flagged as "hidden" text, because in fact Google won't even consider it to exist on the page.
An easy way to test is to view the source for the page in question - you'll see that none of the words of text actually exist on the page in any form, even in the code.
For the ultimate example of this - go into Google Webmaster Tools and use the Fetch as Googlebot tool to fetch the page. Then you'll see exactly the content that googlebot will see. It won't see the text, therefor it can't index and rank it. Ergo no SEO benefit at all.
Where you could get into trouble is if you did have text on the page designed to make googlebot think the page is about one thing, while using this kind of scripted text to try to show the visitor something completely different and unrelated. Google could then suspect you of cloaking and penalize accordingly. (Cloaking is when you intentionally show googlebot one thing and the user something different for nefarious purposes)
But if you're adding the text as a usability enhancement for your visitors in a way that googlebot doesn't happen to understand, you won't get any SEO benefit from it, but you also shouldn't be penalized for it.
Paul
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Indexed pages
Just started a site audit and trying to determine the number of pages on a client site and whether there are more pages being indexed than actually exist. I've used four tools and got four very different answers... Google Search Console: 237 indexed pages Google search using site command: 468 results MOZ site crawl: 1013 unique URLs Screaming Frog: 183 page titles, 187 URIs (note this is a free licence, but should cut off at 500) Can anyone shed any light on why they differ so much? And where lies the truth?
Technical SEO | | muzzmoz1 -
Is it good to redirect million of pages on a single page?
My site has 10 lakh approx. genuine urls. But due to some unidentified bugs site has created irrelevant urls 10 million approx. Since we don’t know the origin of these non-relevant links, we want to redirect or remove all these urls. Please suggest is it good to redirect such a high number urls to home page or to throw 404 for these pages. Or any other suggestions to solve this issue.
Technical SEO | | vivekrathore0 -
Using the Google Remove URL Tool to remove https pages
I have found a way to get a list of 'some' of my 180,000+ garbage URLs now, and I'm going through the tedious task of using the URL removal tool to put them in one at a time. Between that and my robots.txt file and the URL Parameters, I'm hoping to see some change each week. I have noticed when I put URL's starting with https:// in to the removal tool, it adds the http:// main URL at the front. For example, I add to the removal tool:- https://www.mydomain.com/blah.html?search_garbage_url_addition On the confirmation page, the URL actually shows as:- http://www.mydomain.com/https://www.mydomain.com/blah.html?search_garbage_url_addition I don't want to accidentally remove my main URL or cause problems. Is this the right way this should look? AND PART 2 OF MY QUESTION If you see the search description in Google for a page you want removed that says the following in the SERP results, should I still go to the trouble of putting in the removal request? www.domain.com/url.html?xsearch_... A description for this result is not available because of this site's robots.txt – learn more.
Technical SEO | | sparrowdog1 -
SEO value of InDesign pages?
Hi there, my company is exploring creating an online magazine built with Adobe's InDesign toolset. If we proceeded with this, could we make these pages "as spiderable" as normal html/css webpages? Or are we limited to them being less spiderable, or not at all spiderable?
Technical SEO | | TheaterMania1 -
How to block text on a page to be indexed?
I would like to block the spider indexing a block of text inside a page , however I do not want to block the whole page with, for example , a noindex tag. I have tried already with a tag like this : chocolate pudding chocolate pudding However this is not working for my case, a travel related website. thanks in advance for your support. Best regards Gianluca
Technical SEO | | CharmingGuy0 -
Splitting Page Authority with two URLs for the same page.
Hello guys, My website is currently holding two different URLs for the same page and I am under the impression such set up is dividing my Page Authority and Link Juice. We currently have the following page with both URLs below: www.wbresearch.com/soldiertechnologyusa/home.aspx
Technical SEO | | JoaoPdaCosta-WBR
www.wbresearch.com/soldiertechnologyusa/ Analysing the page authority and backlinks I identified that we are splitting the amount of backlinks (links from sites, social media and therefore authority). "/home.aspx"
PA: 67
Linking Root Domains: 52
Total Links: 272 "/"
PA: 64
Linking Root Domains: 29
Total Links: 128 I am under the impression that if the URLs were the same we would maximise our backlinks and therefore page authority. My Question: How can I fix this? Should I have a 301 redirect from the page "/" to the "/home.aspx" therefore passing the authority and link juice of “/” directly to “/homes.aspx”? Trying to gather thoughts and ideas on this, suggestions are much appreciated? Thanks!0 -
What should be use 301 or 302 redirection for 404 pages
Please suggest which redirection we should use for 404 pages- 301 or 302. If you can elaborate it with reason then it will be highly appreciated.
Technical SEO | | koamit0 -
What's the difference between a category page and a content page
Hello, Little confused on this matter. From a website architectural and content stand point, what is the difference between a category page and a content page? So lets say I was going to build a website around tea. My home page would be about tea. My category pages would be: White Tea, Black Tea, Oolong Team and British Tea correct? ( I Would write content for each of these topics on their respective category pages correct?) Then suppose I wrote articles on organic white tea, white tea recipes, how to brew white team etc...( Are these content pages?) Do I think link FROM my category page ( White Tea) to my ( Content pages ie; Organic White Tea, white tea receipes etc) or do I link from my content page to my category page? I hope this makes sense. Thanks, Bill
Technical SEO | | wparlaman0