AJAX & JQuery Tabs: Indexation & Navigation
-
Hi
I've two questions about indexing Tabs.
1. Let's say I have tabs, or an accordion that is triggered with Jquery. That means that all HTML is accessible and indexed by search engines. But let's say a search query is relevant to the content in Tab#3, while Tab#1 is the one that's open by default. Is there any way that Tab#3 would be open directly if it's more relevant to the search query?
2. AJAX Tabs: We have pages that have Tabs triggered by AJAX (example: http://www.swisscom.ch/en/residential/help/loesung/entfernen-sie-sim-lock.html). I'm wondering about the current best practice. Google recommends HTML Snapshots. A newer SEOMoz Article talks about pushState(). What's the way to go here?
Or in other words: How to get Tabs & Accordion content indexed and allow users to navigate directly to it?
-
hello Philip,
robots.txt file allows you to tell the web bots that crawl your site what is a link and what is not a link that you want to show to the world and to Google
Most search engines will analyze and follow a link only if it contains three query string parameters or fewer.
many parameters in the link shown you have 5 parameters they are what come after the1st / as shown below you have 5. You can block off certain parameters with robots.txt
/en/residential/help/loesung/entfernen-sie-sim-lock
http://www.swisscom.ch/en/residential/help/loesung/entfernen-sie-sim-lock.htm
for some reason whenever I go to your link you have posted I get this error
The requested URL /system/sling/cqform/defaultlogin.html was not found on this server.
http://msdn.microsoft.com/en-us/library/ff723936(v=expression.40).aspx
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=1235687
please links above regarding URL parameters and Microsoft and Google agree that too many parameters and they will not search the link.
I hope this is been helpful
sincerely,
Thomas
-
Hi Thomas
Thanks for the resources. I'll have to check with IT which solution seems most practical.
Though I don't understand two points:
- Where does robots.txt come into the game?
- How do we have too many parameters in http://www.swisscom.ch/en/residential/help/loesung/entfernen-sie-sim-lock.htm? Can you specify this?
Thanks!
/Philipp
-
I forgot to add this but this is very relevant the software could make your script is Seo friendly
-
I would do some serious decoding
if you want to see what Google sees this as a great tool http://www.screamingfrog.co.uk/seo-spider/
see what screaming frog tells you then try to fix it in Unicode 8
can fix some of this by going to the bottom of this page and changing your robot.txt.
https://developers.google.com/webmasters/ajax-crawling/docs/learn-more
https://developers.google.com/webmasters/control-crawl-index/docs/robots_txt
you have too many parameters http://www.swisscom.ch/en/residential/help/loesung/entfernen-sie-sim-lock.html
in my opinion you forgot some serious code changing To do.
if you want to use one of the tools Google recommended you'll need I-frame crawling and you can do that with this along with the 2nd URL
https://github.com/crawljax/crawljax/blob/master/CHANGELOG.md
http://code.google.com/p/selenium/issues/detail?id=387
http://www.unicode.org/faq/utf_bom.html
because it is Java based there are some great tools found here as well
I wish you the best and hope that this is helpful,
Thomas
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
404 errors & old unused pages
I am using shopify and I need to delete some old pages which are coming up as 404 errors (product no longer available!) does anyone know where you go to delete these pages which are no longer needed?
Web Design | | carleyb0 -
NO Meta description pulling through in SERP with react website - Requesting Indexing & Submitting to Google with no luck
Hi there, A year ago I launched a website using react, which has caused Google to not read my meta descriptions. I've submitted the sitemap and there was no change in the SERP. Then, I tried "Fetch and Render" and request indexing for the homepage, which did work, however I have over 300 pages and I can't do that for every one. I have requested a fetch, render and index for "this url and linked pages," and while Google's cache has updated, the SERP listing has not. I looked in the Index Coverage report for the new GSC and it says the urls and valid and indexable, and yet there's still no meta description. I realize that Google doesn't have to index all pages, and that Google may not also take your meta description, but I want to make sure I do my due diligence in making the website crawlable. My main questions are: If Google didn't reindex ANYTHING when I submitted the sitemap, what might be wrong with my sitemap? Is submitting each url manually bad, and if so, why? Am I simply jumping the gun since it's only been a week since I requested indexing for the main url and all the linked urls? Any other suggestions?
Web Design | | DigitalMarketingSEO1 -
International Websites - Hosting, Domain Names, & Configuration
What is the best way to configure a website that targets a number of countries and languages around the world? For example, Apple has websites optimized for just about every country and language in the world (see: https://www.apple.com/choose-country-region/). When you choose the UK it takes you to: https://www.apple.com/uk/ When you choose China it take you to: https://www.apple.com/cn/ Etc. When you go to apple.co.uk it forwards you to the UK version of the website. The same is true for apple.cn. Is this the ideal way to set it up? I have also seen websites that have each version of the website on its own TLD such as exampleBrand.co.uk and exampleBrand.cn - in this example they don't forward to the .com. My concern with Apple's solution is SEO and hosting. Do consumers favor seeing their country's TLD in search results over exampleBrand.com/uk? For hosting, shouldn't the mainland China version of the website be hosted in China? Is it possible to just host a folder of a website in a certain country such as the cn folder for China? Any suggestions would be appreciated. I was unable to find much info on this.
Web Design | | rx3000 -
What is your opinion in the use of jquery for a continuous scroll type of page layout?
So, I'm in 2 minds about this; let me start with a bit of background info. Context
Web Design | | ChrisAshton
We have a new client who is in the final days of their new site design and were when they first contacted us. Their design essentially uses 5 pages, each with several pages worth of content on each, separated with the use of jquery. What this means is a user can click a menu item from a drop-down in the nav and be taken directly to that section of content like using internal anchor links as if it were a separate page, or they can click the top-level nav item and scroll through each "sub-page" without having to click other links. Vaguely similar to Google's "How Search Works" page if each sector of that page had it's own URL, only without the heavy design elements and slow load time. In this process, scrolling down to each new "sub-page" changes the URL in the address bar and is treated as a new page as far as referencing the page, adding page titles, meta descriptions, backlinks etc. From my research this also means search engines don't see the entire page, they see each sub-page as their own separate item like a normal site. My Reservations I'm worried about this for several reasons, the largest of them being that you're essentially presenting the user with something different to the search engines. The other big one being that I just don't know if search engines really can render this type of formatting correctly or if there's anything I need to look out for here. Since they're so close to launching their new site, I don't have time to set up a test environment and I'm not going to gamble with a new corporate website but they're also going to be very resistant to the advice of "start the design over, it's too dangerous". The Positives
For this client in particular, the design actually works very well. Each of these long pages is essentially about a different service they offer and the continuous scrolling through the "sub-pages" acts as almost a workflow through the process, covering each step in order. It also looks fantastic, loads quickly and has a very simple nav so the overall user experience is great. Since the majority of my focus in SEO is on UX, this is my confusion. Part of me thinks that obscuring the other content on these pages and only showing each individual "sub-page" to search engines is an obvious no-no, the other part of me feels that this kind of user experience and the reasonable prevalence of AJAX/Paralax etc means search engines should be more capable of understanding what's going on here. Can anyone possibly shed some light on this with either some further reading or first-hand experience?0 -
Duplicate Content? Designing new site, but all content got indexed on developer's sandbox
An ecommerce I'm helping is getting a complete redesign. Their developer had a sandbox version of their new site for design & testing. Several thousand products were loaded into the sandbox site. Then Google/Bing crawled and indexed the site (because developer didn't have a robots.txt), picking up and caching about 7,200 pages. There were even 2-3 orders placed on the sandbox site, so people were finding it. So what happens now?
Web Design | | trafficmotion
When the sandbox site is transferred to the final version on the proper domain, is there a duplicate content issue?
How can the developer fix this?0 -
Duplicate Content & Canonicals
I am a bit confused about canonicals and whether they are "working" properly on my site. In Webmaster Tools, I'm showing about 13,000 pages flagged for duplicate content, but nearly all of them are showing two pages, one URL as the root and a second with parameters. Case in point, these two are showing as duplicate content: http://www.gallerydirect.com/art/product/vincent-van-gogh/starry-night http://www.gallerydirect.com/art/product/vincent-van-gogh/starry-night?substrate_id=3&product_style_id=8&frame_id=63&size=25x20 We have a canonical tag on each of the pages pointing to the one without the parameters. Pages with other parameters don't show as duplicates, just one root and one dupe per listing, So, am I not using the canonical tag properly? It is clearly listed as:Is the tag perhaps not formatted properly (I saw someone somewhere state that there needs to be a /> after the URL, but that seems rather picky for Google)?Suggestions?
Web Design | | sbaylor0 -
CSS vs Javascript vs JQuery drop down navigation
For a user / seo perspective, what is the best way to code a drop down menu nav bar? Is it best to use css, javascript or a scripting library like jquery? I am thinking about overall best practice that will not have a negative impact on serps. I am also thinking about what will work best on all types of devices i.e. desk tops, lap tops, smart phones and tablets. What are the Pro's & Cons of Using CSS for Drop Down Menus. What are the Pro's & cons of using Javascript for drop down menus. And the same question for jquery. Thank you all in advance for your ideas.
Web Design | | bronxpad0 -
3 Products & 50 Options each, How does Google handle product variant or options?
We are selling furnace filter and we might move our existing store host by BigCommerce to Americommerce or Corecommerce. Before moving the store, I have a questions about our online store structure. We are selling 3 different furnace filters, GOLD, SILVER and BRONZE Series. Each furnace filter come in about 50 different sizes, for a total of about 150 different products. The way our store is setup now, it is 150 different product, 150 different URL, 150 different page name... The way it is setup now, might look like duplicate content. All the product page are the same, all the pictures are the same, the only thing that change, is the furnace filter size in the product description. Look at those pages for example: http://www.furnacefilterscanada.com/20x20x4-Furnace-Filters/ http://www.furnacefilterscanada.com/categories/2-Inches-Thick-Filters/10x20x2-inches/ http://www.furnacefilterscanada.com/categories/2-Inches-Thick-Filters/16x25x2-inches/ Would it be better to only have 3 products and 50 variables or size options? What would be the best structure in a SEO point of view? One thing we have to keep in mind, when searching for a furnace filter, shooper will use keywords like: 16x25x4 furnace filter filter 20x20x1 air furnace filter 10x20x1 furnace filter 24x24x4 canada furnace filter Most of the Google search will included the filter size_._ How does Google handle product variant or options_?_ If I have 3 products, I will have only 3 URL and 3 different page name. I know for the shoppers, 3 products with sizes options might provide a better experience, but what about Google ranking the products? What is opinion the best online store structure in our case? Thank you for your help, preciouse time and support. BigBlaze www.furnacefilterscanada.com/
Web Design | | BigBlaze2050