Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
The main navigation is using JS, will this have a negative impact on SEO?
-
Hi mozzers,
We just redesigned our homepage and discovered that our main nav is using JS and when disabling JS, no main nav links was showing up. Is this still considered bad practice for SEO? https://cl.ly/14ccf2509478
thanks
-
Thanks Jose!
-
Googlebot does not render JavaScript
We hear very often about JavaScript rendering is the assumption that the crawler is responsible for the rendering. But that's not the case. The rendering is done by the indexer (from Google Caffeine).More specifically, the Web Rendering Service (WRS), which is part of Google Caffeine. And this is exactly where the second great difficulty lies in the area of JavaScript SEO.
Gary Illyes says: "Also, to clarify, WRS is a subsystem of Caffeine, just like the PageRanker (cos yes, we still use it) or the canonicalization algorithm."
The process flow between the crawler and the indexer:
HTML files are downloaded.
CSS and JavaScript data will be downloaded in parallel.
WRS (Web Rendering Service) (part of Google Caffeine) is used to run JS.
WRS renders all files.
Now Google Caffeine can index the content if necessary.
And that's when Googlebot can crawl new links.
Why is that a problem? For the crawler to be able to enter the content and follow the links, it must first pass the data to the indexer. The indexer renders the content and returns it to the crawlers. Only now can the crawler see the content and follow the links. On the next page, the game starts again. Page by page. (many processes)Unlike conventional pages, where the crawler crawls numerous pages and, after crawling, passes everything collectively to the indexer. (few processes)
This process presents another challenge to search engines.
Regards
-
The only bad SEO Practice regarding JS that I've of is trying to hide the JS files from Search Engines, apparently, Google needs to render a whole page as it indexing it; so blocking any JS (& CSS for that matter) isn't a good thing.
-
Hi There,
JS rendered UI isn't negative, it is however not a good practice also.
You can find complete information to the concept of JS rendered pages and UI on this link, https://www.portent.com/blog/seo/javascript-bad-for-seo.htm .
I hope this helps.
Regards,
Vijay
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Does content in collapsible menus negatively affect SEO or featured snippets?
We want to confirm whether content in collapsible menus negatively affects SEO and/or featured snippets on Google. We're hoping to add a menu to answer some frequently asked questions and attract featured snippets, while also creating a positive user experience/not clogging up the page. Here is an example of the style of menu we're using now, the troubleshooting menu: http://www.lynden.com/help/index.html Appreciate your help!
Intermediate & Advanced SEO | | RyanD.1 -
Does Disavowing Links Negate Anchor Text, or Just Negates Link Juice
I'm not so sure that disavowing links also discounts the anchor texts from those links. Because nofollow links absolutely still pass anchor text values. And disavowing links is supposed to be akin to nofollowing the links. I wonder because there's a potential client I'm working on an RFP for and they have tons of spammy directory links all using keyword rich anchor texts and they lost 98% of their traffic in Pengiun 1.0 and haven't recovered. I want to know what I'm getting into. And if I just disavow those links, I'm thinking that it won't help the anchor text ratio issues. Can anyone confirm?
Intermediate & Advanced SEO | | MiguelSalcido0 -
SEO impact difference between a URL Rewrite and 301 redirect
Hi guys and girls! Just putting a new site live, we changed the URL from one thing to another and I created a 301 file redirecting the urls like for like. The developer installing it has created a different file with columns like: RewriteRule ^page/ http://www.site/page [R=301,L] RewriteRule ^/page/ http://www.site/page [R=301,L] What's the difference? The page redirects but is there a difference between the 301 redirect and this URL rewrite in terms of SEO and link value?
Intermediate & Advanced SEO | | shloy23-2945840 -
If a website Uses <select>to dropdown some choices, will Google see every option as Content Or Hyperlink?</select>
If a website Uses <select> to dropdown some choices, will Google see every option as Content Or Hyperlink?</select>
Intermediate & Advanced SEO | | Zanox0 -
Will changing a subdirectory name negatively effect local ranking?
We submitted a group of 50+ franchise stores into UBL to fulfill directory listings back in September. We are now looking at changing the some of the URL structure to include city names. Example: website.com/store/store-name(not city) to website.com/location/city-store-name Will changing the subdirectory and resubmitting to the directory aggregators negatively effect their search results? Thanks, Jake
Intermediate & Advanced SEO | | AESEO0 -
Are there any negative effects to using a 301 redirect from a page to another internal page?
For example, from http://www.dog.com/toys to http://www.dog.com/chew-toys. In my situation, the main purpose of the 301 redirect is to replace the page with a new internal page that has a better optimized URL. This will be executed across multiple pages (about 20). None of these pages hold any search rankings but do carry a decent amount of page authority.
Intermediate & Advanced SEO | | Visually0 -
How do you implement dynamic SEO-friendly URLs using Ajax without using hashbangs?
We're building a new website platform and are using Ajax as the method for allowing users to select from filters. We want to dynamically insert elements into the URL as the filters are selected so that search engines will index multiple combinations of filters. We're struggling to see how this is possible using symfony framework. We've used www.gizmodo.com as an example of how to achieve SEO and user-friendly URLs but this is only an example of achieving this for static content. We would prefer to go down a route that didn't involve hashbangs if possible. Does anyone have any experience using hashbangs and how it affected their site? Any advice on the above would be gratefully received.
Intermediate & Advanced SEO | | Sayers1 -
Should I nofollow the main navigation on certain pages?
We have several large Ecommerce sites with hundreds of links on each page. I have been trying to think of ways to focus our internal linking to increase certain pages relevancy. My thought was to put nofollow in the main navigation (since there are hundreds of links there controlled by dropdowns) and only follow the links on each page for the products we are selling and promoting (15-20 links). I would still be using a sitemap that includes the links. Is this a terrible idea? if a link is nofollowed in the main navigation does that still count as the one mention for google if it points to the same page that a normal link points too that is in the content of the page? since all of the main navigation is the same on every page of the website would it be good to only put nofollow on the subpages/subsections and leave the home page navigation alone (that would allow the spiders to crawl all of those links on the home page but not crawl those same links on the subsections where I could then focus the linking).
Intermediate & Advanced SEO | | bigtimeseo0