Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is The HREF Link "Title" Tag Needed on Mobile Websites?
-
Hello To Those Who Are Wiser Than I,
I am wondering if the href link "title" tag is needed, or serves any purpose, on mobile websites? Also, does it effect SEO in any way?
I ask because generally the href link title tag provides more information to the user when they scroll their mouse over the link - but this action does not happen on mobile! Users have no mouse and thus no extra information would be displayed.
I'm really wondering if it still matters for SEO purposes on mobile though.
-The UnEnlightened
-
Hi Chris,
Thank you - that is what I am referring too. I work in taxes so I'm sure that we have some visually impaired customers - I will continue to put the attribute on my links!
-The SEO Impaired
-
I think Linda is referring to the head <title>tag. The original question references link title attribute which is not shown on search engines.</p></title>
-
The title tag is what Google often shows in the search results. And external websites, especially social, often use title tags as anchor text for links. So even though you don't see it at the top of the page in mobile, it is an important element.
-
Hello Michael, valid question. Unfortunately, the title attribute has no direct impact on search rankings. The title tag is primarily for usability purposes.
I'm not positive, but I believe there may be apps for the visually impaired that read the content of title tags and alt image tags. If you're using the same content on desktop and mobile devices I would go ahead and add the title if it helps clarify what the link is pointing to.
However, it is not recommend to use the title tag if it is the same as your anchor text. It should be unique and clarify what the link is. Hope this helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Do we need both an .XML Sitemap and a .aspx sitemap?
Hi Mozers, We recently switched servers and it came to my attention that we have two sitemaps a XML version of the sitemap and a .aspx version of the sitemap. This came to light as the .aspx version of the sitemap is causing the site to come to a screeching halt as it has some complex code and lists over 80,000 products. My question is do we need both versions of the sitemap? My understanding is that the XML version is for Search Engine bots and the .aspx version is for customers. I can't imagine that anyone is using our .aspx version as it is basically a page with 80,000 links and it's buried away on the site, so we were hoping to kill off the .aspx version of the sitemap and keep the .xml version for Search Engine Bots. I wanted to check here first to make sure we did not any negative search engine implications. Any help would be most appreciated. Thanks so much! Patrick
Web Design | | gatorpool0 -
Bing Indexation and handling of X-ROBOTS tag or AngularJS
Hi MozCommunity, I have been tearing my hair out trying to figure out why BING wont index a test site we're running. We're in the midst of upgrading one of our sites from archaic technology and infrastructure to a fully responsive version.
Web Design | | AU-SEO
This new site is a fully AngularJS driven site. There's currently over 2 million pages and as we're developing the new site in the backend, we would like to test out the tech with Google and Bing. We're looking at a pre-render option to be able to create static HTML snapshots of the pages that we care about the most and will be available on the sitemap.xml.gz However, with 3 completely static HTML control pages established, where we had a page with no robots metatag on the page, one with the robots NOINDEX metatag in the head section and one with a dynamic header (X-ROBOTS meta) on a third page with the NOINDEX directive as well. We expected the one without the meta tag to at least get indexed along with the homepage of the test site. In addition to those 3 control pages, we had 3 pages where we had an internal search results page with the dynamic NOINDEX header. A listing page with no such header and the homepage with no such header. With Google, the correct indexation occured with only 3 pages being indexed, being the homepage, the listing page and the control page without the metatag. However, with BING, there's nothing. No page indexed at all. Not even the flat static HTML page without any robots directive. I have a valid sitemap.xml file and a robots.txt directive open to all engines across all pages yet, nothing. I used the fetch as Bingbot tool, the SEO analyzer Tool and the Preview Page Tool within Bing Webmaster Tools, and they all show a preview of the requested pages. Including the ones with the dynamic header asking it not to index those pages. I'm stumped. I don't know what to do next to understand if BING can accurately process dynamic headers or AngularJS content. Upon checking BWT, there's definitely been crawl activity since it marked against the XML sitemap as successful and put a 4 next to the number of crawled pages. Still no result when running a site: command though. Google responded perfectly and understood exactly which pages to index and crawl. Anyone else used dynamic headers or AngularJS that might be able to chime in perhaps with running similar tests? Thanks in advance for your assistance....0 -
Spanish website indexed in English, redirect to spanish or english version if i do a new website design?
Hi MOZ users, i have this problem. We have a website in Spanish Language but Google crawls it on English (it is not important the reasons). We re made the entire website and now we are planning the move. The new website will have different language versions, english, spanish and portuguese. Somebody tells me that we have to redirect the old urls (crawled on english) to the new english versions, not to the spanish (the real language of the firsts). Example: URL1 Language: Spanish - Crawled on English --> redirect to Language English version. the other option will be redirect to the spanish new version, which the visitor is waiting to find. URL1 Language: Spanish - Crawled on English --> redirect to Language Spanish version. What do you think? Which is the better option?
Web Design | | NachoRetta0 -
Spotted Hidden Omiod Links in Footer - What do you think is Going on Here?
Hi guys, Hoping one of you have come across this before. While taking a look at the source code for a website I've recently started working on, I spotted some 'display:none' code in the footer of the page. Here's a snapshot of the code: close XMETAhead title : 404 Page Not Found | ( 39 chrs ) [http://www.omiod.com/chrome-extensions/meta-seo-inspector/info.php?meta=description&cont=404 Page Not Found.](<a href=)" title="more about description" target="_blank" class="ad_seo_link">description : 404 Page Not Found( 170 chrs )[http://www.omiod.com/chrome-extensions/meta-seo-inspector/info.php?meta=keywords&cont=404, 404 error page,](<a href=) " title="more about keywords" target="_blank" class="ad_seo_link">keywords : 404, 404 error page ( 7 items )SCRIPThttp://www.google.com/s2/favicons?domain=www.google-analytics.com">www.google-analytics.com http://www.google-analytics.com/ga.js <div< a="">class="ad_seo_title">HTML5 report</div<>Doctype is not HTML5, there are no HTML5 tags, but at least no obsolete HTML tags were found. 1/5
Web Design | | ecommercebc0 -
Does Google follow links inside a <noscript>tag?</noscript>
I'm looking at making an embedable calculator and asking users to embed it to their website. I had the idea of using javascript to include the calculator which would also conatain a text link back to my site in order to gain some back links. If it's possible Google won't see the link (as they may not execute the javascript), is it safe to place the link in the <noscript>tag? If so, Will it be indexed and will Page Rank be passed?</span></p> <p>Thanks in advance for your answers. </p> <p>Anthony</p> <p><span style="color: #5e5e5e;"><br /></span></p></noscript>
Web Design | | BallyhooLtd0 -
Separate .mobi site or make .com site mobile friendly?
Our website now has enough mobile traffic to justify going mobile friendly, which it is not at this time. I am in favor of making a separate .mobi site designed specifically for mobile phones and smart phones for several reasons. It is cheaper, faster, and easier to accomplish. I think our mobile users will have a good experience though obviously not as much info as our full site. I would use ourdomain.mobi with link or a redirect for mobile users from from the main site. My top three choices for implementing that are http://allwebcodesign.com/setup/mobi-templates.htm#detailsarea
Web Design | | zharriet
Template that can be viewed by mobile or desktop. http://www.onbile.com/ http://www.networksolutions.com/mobile-website/index.jsp Does this seem like a good solution?1 -
How is link juice split between navigation?
Hey All, I am trying to understand link juice as it relates to duplicate navigation Take for example a site that has a main navigation contained in dropdowns containing 50 links (fully crawl-able and indexable), then in the footer of said page that navigation is repeated so you have a total of 100 links with the same anchor text and url. For simplicity sake will the link juice be divided among those 100 and passed to the corresponding page or does the "1st link rule" still apply and thus only half of the link juice will be passed? What I am getting at is if there was only one navigation menu and the page was passing 50 link juice units then each of the subpages would get passed 1link juice unit right? but if the menu is duplicated than the possible link juice is divided by 100 so only .5 units are being passed through each link. However because there are two links pointing to the same page is there a net of 1 unit? We have several sites that do this for UX reasons but I am trying to figure out how badly this could be hurting us in page sculpting and passing juice to our subpages. Thanks for your help! Cheers.
Web Design | | prima-2535090