Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How can I make sure Google is crawling a link from an iframe (video)?
-
Do they crawl backlinks from an iframe example from a Youtube video embedded in a blog post? TIA!
-
That test in no-way proves that the link passed PageRank.
I'm not sure about an href within an iframe, but equally I don't see any reason why you would do that?
-
So this test is not true:
https://www.seroundtable.com/google-iframe-link-14558.html
What if there's an a href tag (can be seen through "inspect element") within the iframe embedded in a webpage?
-
Nope, they can't. Those links are within the iframe.
If you're trying to get link equity from embedded video, it'll need to be with crediting text links or images outside of the actual player, as overlays are done with JavaScript, which doesn't pass PageRank by all accounts.
-
We created videos that will be embedded to other sites and we have to take advantage of the backlinks.
Regarding Youtube videos, it was just another question if Google can crawl backlinks (the hyperlinked Youtube logo within a video embedded to an external site or webpage) of a Youtube page.
-
Not quite sure I understand... Are you trying to build links to the YouTube video on youtube.com? if so, an embed through an iframe does sort of count as a "link" to the video (i.e. it provides ranking equity), but only in the unique case of YouTube videos.
Or are you wondering if links in the descriptions/overlays will pass PageRank?
-
This is for link building, Ryan. Do you think Google will still follow the links from an iframe?
-
Ah, ya, since OP is asking about these videos after they're embedded in other sites maybe he means will this be like link building? Or maybe he's asking about view lift and popularity of the original video? OP, I think you'll need to clarify some to get a more precise answer.
-
I wrote that post, and it's now nearly three years out of date.
They have got better at crawling iframes since then, and now you can get a video indexed if it's embedded in an iframe.
However, I don't think this relates to the question specifically?
-
Google see iframes as basically "a window into other sites". As such, no, they don't crawl links in this way - however, they will likely then crawl the page which the iframe references, and thereby pick up any backlinks there.
I'm not quite sure how this might related to an embedded YouTube video, but YouTube videos are indexed instantly anyway once they're published because Google don't need to crawl YouTube (they have access to the database)
-
Thanks Ryan! These videos will be embedded in other websites.
-
This post is fairly extensive in regards to your question and should give you ideas even beyond what you're asking here, http://moz.com/blog/hosting-and-embedding-for-video-seo but if you're looking for the highlight, this is still pretty applicable advice today:
Embed the content with HTML5 and JavaScript or Flash, but not an iframe
Unfortunately, Google are not very good at crawling iframes at the moment; so if you want videos to be indexed, you need to make sure you’re embedding content in an HTML5 player with Flash fallback, or a pure Flash player.
If the video is being embedded on your own site you can also create a transcription for it on the same page and then place the links in according spots. Cheers!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How can I find all broken links pointing to my site?
I help manage a large website with over 20M backlinks and I want to find all of the broken ones. What would be the most efficient way to go about this besides exporting and checking each backlink's reponse code? Thank you in advance!
Intermediate & Advanced SEO | | StevenLevine3 -
My site shows 503 error to Google bot, but can see the site fine. Not indexing in Google. Help
Hi, This site is not indexed on Google at all. http://www.thethreehorseshoespub.co.uk Looking into it, it seems to be giving a 503 error to the google bot. I can see the site I have checked source code Checked robots Did have a sitemap param. but removed it for testing GWMT is showing 'unreachable' if I submit a site map or fetch Any ideas on how to remove this error? Many thanks in advance
Intermediate & Advanced SEO | | SolveWebMedia0 -
Does anyone know of any tools that can help split up xml sitemap to make it more efficient and better for seo?
Hello All, We want to split up our Sitemap , currently it's almost 10K pages in one xml sitemap but we want to make it in smaller chunks splitting it by category or location or both. Ideally into 100 per sitemap is what I read is the best number to help improve indexation and seo ranking. Any thoughts on this ? Does anyone know or any good tools out there which can assist us in doing this ? Also another question I have is that should we put all of our products (1250) in one site map or should this also be split up in to say products for category etc etc ? thanks Pete
Intermediate & Advanced SEO | | PeteC120 -
Can a large fluctuation of links cause traffic loss?
I've been asked to look at a site that has lost 70/80% if their search traffic. This happened suddenly around the 17th April. Traffic dropped off over a couple of days and then flat-lined over the next couple of weeks. The screenshot attached, shows the impressions/clicks reported in GWT. When I investigated I found: There had been no changes/updates to the site in question There were no messages in GWT indicating a manual penalty The number of pages indexed shows no significant change There are no particular trends in keywords/queries affected (they all were.) I did discover that ahrefs.com showed that a large number of links were reported lost on the 17th April. (17k links from 1 domain). These links reappeared around the 26th/27th April. But traffic shows no sign of any recovery. The links in question were from a single development server (that shouldn't have been indexed in the first place, but that's another matter.) Is it possible that these links were, maybe artificially, boosting the authority of the affected site? Has the sudden fluctuation in such a large number of links caused the site to trip an algorithmic penalty (penguin?) Without going into too much detail as I'm bound by client confidentiality - The affected site is really a large database and the links pointing to it are generated by a half dozen or so article based sister sites based on how the articles are tagged. The links point to dynamically generated content based on the url. The site does provide a useful/valuable service/purpose - it's not trying to "game the system" in order to rank. That doesn't mean to say that it hasn't been performing better in search than it should have been. This means that the affected site has ~900,000 links pointing to is that are the names of different "entities". Any thoughts/insights would be appreciated. I've expresses a pessimistic outlook to the client, but as you can imaging they are confused and concerned. LVSceCN.png
Intermediate & Advanced SEO | | DougRoberts0 -
Do search engines crawl links on 404 pages?
I'm currently in the process of redesigning my site's 404 page. I know there's all sorts of best practices from UX standpoint but what about search engines? Since these pages are roadblocks in the crawl process, I was wondering if there's a way to help the search engine continue its crawl. Does putting links to "recent posts" or something along those lines allow the bot to continue on its way or does the crawl stop at that point because the 404 HTTP status code is thrown in the header response?
Intermediate & Advanced SEO | | brad-causes0 -
Do 404 Pages from Broken Links Still Pass Link Equity?
Hi everyone, I've searched the Q&A section, and also Google, for about the past hour and couldn't find a clear answer on this. When inbound links point to a page that no longer exists, thus producing a 404 Error Page, is link equity/domain authority lost? We are migrating a large eCommerce website and have hundreds of pages with little to no traffic that have legacy 301 redirects pointing to their URLs. I'm trying to decide how necessary it is to keep these redirects. I'm not concerned about the page authority of the pages with little traffic...I'm concerned about overall domain authority of the site since that certainly plays a role in how the site ranks overall in Google (especially pages with no links pointing to them...perfect example is Amazon...thousands of pages with no external links that rank #1 in Google for their product name). Anyone have a clear answer? Thanks!
Intermediate & Advanced SEO | | M_D_Golden_Peak0 -
Can Google read my backlink in Javascript??
Hi SeoMoz community! I have a software product, which our clients implement onto their websites. It is like a pop up box. I know that backlinks are very important for SEO ranking, and I really want to give our clients 2 options of product: 1. you can get the free/cheaper option if you use the code which has a keyworded backlink to our site on it 2. you can pay small fee if you don't want to use the version with a link to our site on it Now, the problem is that the product is written entirely in Javascript, and I don't think that Google crawls this, do they? Is there a way around this? Thanks for your help!
Intermediate & Advanced SEO | | qdigi0 -
How to prevent Google from crawling our product filter?
Hi All, We have a crawler problem on one of our sites www.sneakerskoopjeonline.nl. On this site, visitors can specify criteria to filter available products. These filters are passed as http/get arguments. The number of possible filter urls is virtually limitless. In order to prevent duplicate content, or an insane amount of pages in the search indices, our software automatically adds noindex, nofollow and noarchive directives to these filter result pages. However, we’re unable to explain to crawlers (Google in particular) to ignore these urls. We’ve already changed the on page filter html to javascript, hoping this would cause the crawler to ignore it. However, it seems that Googlebot executes the javascript and crawls the generated urls anyway. What can we do to prevent Google from crawling all the filter options? Thanks in advance for the help. Kind regards, Gerwin
Intermediate & Advanced SEO | | footsteps0