Indexing content behind a login
-
Hi,
I manage a website within the pharmaceutical industry where only healthcare professionals are allowed to access the content. For this reason most of the content is behind a login.
My challenge is that we have a massive amount of interesting and unique content available on the site and I want the healthcare professionals to find this via Google!
At the moment if a user tries to access this content they are prompted to register / login. My question is that if I look for the Google Bot user agent and allow this to access and index the content will this be classed as cloaking? I'm assuming that it will.
If so, how can I get around this? We have a number of open landing pages but we're limited to what indexable content we can have on these pages!
I look forward to all of your suggestions as I'm struggling for ideas now!
Thanks
Steve
-
Thanks everyone... It's not as restrictive as patient records... Basically, because of the way our health service works in the UK we are not allowed to promote material around our medicines to patients, it should be restricted only to HCP's. If we are seen to be actively promoting to patients we run the risk of a heavy fine.
For this reason we need to take steps to ensure that we only target this information towards HCP's and therefore we require them to register before being able to access the content...
My issue is that HCP's may search for a Brand that we supply but we have to be very careful what Brand information we provide outside of log-in. Therefore the content we can include on landing pages cannot really be optimised for the keywords that they are searching for! Hence why I want the content behind log-in indexed but not easily available without registering...
It's a very difficult place to be!
-
I guess I was just hoping for that magic answer that doesn't exist! It's VERY challenging to optimise a site with these kinds of restrictions but I get I just need to put what I can on the landing pages and optimise as best I can with the content I can show!
We also have other websites aimed at patients where all the content is open so I guess I'll just have to enjoy optimising these instead
Thanks for all your input!
Steve
-
Steve,
Yes that would be cloaking. I wouldn't do that.
As Pete mentioned below, your only real options at this point are to make some of the content, or new content, available for public use. If you can't publish abstracts at least, then you'll have to invest in copywriting content that is legally available for the public to get traffic that way, and do your best to convert them into subscribers.
-
Hi Steve
If it can only be viewed legally by health practitioners who are members of your site, then it seems to me you don't have an option as by putting any of this content into the public domain on Google by whatever method you use will be deemed illegal by whichever body oversees it.
Presumably you cannot also publish short 25o word summaries of the content?
If not, then I think you need to create pages that are directly targeted at marketing the site to health practitioners. Whilst the pages won't be able to contain the content you want to have Google index, they could still contain general information and the benefits of becoming a subscriber.
Isn't that the goal of the site anyway, i.e. to be a resource to health practitioners? So, without being able to make the content public, you have to market to them through your SEO or use some other form or indirect or direct marketing to encourage them to the site to sign up.
I hope that helps,
Peter -
Thanks all... Unfortunately it is a legal requirement that the content is not made publicly available but the challenge then is how do people find it online!
I've looked at first click free and pretty much ever other option I could think of and yet to find a solution
My only option is to allow Google Bot through the authentication which will allow it to index the content but my concern is that this is almost certainly cloaking...
-
Please try looking at "First Click Free" by Google
https://support.google.com/webmasters/answer/74536?hl=en
I think this is along the lines of what you are looking for.
-
Hi Steve
As you already know, if a page is not crawlable it's not indexable. I don't think there is any way around this without changing the strategy of the site. You said, _"We have a number of open landing pages but we're limited to what indexable content we can have on these pages". _Is that limitation imposed by a legal requirement or something like that, or by the site owners because they don't want to give free access?
If the marketing strategy for the site is to grow the membership, then as it's providing a content service to its members then it has to give potential customers a sample of its wares.
I think there are two possible solutions.
(1) increase the amount of free content available on the site to give the search engines more content to crawl and make available to people searching or
(2) Provide a decent size excerpt, say the first 250 words of each article as a taster for potential customers and put the site login at the point of the "read more". That way you give the search engines something to get their teeth into which is of a decent length but it's also a decent size teaser to give potential customers an appetite to subscribe.
I hope that helps,
Peter
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Somebody took an article from my site and posted it on there own site but gave it credit back to my site is this duplicate content?
Hey guys, This question may sound a bit drunk, but someone copied our article and re-posted it on their site the exact article, however the article was credited to our site and the original author of the article had approved the other site could do this. We created the article first though, Will this still be regarded as duplicate content? The owner of the other site has told us it wasn't because they credited it. Any advice would be awesome Thanks
White Hat / Black Hat SEO | | edward-may0 -
Site De-Indexed except for Homepage
Hi Mozzers,
White Hat / Black Hat SEO | | emerald
Our site has suddenly been de-indexed from Google and we don't know why. All pages are de-indexed in Google Webmaster Tools (except for the homepage and sitemap), starting after 7 September: Please see screenshot attached to show this: 7 Sept 2014 - 76 pages indexed in Google Webmaster Tools 28 Sept until current - 3-4 pages indexed in Google Webmaster Tools including homepage and sitemaps. Site is: (removed) As a result all rankings for child pages have also disappeared in Moz Pro Rankings Tracker. Only homepage is still indexed and ranking. It seems like a technical issue blocking the site. I checked for robots.txt, noindex, nofollow, canonical and site crawl for any 404 errors but can't find anything. The site is online and accessible. No warnings or errors appear in Google Webmaster Tools. Some recent issues were that we moved from Shared to Dedicated Server around 7 Sept (using same host and location). Prior to the move our preferred domain was www.domain.com WITH www. However during the move, they set our domain as domain.tld WITHOUT the www. Running a site:domain.tld vs site:www.domain.tld command now finds pages indexed under non-www version, but no longer as www. version. Could this be a cause of de-indexing? Yesterday we had our host reset the domain to use www. again and we resubmitted our sitemap, but there is no change yet to the indexing. What else could be wrong? Any suggestions appeciated. Thanks. hDmSHN9.gif0 -
Cloaking for better user experience and deeper indexing - grey or black?
I'm working on a directory that has around 800 results (image rich results) in the top level view. This will likely grow over time so needs support thousands. The main issue is that it is built in ajax so paginated pages are dynamically generated and look like duplicate content to search engines. If we limit the results, then not all of the individual directory listing pages can be found. I have an idea that serves users and search engines what they want but uses cloaking. Is it grey or black? I've read http://moz.com/blog/white-hat-cloaking-it-exists-its-permitted-its-useful and none of the examples quite apply. To allow users to browse through the results (without having a single page that has a slow load time) we include pagination links but which are not shown to search engines. This is a positive user experience. For search engines we display all results (since there is no limit the number of links so long as they are not spammy) on a single page. This requires cloaking, but is ultimately serving the same content in slightly different ways. 1. Where on the scale of white to black is this? 2. Would you do this for a client's site? 3. Would you do it for your own site?
White Hat / Black Hat SEO | | ServiceCrowd_AU0 -
Have just submitted Disavow file to Google: Shall I wait until after they have removed bad links to start new content lead SEO campaign?
Hi guys, I am currently conducting some SEO work for a client. Their previous SEO company had built a lot of low quality/spam links to their site and as a result their rankings and traffic have dropped dramatically. I have analysed their current link profile, and have submitted the spammiest domains to Google via the Disavow tool. The question I had was.. Do I wait until Google removes the spam links that I have submitted, and then start the new content based SEO campaign. Or would it be okay to start the content based SEO campaign now, even though the current spam links havent been removed yet.. Look forward to your replies on this...
White Hat / Black Hat SEO | | sanj50500 -
Can I Point Multiple Exact Match Domains to a Primary Domain? (Avoiding Duplicate Content)
For example, lets say I have these 3 domains: product1.com product2.com product.com The first 2 domains will have very similar text content, with different products. The product.com domain will be similar content, with all of the products in one place. Transactions would be handled through the Primary domain (product.com) The purpose of this would be to capitalize on the Exact match domain opportunities. I found this seemingly old article: http://www.thesitewizard.com/domain/point-multiple-domains-one-website.shtml The article states that you can avoid duplicate content issues, and have all links attributed to the Primary domain. What do you guys think about this? Is it possible? Is there a better way of approaching this while still taking advantage of the EMD?
White Hat / Black Hat SEO | | ClearVisionDesign0 -
DIV Attribute containing full DIV content
Hi all I recently watched the latest Mozinar called "Making Your Site Audits More Actionable". It was presented by the guys at seogadget. In the mozinar one of the guys said he loves the website www.sportsbikeshop.co.uk and that they have spent a lot of money on it from an SEO point of view (presumably with seogadget) so I decided to look through the source and noticed something I had not seen before and wondered if anyone can shed any light. On this page (http://www.sportsbikeshop.co.uk/motorcycle_parts/content_cat/852/(2;product_rating;DESC;0-0;all;92)/page_1/max_20) there is a paragraph of text that begins with 'The ever reliable UK weather...' and when you via the source of the containing DIV you will notice a bespoke attribute called "threedots=" and within it, is the entire text content for that DIV. Any thoughts as to why they would put that there? I can't see any reason as to why this would benefit a site in any shape or form. Its invalid markup for one. Am I missing a trick..? Thoughts would be greatly appreciated. Kris P.S. for those who can't be bothered to visit the site, here is a smaller version of what they have done: This is an introductory paragraph of text for this page.
White Hat / Black Hat SEO | | yousayjump0 -
Does posting a source to the original content avoid duplicate content risk?
A site I work with allows registered user to post blog posts (longer articles). Often, the blog posts have been published earlier on the writer's own blog. Is posting a link to the original source a sufficient preventative solution to possibly getting dinged for duplicate content? Thanks!
White Hat / Black Hat SEO | | 945010 -
Possibly a dumb question - 301 from a banned domain to new domain with NEW content
I was wondering if banned domains pass any page rank, link love, etc. My domain got banned and I AM working to get it unbanned, but in the mean time, would buying a new domain, and creating NEW content that DOES adhere to the google quality guidelines, help at all? Would this force an 'auto-evaluation' or 're-evaluation' of the site by google? or would the new domain simply have ZERO effect from the 301 unless that old domain got into google's good graces again.
White Hat / Black Hat SEO | | ilyaelbert0