Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Client Portal and SEO Considerations?
-
Hi Moz and Moz fans,
We are looking to add a client portal to the website. Basically, I haven't found too much on this with regard to SEO. The idea would be that certain parts of the website would be hidden under a pay wall and for subscribers, they would be able to see all content. I am wondering if anyone has any experience with that and what SEO considerations to take into account.
One thing we are particularly concerned about is how Google will index the portions of the website behind the pay wall, if at all. Obviously, we would rather that they don't index it, so that people can't find a way to get to the info without paying. I would imagine it would have to do with the type of coding, however, I am not a coding guru, so I am not 100% on that.
Anyway, anyone that has any experience in this kind of thing and can comment on this at all, any comment is welcome. Also, any documentation that could be helpful would be welcome too.
Thanks
-
Hi Sergey,
Thanks for this. I have read that article at some point a while back. Interesting read. I was just sort of sounding out a few ideas in my last post. The second option seems to be the best way to go.
Do you know of any other methods to do this that is common in the industry?
Thanks,
Brian -
Hi Brian,
Google can definitely crawl Javascript, although it is limited in a sense. Here is a great article addressing this.
I have seen many websites use the solution that you mention in your second paragraph. That is, creating a separate section of the domain (or creating a subdomain) for users who are logging in. GoogleBot will not be able to crawl anything that is beyond a login portal, so this is the most common solution. Of course you can also disallow the specific path or subdomain in robots.txt as well!
It comes down to how much content that you have behind the login portal. You want to figure out which part of your site you want to be crawled, and which parts you don't want. After that point it's a matter of figuring out the best solution for blocking that content from crawlers.
Hope that helps!
-
Hi Andy,
Thanks for the reply.
This documentation was helpful. You mentioned the potential for the content to be seen in the cached version of pages. I read the Search Engine Land article there and shes sort of getting at how you don't want to block your content to search engines because then they can't be indexed and ranked. Let's say we don't really care about those parts of the pages that we wan't to block in terms of being indexed, which I sort of mentioned above. If we were to some how display those parts of the pages, images and text, in JavaScript, search engines wouldn't be able to crawl. Would those portions of the page still be visible in the cached version? I would imagine not if the search engines can't crawl JS, but not sure.
A second option I thought of was to create a separate subdomain for logged in users that would only be accessible through a log in form. Search engines cannot follow forms, so I would imagine, that gated content wouldn't be visible to crawlers and therefore not indexed. You could go one step farther and block the subdomain in robots.txt. Do you think that could be a possible solution?
Thanks again for the reply last week.
-Brian
-
Hi Brian,
I have a client working on correcting this issue with his site at the moment. They run a big media site that allows access once paid, but so may of these sites suffer with the same issue and because they allow Google to index the whole text, but only show a portion of it, this means that if you just look at the cached version, you can read it without paying.
In terms of correcting it, I would first have a read on how Google handles subscription sites. You can find that info here. Google prefers the "First click free" model.
There is additional reading on this subject over at Search Engine Land. First Click Free (FCF) is what you want to be looking into in more detail.
I hope this helps a little.
-Andy
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What type of website is best for seo.
I need a new website for my health insurance business. What type is best for SEO? Many thanks
Web Design | | laurentjb1 -
Has anyone ever tested to see if having an ads.txt file provided any SEO lift?
I know that the ads.txt system is designed to prevent ad fraud and technically has nothing to do with search. That said, the presence of such a file would seem to be an indicator of overall site quality because it would show that a site owner wants to participate in a fraud-free system. Has anyone ever tested that? If so, they don't seem to have published their results. Maybe it's a secret weapon that some pros are using and not sharing?
Web Design | | scodtt0 -
How is Single Page Application (SPA) bad for SEO
Hi guys. I am quite inspired of SPA technique. It's really amazing when all your interaction with the site is going on the fly and you don't see any page reloads. I've started implementing the site with this instruction and already found nice guys to make the design. The only downside of the using SPA which I can see **is the **SEO part. That's because the URL does not really change and different pages don't have their unique URL addresses.
Web Design | | Billy_gym
Actually they have, but it looks like: yoursite.com/#/products yoursite.com/#/prices yoursite.com/#/contact So all of them goes after # and being just anchors. For Google this mean all of these pages is just yoursite.com/ My question is what is really proven method to implement the URL structure in Single Page Application, so all the pages indexed by Google correctly (sorry I don't mention the other search engines because of market share). The other question, of course, is examples. It will be great to see real life site examples, better authority sites, which use SPA technique and well indexed by search engines.1 -
Will having two wordpress themes installed hurt seo?
We currently have 3 sites built on WordPress that have little to no blogging capabilities. Currently, all published posts show up on a /category page which does not resemble the traditional blog format and is not aesthetically pleasing. We would like to have a more traditional blog and are considering installing a second wordpress theme on the site which will strictly be used for /blog and all the posts. My question is will having the second WordPress installation on the sites hurt us in any way on the SEO front and if we go this way should we place the install in a subfolder or on a subdomain? Is there anything else we need to worry about with making this transition? Thank you in advance for the advice! Patrick
Web Design | | PlanetDISH0 -
Best SEO practice - Umbrella brand with several domains
Hi, we have several blogs and comparison sites on specific topics. All the domains rank on top positions in very competitive niche markets. We think that we can get more profit out of the domains when we put them under an umbrella brand. Customers that visit domain A can then also find products easily on domain B. We see this for example on health.com, with several brands in the top. To maintain or improve our rankings i'm looking for specific information for the link structure. For example, is it better to have the 'about us'/rel=author on each domain, with contributors on that specific domain or is it better to have them all in the (umbrella) brand domain. At the moment we have the structure like this: domainA.com, domainA.com/blog, domainA.com/about-us and domainB.com, domainB.com/blog, domainB.com/about-us. I think to maintain the rankings it is best to keep specific content (like blog/ about us) on the domain. So is it the best to just do side wide links with a logo (like health.com) and what about hosting? We work with wordpress, so all domains will be hosted on one ip? when we use the multiple site option of WP? All information on this topic is more than welcome 🙂
Web Design | | remkoallertz0 -
Can white text over images hurt your SEO?
Hi everyone, I run a travel website that has about 30 pre-search city landing pages. In a redesign last year we added large "hero" images to the top of the page, and put our h1 headlines on top of them in white. The result is attractive, but I'm wondering if Google could be reading this page as "white text on white page", which is an obvious no-no, especially if it could seem that we're trying to hide text. Here's an example: http://www.eurocheapo.com/paris/ H1: Expert reviews of cheap hotels in Paris I should add that our SERPs for these city pages has dropped (for "Cheap hotels in X"), but it could obviously be related to other issues. Any advice would be appreciated. Many thanks! Tom
Web Design | | TomNYC0 -
What is the difference of HTML5 and web 2.0? What is web 2.0 and is this better for seo?
A little bit confused with the new stuff. The web 2.0 webpages are so much better? What changes?
Web Design | | Naghirniac0 -
Combining web pages and it's affects on SEO?
We are looking into amending a website we are working on to try and combine 2 or 3 current pages onto one page. This site is similar to an estate agents site and currently has images, map, floor plan sub pages etc. Can anyone tell me, if we were to combine these pages and include the above details on one page, how that would affect the current search engine rankings?
Web Design | | SoundinTheory0