Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Client Portal and SEO Considerations?
-
Hi Moz and Moz fans,
We are looking to add a client portal to the website. Basically, I haven't found too much on this with regard to SEO. The idea would be that certain parts of the website would be hidden under a pay wall and for subscribers, they would be able to see all content. I am wondering if anyone has any experience with that and what SEO considerations to take into account.
One thing we are particularly concerned about is how Google will index the portions of the website behind the pay wall, if at all. Obviously, we would rather that they don't index it, so that people can't find a way to get to the info without paying. I would imagine it would have to do with the type of coding, however, I am not a coding guru, so I am not 100% on that.
Anyway, anyone that has any experience in this kind of thing and can comment on this at all, any comment is welcome. Also, any documentation that could be helpful would be welcome too.
Thanks
-
Hi Sergey,
Thanks for this. I have read that article at some point a while back. Interesting read. I was just sort of sounding out a few ideas in my last post. The second option seems to be the best way to go.
Do you know of any other methods to do this that is common in the industry?
Thanks,
Brian -
Hi Brian,
Google can definitely crawl Javascript, although it is limited in a sense. Here is a great article addressing this.
I have seen many websites use the solution that you mention in your second paragraph. That is, creating a separate section of the domain (or creating a subdomain) for users who are logging in. GoogleBot will not be able to crawl anything that is beyond a login portal, so this is the most common solution. Of course you can also disallow the specific path or subdomain in robots.txt as well!
It comes down to how much content that you have behind the login portal. You want to figure out which part of your site you want to be crawled, and which parts you don't want. After that point it's a matter of figuring out the best solution for blocking that content from crawlers.
Hope that helps!
-
Hi Andy,
Thanks for the reply.
This documentation was helpful. You mentioned the potential for the content to be seen in the cached version of pages. I read the Search Engine Land article there and shes sort of getting at how you don't want to block your content to search engines because then they can't be indexed and ranked. Let's say we don't really care about those parts of the pages that we wan't to block in terms of being indexed, which I sort of mentioned above. If we were to some how display those parts of the pages, images and text, in JavaScript, search engines wouldn't be able to crawl. Would those portions of the page still be visible in the cached version? I would imagine not if the search engines can't crawl JS, but not sure.
A second option I thought of was to create a separate subdomain for logged in users that would only be accessible through a log in form. Search engines cannot follow forms, so I would imagine, that gated content wouldn't be visible to crawlers and therefore not indexed. You could go one step farther and block the subdomain in robots.txt. Do you think that could be a possible solution?
Thanks again for the reply last week.
-Brian
-
Hi Brian,
I have a client working on correcting this issue with his site at the moment. They run a big media site that allows access once paid, but so may of these sites suffer with the same issue and because they allow Google to index the whole text, but only show a portion of it, this means that if you just look at the cached version, you can read it without paying.
In terms of correcting it, I would first have a read on how Google handles subscription sites. You can find that info here. Google prefers the "First click free" model.
There is additional reading on this subject over at Search Engine Land. First Click Free (FCF) is what you want to be looking into in more detail.
I hope this helps a little.
-Andy
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Mega Menus and SEO
Hi Everyone, I know this has been brought up before, but wanted your opinion for 2020. I have a new client that is hesitant to do a mega menu for their huge site due to the amount of links and "dilution". I have quite a few clients with mega menus with no problems at all from an SEO standpoint. But I can understand his perspective. I am suggesting that we have the main links (looking at GA) as the the navigation, then clicking them takes you to subcategory page listing all the subcats within. Problem is that the developer/designer has made this mega menu already and it is pretty slick. Now they already are killing it search-wise on Google, but don't have a mega menu or a secondary category page. Just a a category with too many products, so we are trying to go one way or the other. Any opinions on which route to best take from a user and SEO perspective?
Web Design | | vetofunk0 -
How do elements that are displayed when scrolled impact SEO?
Hi, We are wanting to implement Animate.css and Wowjs on our site and were concerned about the SEO impacts. Basically when the page is loaded, if the element is not within the viewport then the HTML tag (i.e. div tag) have a style="visibility: hidden" and once the element is within the viewport it will change to have style="visibility: visible". Would having the style="visibility: hidden" negatively impact SEO?
Web Design | | KendallHershey0 -
How is Single Page Application (SPA) bad for SEO
Hi guys. I am quite inspired of SPA technique. It's really amazing when all your interaction with the site is going on the fly and you don't see any page reloads. I've started implementing the site with this instruction and already found nice guys to make the design. The only downside of the using SPA which I can see **is the **SEO part. That's because the URL does not really change and different pages don't have their unique URL addresses.
Web Design | | Billy_gym
Actually they have, but it looks like: yoursite.com/#/products yoursite.com/#/prices yoursite.com/#/contact So all of them goes after # and being just anchors. For Google this mean all of these pages is just yoursite.com/ My question is what is really proven method to implement the URL structure in Single Page Application, so all the pages indexed by Google correctly (sorry I don't mention the other search engines because of market share). The other question, of course, is examples. It will be great to see real life site examples, better authority sites, which use SPA technique and well indexed by search engines.1 -
Does an age verification home page hurt SEO?
There's a microbrewery in our area that just launched its first website. It has the "verify your age" homepage (which is not really their homepage, but I don't know what it's called) before you can enter. It looks like this: http://angrychairbrewing.com/ Anyway, does this hurt them at all from a rankings standpoint? Also, assuming bots/spiders/ROGER can crawl sites like this, (which I think they would have to be able to do) how do they get around this verification? Thanks, Ruben
Web Design | | KempRugeLawGroup0 -
Too Many Outbound Links on the Home Page - Bad for SEO?
Hello Again Moz community, This is my last Q of the day: I have a LOT of outbound links on the home page of www.web3.ca Some are to clients projects, most are to other pages on the website. Can reducing this to the core pages have a positive impact on SEO? Thanks, Anton
Web Design | | Web3Marketing870 -
Having a second homepage for a site would affect my SEO?
Hello guys, One of our clients is planning to have a new landing page for any users hitting the site for the first time. (returning users will still see the current homepage based on cookies ... in other words, the site would technically have 2 home pages). According to this client, they are planning to do something like this: https://www.websitename.com/ (for returning visitors) https://www.websitename.com/newuser (for first time visitors) Our instinct is that is not great to have 2 home pages (that would affect the SEO campaign we are managing for this company) and we are not sure how to handle this. That's why we would appreciate your opinion regarding this topic: From an SEO perspective, do you think this is a good idea? If not, what would you guys do differentiate first-time visitors vs returning visitors without affecting SEO? Maybe just a pop-up? Thanks in advance for your help !
Web Design | | Robertnweil10 -
Script tags and seo
Hi, I have a page on my site with a google map embed, and a path drawn on the map. The path is made from a long string of coordinates. For ease I have the co-ordinates placed in a script tag at the foot of the page, amongst my javascript My question is, will this script tag hurt the seo for the page? I've read that inline js and 'data islands' can be bad, so I've been careful to keep it out of the main body of the page. Thanks, any help appreciated!
Web Design | | madegood0 -
Yes or No for Ampersand "&" in SEO URLs
Hi Mozzers I would like to know how crawlers see the ampersand (& or &) in your URLs and if Google frown upon this or not? As far as I know they purely recognise this as "and" is this correct and is there any best practice for implementing this, as I know a lot of people complained before about & in links and that it is better to use it as &, but this is not on links, this is on URLs. Reason for this is that we looking to move onto an ASP.Net MVC framework (any suggestions for a different framework are welcome, we still just planning out future development) and in order to make use of the filter options we have on our site we need a parameter to indicate the difference on a routing level (routing sends to controller, controller sends to model, model sends to controller and controller sends to view < this is pattern of a request that comes in on the framework we will be using). I already have -'s and /'s in the URLs (which is for my SEO structuring) so these syntax can't be used for identifying filters the user clicks or uses to define their search as it will create a complete mess in the system. Now we looking at & to say; OK, when a user lands on /accommodation and they selects De Kelders (which is a destination in our area) the page will be /accommodation/de-kelders on this page they can define their search further to say they are looking for 5 star accommodation and it should be close to the beach, this is where the routing needs some guidance and we looking to have it as follow: /accommodation/de-kelders/5-star&close-to-the-beach. Now, does the "&" get identified by search engines on a URL level as "and" and does this cause any issues with crawling or indexation or would it be best to look at another solution? Thanks, Chris Captivate
Web Design | | DROIDSTERS0