undefined
Skip to content
Moz logo Menu open Menu close
  • Products
    • Moz Pro
    • Moz Pro Home
    • Moz Local
    • Moz Local Home
    • STAT
    • Moz API
    • Moz API Home
    • Compare SEO Products
    • Moz Data
  • Free SEO Tools
    • Domain Analysis
    • Keyword Explorer
    • Link Explorer
    • Competitive Research
    • MozBar
    • More Free SEO Tools
  • Learn SEO
    • Beginner's Guide to SEO
    • SEO Learning Center
    • Moz Academy
    • MozCon
    • Webinars, Whitepapers, & Guides
  • Blog
  • Why Moz
    • Digital Marketers
    • Agency Solutions
    • Enterprise Solutions
    • Small Business Solutions
    • The Moz Story
    • New Releases
  • Log in
  • Log out
  • Products
    • Moz Pro

      Your all-in-one suite of SEO essentials.

    • Moz Local

      Raise your local SEO visibility with complete local SEO management.

    • STAT

      SERP tracking and analytics for enterprise SEO experts.

    • Moz API

      Power your SEO with our index of over 44 trillion links.

    • Compare SEO Products

      See which Moz SEO solution best meets your business needs.

    • Moz Data

      Power your SEO strategy & AI models with custom data solutions.

    Let your business shine with Listings AI
    Moz Local

    Let your business shine with Listings AI

    Learn more
  • Free SEO Tools
    • Domain Analysis

      Get top competitive SEO metrics like DA, top pages and more.

    • Keyword Explorer

      Find traffic-driving keywords with our 1.25 billion+ keyword index.

    • Link Explorer

      Explore over 40 trillion links for powerful backlink data.

    • Competitive Research

      Uncover valuable insights on your organic search competitors.

    • MozBar

      See top SEO metrics for free as you browse the web.

    • More Free SEO Tools

      Explore all the free SEO tools Moz has to offer.

    NEW Keyword Suggestions by Topic
    Moz Pro

    NEW Keyword Suggestions by Topic

    Learn more
  • Learn SEO
    • Beginner's Guide to SEO

      The #1 most popular introduction to SEO, trusted by millions.

    • SEO Learning Center

      Broaden your knowledge with SEO resources for all skill levels.

    • On-Demand Webinars

      Learn modern SEO best practices from industry experts.

    • How-To Guides

      Step-by-step guides to search success from the authority on SEO.

    • Moz Academy

      Upskill and get certified with on-demand courses & certifications.

    • MozCon

      Save on Early Bird tickets and join us in London or New York City

    Unlock flexible pricing & new endpoints
    Moz API

    Unlock flexible pricing & new endpoints

    Find your plan
  • Blog
  • Why Moz
    • Digital Marketers

      Simplify SEO tasks to save time and grow your traffic.

    • Small Business Solutions

      Uncover insights to make smarter marketing decisions in less time.

    • Agency Solutions

      Earn & keep valuable clients with unparalleled data & insights.

    • Enterprise Solutions

      Gain a competitive edge in the ever-changing world of search.

    • The Moz Story

      Moz was the first & remains the most trusted SEO company.

    • New Releases

      Get the scoop on the latest and greatest from Moz.

    Surface actionable competitive intel
    New Feature

    Surface actionable competitive intel

    Learn More
  • Log in
    • Moz Pro
    • Moz Local
    • Moz Local Dashboard
    • Moz API
    • Moz API Dashboard
    • Moz Academy
  • Avatar
    • Moz Home
    • Notifications
    • Account & Billing
    • Manage Users
    • Community Profile
    • My Q&A
    • My Videos
    • Log Out

The Moz Q&A Forum

  • Forum
  • Questions
  • Users
  • Ask the Community

Welcome to the Q&A Forum

Browse the forum for helpful insights and fresh discussions about all things SEO.

  1. Home
  2. SEO Tactics
  3. Technical SEO
  4. How to block "print" pages from indexing

Moz Q&A is closed.

After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.

How to block "print" pages from indexing

Technical SEO
5
23
9.8k
Loading More Posts
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as question
Log in to reply
This topic has been deleted. Only users with question management privileges can see it.
  • dreadmichael
    dreadmichael last edited by Mar 16, 2012, 6:39 PM

    I have a fairly large FAQ section and every article has a "print" button. Unfortunately, this is creating a page for every article which is muddying up the index - especially on my own site using Google Custom Search.

    Can you recommend a way to block this from happening?

    Example Article:

    http://www.knottyboy.com/lore/idx.php/11/183/Maintenance-of-Mature-Locks-6-months-/article/How-do-I-get-sand-out-of-my-dreads.html

    Example "Print" page:

    http://www.knottyboy.com/lore/article.php?id=052&action=print

    1 Reply Last reply Reply Quote 0
    • NakulGoyal
      NakulGoyal @dreadmichael last edited by Mar 16, 2012, 8:32 PM Mar 16, 2012, 8:32 PM

      Donnie, I agree. However, we had the same problem on a website and here's what we did the canonical tag:

      Over a period of 3-4 weeks, all those print pages disappeared from the SERP. Now if I take a print URL and do a cache: for that page, it shows me the web version of that page.

      So yes, I agree the question was about blocking the pages from getting indexed. There's no real recipe here, it's about getting the right solution. Before canonical tag, robots.txt was the only solution. But now with canonical there (provided one has the time and resources available to implement it vs adding one line of text to robots.txt), you can technically 301 the pages and not have to stop/restrict the spiders from crawling them.

      Absolutely no offence to your solution in any way. Both are indeed workable solutions. The best part is that your robots.txt solution takes 30 seconds to implement since you provided the actually disallow code :), so it's better.

      1 Reply Last reply Reply Quote 0
      • dreadmichael
        dreadmichael @SEODinosaur last edited by Mar 16, 2012, 8:30 PM Mar 16, 2012, 8:30 PM

        Thanks Jennifer, will do! So much good information.

        1 Reply Last reply Reply Quote 0
        • Dr-Pete
          Dr-Pete Staff @SEODinosaur last edited by Mar 16, 2012, 8:30 PM Mar 16, 2012, 8:30 PM

          Sorry, but I have to jump in - do NOT use all of those signals simultaneously. You'll make a mess, and they'll interfere with each other. You can try Robots.txt or NOINDEX on the page level - my experience suggests NOINDEX is much more effective.

          Also, do not nofollow the links yet - you'll block the crawl, and then the page-level cues (like NOINDEX) won't work. You can nofollow later. This is a common mistake and it will keep your fixes from working.

          1 Reply Last reply Reply Quote 1
          • jennita
            jennita @SEODinosaur last edited by Mar 16, 2012, 8:28 PM Mar 16, 2012, 8:28 PM

            Josh, please read my and Dr. Pete's comments below. Don't nofollow the links, but do use the meta noindex,follow on the page.

            1 Reply Last reply Reply Quote 0
            • Dr-Pete
              Dr-Pete Staff @SEODinosaur last edited by Mar 16, 2012, 8:26 PM Mar 16, 2012, 8:26 PM

              Rel-canonical, in practice, does essentially de-index the non-canonical version. Technically, it's not a de-indexation method, but it works that way.

              1 Reply Last reply Reply Quote 0
              • dreadmichael
                dreadmichael @SEODinosaur last edited by Mar 16, 2012, 8:21 PM Mar 16, 2012, 8:21 PM

                You are right Donnie. I've "good answered" you too.

                I've gone ahead and updated my robots.txt file. As soon as I am able, I will use no indexon the page, no follow on the links, and rel=canonical.

                This is just what I needed, a quick fix until I can make a more permanent solution.

                1 Reply Last reply Reply Quote 0
                • SEODinosaur
                  SEODinosaur @dreadmichael last edited by Mar 16, 2012, 8:16 PM Mar 16, 2012, 8:16 PM

                  Your welcome : )

                  1 Reply Last reply Reply Quote 0
                  • SEODinosaur
                    SEODinosaur @SEODinosaur last edited by Mar 16, 2012, 8:16 PM Mar 16, 2012, 8:16 PM

                    Although you are correct... there is still more then one way to skin a chicken.

                    1 Reply Last reply Reply Quote 0
                    • SEODinosaur
                      SEODinosaur @dreadmichael last edited by Mar 16, 2012, 8:15 PM Mar 16, 2012, 8:15 PM

                      But the spiders still run on the page and read the canonical link, however with the robot text the spiders will not.

                      1 Reply Last reply Reply Quote 0
                      • SEODinosaur
                        SEODinosaur @NakulGoyal last edited by Mar 16, 2012, 8:11 PM Mar 16, 2012, 8:11 PM

                        Yes, but Rel=Canonical does not block a page it only tells google which page to follow out of two pages.The question was how to block, not how to tell google which link to follow. I believe you gave credit to the wrong answer.

                        http://en.wikipedia.org/wiki/Canonical_link_element

                        This is not fair. lol

                        dreadmichael Dr-Pete jennita 5 Replies Last reply Mar 16, 2012, 8:30 PM Reply Quote 0
                        • Dr-Pete
                          Dr-Pete Staff @jennita last edited by Mar 16, 2012, 8:05 PM Mar 16, 2012, 8:05 PM

                          I have to agree with Jen - Robots.txt isn't great for getting indexed pages out. It's good for prevention, but tends to be unreliable as a cure. META NOINDEX is probably more reliable.

                          One trick - DON'T nofollow the print links, at least not yet. You need Google to crawl and read the NOINDEX tags. Once the ?print pages are de-indexed, you could nofollow the links, too.

                          1 Reply Last reply Reply Quote 0
                          • NakulGoyal
                            NakulGoyal @dreadmichael last edited by Mar 16, 2012, 8:03 PM Mar 16, 2012, 8:03 PM

                            Yes, it's strongly recommended. It should be fairly simple to populate this tag with the "full" URL of the article based on the article ID. This approach will not only help you get rid of the duplicate content issue, but a canonical tag essentially works like a 301 redirect. So from all search engine perspective you are 301'ing your print pages to the real web urls without redirecting the actual user's who are browsing the print pages if they need to.

                            1 Reply Last reply Reply Quote 0
                            • dreadmichael
                              dreadmichael @NakulGoyal last edited by Mar 16, 2012, 7:53 PM Mar 16, 2012, 7:53 PM

                              Ya it is actually really useful. Unfortunately they are out of business now - so I'm hacking it on my own.

                              I will take your advice. I've shamefully never used rel= canonical before - so now is a good time to start.

                              NakulGoyal SEODinosaur 3 Replies Last reply Mar 16, 2012, 8:32 PM Reply Quote 0
                              • jennita
                                jennita @SEODinosaur last edited by Mar 16, 2012, 7:51 PM Mar 16, 2012, 7:51 PM

                                True but using robots.txt does not keep them out of the index. Only using "noindex" will do that.

                                1 Reply Last reply Reply Quote 1
                                • dreadmichael
                                  dreadmichael last edited by Mar 16, 2012, 7:50 PM Mar 16, 2012, 7:50 PM

                                  Thanks Donnie. Much appreciated!

                                  SEODinosaur 1 Reply Last reply Mar 16, 2012, 8:16 PM Reply Quote 1
                                  • NakulGoyal
                                    NakulGoyal last edited by Mar 16, 2012, 7:51 PM Mar 16, 2012, 7:50 PM

                                    I actually remember Lore from a while ago. It's an interesting, easy to use FAQ CMS.

                                    Anyways, I would also recommend implementing Canonical Tags for any possible duplicate content issues. So whether it's the print or the web version, each one of them will contain a canonical tag pointing to the web url of that article in the section of your website.

                                    rel="canonical" href="http://www.knottyboy.com/lore/idx.php/11/183/Maintenance-of-Mature-Locks-6-months-/article/How-do-I-get-sand-out-of-my-dreads.html" />
                                    dreadmichael SEODinosaur 2 Replies Last reply Mar 16, 2012, 8:11 PM Reply Quote 1
                                    • SEODinosaur
                                      SEODinosaur @dreadmichael last edited by Mar 16, 2012, 7:50 PM Mar 16, 2012, 7:39 PM

                                      http://www.seomoz.org/learn-seo/robotstxt

                                      1 Reply Last reply Reply Quote 1
                                      • SEODinosaur
                                        SEODinosaur @dreadmichael last edited by Mar 16, 2012, 7:38 PM Mar 16, 2012, 7:38 PM

                                        Try This.

                                        User-agent: *

                                        Disallow: /*&action=print

                                        1 Reply Last reply Reply Quote 0
                                        • SEODinosaur
                                          SEODinosaur @jennita last edited by Mar 16, 2012, 7:31 PM Mar 16, 2012, 7:31 PM

                                          Theres more then one way to skin a chicken.

                                          jennita SEODinosaur 2 Replies Last reply Mar 16, 2012, 8:16 PM Reply Quote 0
                                          • jennita
                                            jennita last edited by Mar 16, 2012, 7:02 PM Mar 16, 2012, 7:02 PM

                                            Rather than using robots.txt I'd use a noindex,follow tag instead to the page. This code goes into the tag for each print page. And it will ensure that the pages don't get indexed but that the links are followed.

                                            SEODinosaur Dr-Pete 2 Replies Last reply Mar 16, 2012, 8:05 PM Reply Quote 1
                                            • dreadmichael
                                              dreadmichael @SEODinosaur last edited by Mar 16, 2012, 6:59 PM Mar 16, 2012, 6:59 PM

                                              That would be great. Do you mind giving me an example?

                                              SEODinosaur 2 Replies Last reply Mar 16, 2012, 7:39 PM Reply Quote 0
                                              • SEODinosaur
                                                SEODinosaur last edited by Mar 16, 2012, 6:54 PM Mar 16, 2012, 6:53 PM

                                                you can block in .robot text, every page that ends in action=print

                                                dreadmichael 1 Reply Last reply Mar 16, 2012, 6:59 PM Reply Quote 0
                                                • 1 / 1
                                                1 out of 23
                                                • First post
                                                  1/23
                                                  Last post

                                                Got a burning SEO question?

                                                Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.


                                                Start my free trial


                                                Browse Questions

                                                Explore more categories

                                                • Moz Tools

                                                  Chat with the community about the Moz tools.

                                                • SEO Tactics

                                                  Discuss the SEO process with fellow marketers

                                                • Community

                                                  Discuss industry events, jobs, and news!

                                                • Digital Marketing

                                                  Chat about tactics outside of SEO

                                                • Research & Trends

                                                  Dive into research and trends in the search industry.

                                                • Support

                                                  Connect on product support and feature requests.

                                                • See all categories

                                                Related Questions

                                                • Vanderlindemedia

                                                  Page with "random" content

                                                  Hi, I'm creating a page of 300+ in the near future, on which the content basicly will be unique as it can be. However, upon every refresh, also coming from a search engine refferer, i want the actual content such as listing 12 business to be displayed random upon every hit. So basicly we got 300+ nearby pages with unique content, and the overview of those "listings" as i might say, are being displayed randomly. Ive build an extensive script and i disabled any caching for PHP files in specific these pages, it works. But what about google? The content of the pages will still be as it is, it is more of the listings that are shuffled randomly to give every business listing a fair shot at a click and so on. Anyone experience with this? Ive tried a few things in the past, like a "Last update PHP Month" in the title which sometimes is'nt picked up very well.

                                                  Technical SEO | Jul 20, 2020, 12:02 PM | Vanderlindemedia
                                                  0
                                                • SpodekandCo

                                                  Is there a way to get a list of all pages of your website that are indexed in Google?

                                                  I am trying to put together a comprehensive list of all pages that are indexed in Google and have differing opinions on how to do this.

                                                  Technical SEO | Oct 31, 2019, 8:15 AM | SpodekandCo
                                                  0
                                                • marchexmarketingmcc

                                                  Trying to find all internal links to a specific page (without index)

                                                  Hi guys -- Still waiting on Moz to index a page of mine. We launched a new site over two months ago. In the meantime, I really just need a list of internal links to a specific page because I want to change its URL. Does anybody know how to find that list (of internal links to 1 of my pages) without the Moz index? I appreciate the help!

                                                  Technical SEO | Jan 28, 2023, 11:30 AM | marchexmarketingmcc
                                                  1
                                                • RickyShockley

                                                  Home Page Ranking Instead of Service Pages

                                                  Hi everyone! I've noticed that many of our clients have pages addressing specific queries related to specific services on their websites, but that the Home Page is increasingly showing as the "ranking" page. For example, a plastic surgeon we work with has a page specifically talking about his breast augmentation procedure for Miami, FL but instead of THAT page showing in the search results, Google is using his home page. Noticing this across the board. Any insights? Should we still be optimizing these specific service pages? Should I be spending time trying to make sure Google ranks the page specifically addressing that query because it SHOULD perform better? Thanks for the help. Confused SEO :/, Ricky Shockley

                                                  Technical SEO | Feb 24, 2016, 5:02 AM | RickyShockley
                                                  0
                                                • Iamfaramon

                                                  How to stop google from indexing specific sections of a page?

                                                  I'm currently trying to find a way to stop googlebot from indexing specific areas of a page, long ago Yahoo search created this tag class=”robots-nocontent” and I'm trying to see if there is a similar manner for google or if they have adopted the same tag? Any help would be much appreciated.

                                                  Technical SEO | Sep 26, 2014, 6:46 PM | Iamfaramon
                                                  0
                                                • jim_shook

                                                  Best way to handle pages with iframes that I don't want indexed? Noindex in the header?

                                                  I am doing a bit of SEO work for a friend, and the situation is the following: The site is a place to discuss articles on the web. When clicking on a link that has been posted, it sends the user to a URL on the main site that is URL.com/article/view. This page has a large iframe that contains the article itself, and a small bar at the top containing the article with various links to get back to the original site. I'd like to make sure that the comment pages (URL.com/article) are indexed instead of all of the URL.com/article/view pages, which won't really do much for SEO. However, all of these pages are indexed. What would be the best approach to make sure the iframe pages aren't indexed? My intuition is to just have a "noindex" in the header of those pages, and just make sure that the conversation pages themselves are properly linked throughout the site, so that they get indexed properly. Does this seem right? Thanks for the help...

                                                  Technical SEO | Jun 26, 2013, 6:35 PM | jim_shook
                                                  0
                                                • ProjectLabs

                                                  Determining When to Break a Page Into Multiple Pages?

                                                  Suppose you have a page on your site that is a couple thousand words long. How would you determine when to split the page into two and are there any SEO advantages to doing this like being more focused on a specific topic. I noticed the Beginner's Guide to SEO is split into several pages, although it would concentrate the link juice if it was all on one page. Suppose you have a lot of comments. Is it better to move comments to a second page at a certain point? Sometimes the comments are not super focused on the topic of the page compared to the main text.

                                                  Technical SEO | Oct 18, 2012, 9:44 PM | ProjectLabs
                                                  1
                                                • fthead9

                                                  What is the best method to block a sub-domain, e.g. staging.domain.com/ from getting indexed?

                                                  Now that Google considers subdomains as part of the TLD I'm a little leery of testing robots.txt with something like: staging.domain.com
                                                  User-agent: *
                                                  Disallow: / in fear it might get the www.domain.com blocked as well. Has anyone had any success using robots.txt to block sub-domains? I know I could add a meta robots tag to the staging.domain.com pages but that would require a lot more work.

                                                  Technical SEO | Oct 6, 2011, 10:55 PM | fthead9
                                                  0

                                                Get started with Moz Pro!

                                                Unlock the power of advanced SEO tools and data-driven insights.

                                                Start my free trial
                                                Products
                                                • Moz Pro
                                                • Moz Local
                                                • Moz API
                                                • Moz Data
                                                • STAT
                                                • Product Updates
                                                Moz Solutions
                                                • SMB Solutions
                                                • Agency Solutions
                                                • Enterprise Solutions
                                                Free SEO Tools
                                                • Domain Authority Checker
                                                • Link Explorer
                                                • Keyword Explorer
                                                • Competitive Research
                                                • Brand Authority Checker
                                                • Local Citation Checker
                                                • MozBar Extension
                                                • MozCast
                                                Resources
                                                • Blog
                                                • SEO Learning Center
                                                • Help Hub
                                                • Beginner's Guide to SEO
                                                • How-to Guides
                                                • Moz Academy
                                                • API Docs
                                                About Moz
                                                • About
                                                • Team
                                                • Careers
                                                • Contact
                                                Why Moz
                                                • Case Studies
                                                • Testimonials
                                                Get Involved
                                                • Become an Affiliate
                                                • MozCon
                                                • Webinars
                                                • Practical Marketer Series
                                                • MozPod
                                                Connect with us

                                                Contact the Help team

                                                Join our newsletter
                                                Moz logo
                                                © 2021 - 2025 SEOMoz, Inc., a Ziff Davis company. All rights reserved. Moz is a registered trademark of SEOMoz, Inc.
                                                • Accessibility
                                                • Terms of Use
                                                • Privacy

                                                Looks like your connection to Moz was lost, please wait while we try to reconnect.