Prevent indexing of dynamic content
-
Hi folks!
I discovered bit of an issue with a client's site. Primarily, the site consists of static html pages, however, within one page (a car photo gallery), a line of php coding:
dynamically generates a 100 or so pages comprising the photo gallery - all with the same page title and meta description. The photo gallery script resides in the /gallery folder, which I attempted to block via robots.txt - to no avail. My next step will be to include a:
within the head section of the html page, but I am wondering if this will stop the bots dead in their tracks or will they still be able to pick-up on the pages generated by the call to the php script residing a bit further down on the page?
Dino
-
Hello Steven,
Thank you for providing another perspective. However, all factors considered, I agree with Shane's approach on this one. The pages add very little merit to the site and exist primarily to provide the site users with eye-candy (e.g. photos of classic cars).
-
Just personally, I would still deindex or canonical them - they are just pages with a few images - so not of much value and unless all titles and descriptions are targeting varying keywords and content is added, they will canabalize eachother, and possibly even drag down the site due to 100's of pages of thin content....
So actually from an SEO perspective it probably IS better to deindex or canonical 3 - 5 or so years ago, maybe the advice would have been keep them and keyword target - but not in the age of content
(unless the images were optimized for image searches for sale able products (but I do not think it is)
-
Hi Dino,
I know this won't solve the immediate problem you asked for, but wouldn't it be better for your client's site (and for SEO) to alter the PHP so that the title and meta data description are replaced with variables that can also be dynamic, depending on whichever of the 100 or so pages gets created?
That way, rather than worrying about a robot seeing 100 pages as duplicate content, it could see 100 pages as 100 pages.
-
It depends on how the pages are being created (I would assume it is off of a template page)
So within the template of this dynamically created page you would place
But if this is the global template - you cannot do this as it will noindex every page which of course is bad.
If you want to PM me the URL of the page I can take a look at your code, and see what is going on and how to recitify, as right now i think we are talking about the same principles, but different words are being used.
It really is pretty straightforward. (what I am saying) The pages that you want to be not indexed DO NOT need a nofollow they need a meta noindex
But there are many variables, as if you have already robot.txt disallowed the directory, then no bot will go there to get the updated noindex directive....
If there is no way to add a meta noindex then you need to nofollow and put in for a manual removal
-
I completely understand and agree with all points you have conveyed. However, I am not certain as to the best approach to "noindex" the urls which are being created dynamically from within the static html page? Maybe I am making this more complex than it needs to be...
-
So it is the pages themselves that are dynamically created you want out of index, not the page the contains the links?
If this is so ---
noindex the pages that are created dynamically
Therein lies the problem. I did have the nofollow directive in place specifying the /gallery/ folder, but apparently, the bots still crawled it.
Nofollow does not remove from index, it only tells the bot not to pass authority, as it is still feasible that the bot will crawl the link, so without the noindex, nofollow is not the correct directive due to the page (even though nofollowed) is still being reached and indexed.
PS. also if you have the nofollow on the links, you may want to remove it, so the bots will go straight through to the page and grab the noindex directive, but if you wanted to try to not let any authority "evaporate" you can continue to nofollow, but you may need to request the dynamically generated pages (URLS) be removed using webmaster tools.
-
The goal is to have the page remain in the index, but not follow any dynamically generated links on the page. The nofollow directive (in place for months) has not done the job.
-
?
If a link is coming into the page, and you have Noindex, Nofollow - this would remove from index and prevent the following of any links -
This is NOT instant, and can take months to occur depending on depth of page, crawl schedule ect... (you can try to speed it up by using webmaster tools to remove the URL)
What is the goal You are attempting to achieve?
To get the page out of index, but still followed?
Or remain in index, but just not follow links on page?
?
-
Therein lies the problem. I did have the nofollow directive in place specifying the /gallery/ folder, but apparently, the bots still crawled it. I agree that the noindex removes the page, but I wasn't certain if it prevented crawling of the page, as I have read mixed opinions on this.
I just thought of something else... perhaps an external url is linking to this page - allowing it to be crawled. I am off to examine this perspective.
Thanks for your response!
-
noindex will only remove from Index and dissallow the act of indexing the specific page (or pages created off template) you place the tag in upon the next page crawl.
Bots will still follow the page, and follow any links that are readable as long as there is not a nofollow directive.
I am not sure I fully understand the situation, so I would not say this is my "reccomendation" but an answer to the specific question.....
but I am wondering if this will stop the bots dead in their tracks or will they still be able to pick-up on the pages generated
Hope this helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content - Pricing Plan tables
Hey guys, We're faced with a problem that we want to solve. We're working on the designs for a few pages for a drag & drop email builder we're currently working on, and we will be having the same pricing table on several pages (much like Moz does). We're worried that Google will take this as duplicate content and not be very fond of it. Any ideas about how we could integrate the same flow without potentially harming ranking efforts? And NO, re-writing the content for each table is not an option. It would do nothing but confuse the heck out of our clients. 😄 Thanks everybody!
On-Page Optimization | | andy.bigbangthemes0 -
Duplicate Page Content
Hi, I am new to the MOZ Pro community. I got the below message for many of my pages. We have a video site so all content in the page except the video link would be different. How can i handle such pages. Can we place adsense AD's on these pages? Duplicate Page Content Code and content on this page looks similar or identical to code and content on other pages on your site. Search engines may not know which pages are best to include in their index and rankings. Common fixes for this issue include 301 redirects, using the rel=canonical tag, and using the Parameter handling tool in Google Webmaster Central. For more information on duplicate content, visit http://moz.com/learn/seo/duplicate-content. Please help me to know how to handle this.. Regards
On-Page Optimization | | Nettv0 -
Duplicate Content on our own website
Our website sells tickets for events. We also have an news articles section with information about events / artists / venues. From time to time we release a product page and a related news article on a separate page. Some of the content in the news article would be perfect for our product page. Essentially its our product page we want too rank. Would it harm our SEO if we had some of the same content on both of these pages?
On-Page Optimization | | Alexogilvie0 -
Add content as blog post or to product pages?
Hi, We have around 40 products which we can produce plenty of in-depth and detailed "how to"-type pieces of content for. Our current plan is to produce a "How to make" style post for each as a long blog post, then link that to the product page. There's probably half a dozen or more of these kind of blog posts that we could do for each product. The reason why we planned on doing it like this is that it would give us plenty of extra pages (blog posts) on their own URL which can be indexed and rank for long tail keywords, but also that we can mention these posts in our newsletter. It'd give people a new page full of specific content that they can read instead of us having to say "Hey! We've updated our product page for X!", which seems a little pointless. Most of the products we sell don't get very many searches themselves; Most get a couple dozen and the odd few get 100-300 each, while one gets more than 2,000 per month. The products don't get many searches as it's a relatively unknown niche when it comes to details, but searches for the "categories" these products are in are very well known (Some broad terms that cover the niche get more than 30,000+ searches a month in the UK and 100,000+ world wide) [Exact].
On-Page Optimization | | azu25
Regarding the one product with more than 2,000 searches; This keyword is both the name of the product and also a name for the category page. Many of our competitors have just one of these products, whereas we're one of the first to have more than 6 variations of this product, thus the category page is acting like our other product pages and the information you would usually find on our product pages, is on the category page for just this product. I'm still leaning towards creating each piece of content as it's own blog post which links to the product pages, while the product pages link to the relevant blog posts, but i'm starting to think that it may be be better to put all the content on the product pages themselves). The only problem with this is that it cuts out on more than 200 very indepth and long blog posts (which due to the amount of content, videos and potentially dozens of high resolution images may slow down the loading of the product pages). From what I can see, here are the pros and cons: Pro (For blog posts):
1. More than 200 blog posts (potentially 1000+ words each with dozens of photos and potentially a video)..
2. More pages to crawl, index and rank..
3. More pages to post on social media..
4. Able to comment about the posts in the newsletter - Sounds more unique than "We've just updated this product page"..
5. Commenting is available on blog posts, whereas it is not on product pages..
6. So much information could slow down the loading of product pages significantly..
7. Some products are very similar (ie, the same product but "better quality" - Difficult to explain without giving the niche away, which i'd prefer not to do ATM) and this would mean the same content isn't on multiple pages.
8. By my understanding, this would be better for Google Authorship/Publishership.. Con (Against blog posts. For extended product pages):
1. Customers have all information in one place and don't have to click on a "Related Blog posts" tab..
2. More content means better ability to rank for product related keywords (All but a few receive very few searches per month, but the niche is exploding at an amazing rate at the moment)..
3. Very little chance of a blog post out-ranking the related product page for keywords.. I've run out of ideas for the 'Con' side of things, but that's why I'd like opinions from someone here if possible. I'd really appreciate any and all input, Thanks! [EDIT]:
I should add that there will be a small "How to make" style section on product pages anyway, which covers the most common step by step instructions. In the content we planned for blog posts, we'd explore the regular method in greater detail and several other methods in good detail. Our products can be "made" in several different ways which each result in a unique end result (some people may prefer it one way than another, so we want to cover every possible method), effectively meaning that there's an almost unlimited amount of content we could write.
In fact, you could probably think of the blog posts as more of "an ultimate guide to X" instead of simply "How to X"...0 -
E-Commerce Site - Duplicate Content
We run an e-commerce site with about 250,000 SKUs. Certain items, such as a micro USB car charger, will be applicable to several different phones. Example: http://www.wirelessemporium.com/p-165787-samsung-galaxy-proclaim-illusion-sch-i110-heavy-duty-car-charger.asp http://www.wirelessemporium.com/p-165856-sony-xperia-ion-4g-lte-att-heavy-duty-car-charger.asp As one can imagine with so many items, unique content for each item description page can be a challenge. What would be the best way to address this on a large scale?
On-Page Optimization | | eugeneku0 -
Archive of content
Hi there, I have recently joined a company to look after the e-marketing side of things, anyway the company I work for have been writing articles for a website that they own for over 2 years, probably about 200 or so unique articles on that website, however over the past year or so there has been no contribution to this site and was wondering if it would be worthwhile transferring these article over to our blog?, as this is where all the attention is in terms of marketing etc Kind Regards,
On-Page Optimization | | Paul780 -
Getting pages indexed by Google
Hi SEOMoz, I relaunched a site back in February of this year (www.uniquip.com) with about 1 million URL's. Right now I'm seeing that Google is not going past 110k indexed URL's (based on sitemaps). Do you have any tips on what I can do to make the site more likeable by Google and get more indexed URL's? All the the part pages can be browsed to by going to: http://www.uniquip.com/product-line-card/suppliers/sw-a/p-1 I've tried to make the content as unique as possible by adding random testimonials and random "related part numbers" see here: http://www.uniquip.com/id/246172/electronic-components/infineon/microcontrollers-mcu/sabc161pilfca Do I need to wait more time and be more patient with Google? It just seems like I'm only getting a few thousand URL's per day at the most. Would it help me if I implemented a breadcrumb on all part pages? Thanks, -Carlos
On-Page Optimization | | caneja0 -
What to do with old content in light of the Panda update?
Let's say you operate a laptop review website. After several years, the individual product review URL's (like site.com/dell/xp1234-review/) aren't receiving much traffic, they may have a few links here and there. In general and considering the panda update, would the best option be to 301 the old URL's back to the category page (site.com/dell/)or just keep them where they are? Any potential issues like having excessive 301's which could slow down the site or appear fishy to search engines?
On-Page Optimization | | BryanPhelps-BigLeapWeb0