Just read Travis Loncar's YouMoz post and I have a question about Pagination
-
This was a brilliant post.
I have a question about Pagination on sites that are opting to use Google Custom Search. Here is an example of a search results page from one of the sites I work on:
http://www.ccisolutions.com/StoreFront/category/search-return?q=countryman
I notice in the source code of sequential pages that the rel="next" and rel="prev" tags are not used. I also noticed that the URL does not change when clicking on the numbers for the subsequent pages of the search results.
Also, the canonical tag of every subsequent page looks like this:
Are you thinking what I'm thinking? All of our Google Custom Search pages have the same canonical tag....Something's telling me this just can't be good.
Questions:
1. Is this creating a duplicate content issue?
2. If we need to include rel="prev" and rel="next" on Google Custom Search pages as well as make the canonical tag accurate, what is the best way to implement this?
Given that searchers type in such a huge range of search terms, it seems that the canonical tags would have to be somehow dynamically generated.
Or, (best case scenario!) am I completely over-thinking this and it just doesn't matter on dynamically driven search results pages?
Thanks in advance for any comments, help, etc.
-
-
Considering that the larger of the two sites I work on is on a platform from 1996, I might actually be living "back in the day!" lol - Thanks again Jared!
-
This would all depend on what the site was built on, and the flexibility. There's no questions that this can be done. "Back in the day" we had a few sites that had tens of thousands of page due to sorting, and we had everything generated including:
Title, meta d, meta k, breadcrmb, H1 and short description.
Those were the days!!!
-
For the most part, I would choose to use rel=prev/next for pagination, including both pagination with dynamic urls and static URLs. There are some cases (as with this original thread question) where you should use canonical, but as a whole you should use rel=prev/next.
The best way to explain it is:
Rel Prev/Next:
Your site: Hi Google, I have all of these pages that very similar so I'm just letting you know that I only have duplicate content here for usability reasons and am in no way inferring that you should index all of these pages and rank them #1!
Google: Ok great, thanks for letting us know. We'll index the pages we feel are appropriate, but you wont get penalized for duplicate content. We may only index and serve one page, "page 1", or we may index multiple pages. Thanks for letting us know.
Canonical:
Your site: Hi Google, I have all these paginated pages that look like duplicate content, please do not include any of them in your index, and don't penalize me for duplicate content. For the record, the page you should index is Page 1 and no other pages.Any links that point to the paginated pages should be counted towards Page 1*.
Google: Great, no matter what we will not index any pagination and only Page 1.
With rel=next you are simply letting Google know, but not dictating how Google should act on the situation. If fact with ecomm sites, youll find that a lot of timees when you use rel=next, Google will actually index the 'view all' page if you have "view all" as an option around your pagination links
*many articles suggest that link juice is passed to the canonical URL - I'm have not seen any direct evidence of this but is worth a different discussion.
-
Yes, Jared, this is a great answer. I understand completed. It looks like we are ok then with Google Custom Search as it is. Thanks so much for your thoughtful answer. Now, if we can only get our paginated category pages sorted out, we'll be on the right track!
-
Hi Gerd,
Yes, this is a separate issue we are also struggling with on the site. I believe Travis' YouMoz post from yesterday made a pretty good case for using multiple paginated URLs, and he even illustrated how to accomplish this with sorting parameters like "color" and "price"
You raise a very good point about duplicate titles and descriptions potentially being a problem in this scenario.
Does anyone have any ideas about how to handle that? Could the backend be programmed to dynamically create unique titles and descriiptions based on some rules for naming conventions? (assuming you have access to that level of the code of course)
Really interested to know some points of view on this!
Dana
-
I raised a similar question in the following Q&A - http://www.seomoz.org/q/duplicate-title-tags-with-pagination-and-canonical
My concern or question (we have rel=prev/next) would be more towards what the canoncial should be. There seems to be different opinions:
1. Use the current paginated page as the canonical - in our case GWMT reports duplicate titles (I suppose appending a page-number should sort this out)
2. Use the base search URL as the canonical - perhaps not a bad choice if your site's content changes and Google indexes page 50, but over time you only have results for 40 pages (resulting in an empty result page)
I currently only can conclude that having the prev/next implemented is a good thing as it will hint Google in pagination (in addition to setup the URL parameters in GWMT). I do plan to change the canoncial to the base search URL (and not having multiple paginated URLs) and see how this will affect indexing and SERPs.
-
Dana
Great and informative question,
Jared
Great Answer
-
Hi Dana - Let me see if I understand this correctly:
In question 1 you asked if this would be a duplicate content issue. The canonical tag retains the exact same URL regardless of the search parameter (and resulting search results). Therefore, regardless of the search being made, Google and other crawlers will not index page with a search parameter since the canonical references to the original url (http://www.ccisolutions.com/StoreFront/category/search-return). This means that when Google accidentally lands here http://www.ccisolutions.com/StoreFront/category/search-return?q=countryman it sees the canonical tag and understands that it should not index this page as it is only a variation of the core page.
This would of course be a problem if you actually wanted Google to index every query page. Alternate methods could be to disclude the query parameter in WMT or Robots. But the canonical is built in for you so that you dont have to.
In situations like this I also like to add site search to analytics and block the query parameter so no query pages show up as landing pages.
-
I understand exactly what you are saying Jared. However, here's the problem, the canonical tag is exactly the same....for every single subsequent page in a series across the entire site.
No matter what is searched. The canonical tag remains:
Wouldn't that mean that all search results pages, regardless of search term, are viewed as the same page?
I have heard this discussed before come to think of it. In this case, wouldn't it be proper to block all dynamic search results pages from being crawled or indexed by Google via the htaccess file or robots.txt file?
-
Hi Dana -
I think in the case of Google Custom Search, there is no need to worry about duplication. The reason is that although the rel="prev" etc tags are not being used, a blanket solution already exists: the canonical tag. As you mentioned, the canonical tag never changes, regardless of the search - therefore the crawlers only ever see the Custom Search page as a single page regardless of the queries being made. Thus there is no duplicate issue.
-
I use Google custom search on my site and love it. I would say you have some valid concerns. At first it was a bit of a pain because some of the images didn't line up with the products after a few weeks it worked itself out. We had a 47% increase in conversion from using Google custom search, I use an out of the box type web service so I cannot help you with a few of the questions. There is a lot of customization you can do to fix that you described. Bringing our blog and recipe section was the purpose for trying it and the revenue proved it to be a wise decision.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
No: 'noindex' detected in 'robots' meta tag
Pages on my site show No: 'noindex' detected in 'robots' meta tag. However, when I inspect the pages html, it does not show noindex. In fact, it shows index, follow. Majority of pages show the error and are not indexed by Google...Not sure why this is happening. The page below in search console shows the error above...
Technical SEO | | Sean_White_Consult0 -
These days on Google results, it also shows the site map. I submitted my company's sitemap and it still does not show?What am I doing wrong?
Look at the image in the link. I want my company to look like the "pluralsight" website in Google. I want it to show the sitemap. I have already submitted the sitemap to Google few days back, what am I doing wrong? search?sourceid=chrome-psyapi2&ion=1&espv=2&ie=UTF-8&q=pluralsight&oq=pluralsight&aqs=chrome..69i57j0l5.11024j0j8
Technical SEO | | Deein0 -
Why are these URL's suddenly appearing in WMT?
One of our clients has suddenly experienced a sudden increase in crawl errors for smart phones overnight for pages which no longer exist and there are no links to these pages according to Google. There is no evidence as to why Google would suddenly start to crawl these pages as they have not existed for over 5 years, but it does come after a new site design has been put live. Pages do not appear to be in the index when a site search is used. There was a similar increase in crawl errors on desktop initially after the new site went live, but these quickly returned to normal. Mobile crawl errors only became apparent after this. There are some URL's showing which have no linking page detected so we don't know where these URL's are being found. WMT states "Googlebot couldn't crawl this URL because it points to a non-existent page". Those that do have a linking page are showing an internal page which also doesn't exist so it can't possibly link to any page. Any insight is appreciated. Andy and Mark at Click Consult.
Technical SEO | | ClickConsult0 -
Specific question about pagination prompted by Adam Audette's Presentation at RKG Summit
This question is prompted by something Adam Audette said in this excellent presentation: http://www.rimmkaufman.com/blog/top-5-seo-conundrums/08062012/ First, I will lay out the issues: 1. All of our paginated pages have the same URL. To view this in action, go here: http://www.ccisolutions.com/StoreFront/category/audio-technica , scroll down to the bottom of the page and click "Next" - look at the URL. The URL is: http://www.ccisolutions.com/StoreFront/IAFDispatcher, and for every page after it, the same URL. 2. All of the paginated pages with non-unique URLs have canonical tags referencing the first page of the paginated series. 3. http://www.ccisolutions.com/StoreFront/IAFDispatcher has been instructed to be neither crawled nor indexed by Google. Now, on to what Adam said in his presentation: At about minute 24 Adam begins talking about pagination. At about 27:48 in the video, he is discussing the first of three ways to properly deal with pagination issues. He says [I am somewhat paraphrasing]: "Pages 2-N should have self-referencing canonical tags - Pages 2-N should all have their own unique URLs, titles and meta descriptions...The key is, with this is you want deeper pages to get crawled and all the products on there to get crawled too. The problem that we see a lot is, say you have ten pages, each one using rel canonical pointing back to page 1, and when that happens, the products or items on those deep pages don't get get crawled...because the rel canonical tag is sort of like a 301 and basically says 'Okay, this page is actually that page.' All the items and products on this deeper page don't get the love." Before I get to my question, I'll just throw out there that we are planning to fix the pagination issue by opting for the "View All" method, which Adam suggests as the second of three options in this video, so that fix is coming. My question is this: It seems based on what Adam said (and our current abysmal state for pagination) that the products on our paginated pages aren't being crawled or indexed. However, our products are all indexed in Google. Is this because we are submitting a sitemap? Even so, are we missing out on internal linking (authority flow) and Google love because Googlebot is finding way more products in our sitemap that what it is seeing on the site? (or missing out in other ways?) We experience a lot of volatility in our rankings where we rank extremely well for a set of products for a long time, and then disappear. Then something else will rank well for a while, and disappear. I am wondering if this issue is a major contributing factor. Oh, and did I mention that our sort feature sorts the products and imposes that new order for all subsequent visitors? it works like this: If I go to that same Audio-Technica page, and sort the 125+ resulting products by price, they will sort by price...but not just for me, for anyone who subsequently visits that page...until someone else re-sorts it some other way. So if we merchandise the order to be XYZ, and a visitor comes and sorts it ZYX and then googlebot crawls, google would potentially see entirely different products on the first page of the series than the default order marketing intended to be presented there....sigh. Additional thoughts, comments, sympathy cards and flowers most welcome. 🙂 Thanks all!
Technical SEO | | danatanseo0 -
When do you use 'Fetch as a Google'' on Google Webmaster?
Hi, I was wondering when and how often do you use 'Fetch as a Google'' on Google Webmaster and do you submit individual pages or main URL only? I've googled it but i got confused more. I appreciate if you could help. Thanks
Technical SEO | | Rubix1 -
Webmaster Tools vs Screaming from for 404's
Hey guys, I was just wondering which is better to use to find the 404's effecting your site. I have been using webmaster tools and just purchased screaming frog which has given me a totally different list of 404's compared to WMT. Which do I use, or do I use both? Cheers
Technical SEO | | Adamshowbiz0 -
Feedburner - Why Sending My Blog Posts A Day After I Post Them?
I have my feed setup through feedburner for my wife's blog ktlouise.com. Whenever she posts a new blog post, it doesn't get emailed to her subscribers until the next day. Does anyone know how to change this so that the updates go out the same day? Thanks for the help! REF
Technical SEO | | FergusonSEO0 -
Schema Markup and Google's Rich Snippet Tool
Has anyone ever used the snippet tool and gotten the following error "could not fetch website"? When using the tool and placing an url that does not have markup present it will show that as the error. Or if part of markup is wrong, it will diagnose it accordingly. Did a search online and found limited info...one of which someone had this error but when other users tested it, they were not getting the same error.
Technical SEO | | andrewv0