Any penalty for having rel=canonical tags on every page?
-
For some reason every webpage of our website (www.nathosp.com) has a rel=canonical tag. I'm not sure why the previous SEO manager did this, but we don't have any duplicate content that would require a canonical tag.
Should I remove these tags? And if so, what's the advantage - or disadvantage of leaving them in place?
Thank you in advance for your help.
-Josh Fulfer
-
There isn't a direct penalty for having rel="canonical" tags on every page, no, as long as you are correctly utilizing them (i.e. don't set the href of the tag to an invalid or non-existent URL). If there is even the possibility of duplicate content on your website, it is best to use canonical tags.
For websites serving straight HTML files, both http://www.example.com/index.html and http://www.example.com/ likely serve the same content.
If you use a framework like ASP.NET MVC, it would by default return duplicate content for both http://www.example.com/ and http://www.example.com/Home/Index.
Choose one or the other and set your canonical tag to that:
(note: the trailing slash is optional - just be consistent with including it or not)
-
You can use a canonical tag on page A, to point to A, telling that this is the original, teh reason for this is when people scrape your site they will point back home.
i belive thats is what they were getting at
you would only point it at B if B was a duplicate.
-
Ryan - I appreciate your help. My initial thought too was that I could remove it to clean up the code. However, I was unaware that the tag helps with dynamically generated pages - which ours are.
Thank you for your thorough response.
-
as far as i can see josh, the canonical URLs on your site are doing what they should be doing. I havn't looked to deep into it, but it seems like your products all refer back to product category pages, so that is the right way to use them.
-
I have never heard of anyone being penalised for having it on every page. Plus I can't see that ever happening unless it has been implemented incorrectly of course.
-
page A has content about apples. page B has content about bubblegum. Canonical tag states that page B should refer to page A. What is the point of that? all link juice, all ranking potential is passed to page A, even though page B has very different content. So page A MIGHT appear in search results about bubblegum, but page B will not because it is passing all link juice and rank potential to page A about apples. People stop going to page A when looking for bubblegum because it is irrelevant, and bounce rates increase.
Dont think you need documentation to get this. If you have all pages redirecting bots via canonical urls to the SAME page, it is pointless. If you have several article about apples and point them all to page A that is a different story.
-
not sure what you mean here, I have a canonical on every page, I program my sites to dynamicly to do, the reason i do so, is if someone scraps a page, it will have my address in the canonical tag.
I dont know what you mean by not relative to the tag. it just a href, are we talking about the same thing?
rel="canonical" href=http://mydomain.com/>
-
Having canonical tags on pages that don't have any duplicate content is pointless, as it may actually stop you for ranking on keywords specific to pages not relative to the tag.
Please, may you present me a document that assess what are you saying? because it is the first time I hear this thing.
#curious
-
The disadvantage to keeping a canonical tag on a page which does not require it would be, as a rule, you want to present your web page with the least amount of code possible. Unnecessary code causes extra confusion and adds to the processing time of web pages.
I use the canonical tag on all pages, but not everyone agrees. If you would like further support, SEOmoz uses the tag on all pages as well. If you use any CMS, ecommerce software, forum software or any system which generates pages dynamically then I would highly recommend a canonical tag on every page. At times a system will generate pages which you might not be aware of, but a crawler will find.
Sometimes a page will offer a print version, the ability to sort on ascending/descending, and numerous other changes. You might think you only have one version of your page but have many versions which you do not realize exist. A proper canonical tag ensures the correct version of your URL is always offered for indexing, and you avoid duplicate content issues. With that said, if you have a basic html/css/php site and you are 100% confident in your programmer, then it is not necessary.
EDIT: In your case, it seems the canonical tags are performing a necessary function. From your home page I clicked on your featured item and I landed on the following URL:
http://www.nathosp.com/product/r1212_c
You have the identical page offered under another URL: http://www.nathosp.com/product/r1212_c/hotel_towels.
If you were to remove the canonical, you would have duplicate content issues on your site.
-
rel=canonical just passes all link juice from one page to the next, it tells bots to use the page specified in the tag to assess link value and page authority. Having canonical tags on pages that don't have any duplicate content is pointless, as it may actually stop you for ranking on keywords specific to pages not relative to the tag. I would look at it closely or ask the last SEO why they did this before removing them. But by the sounds of it, you dont really need them.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best Practice Approaches to Canonicals vs. Indexing in Google Sitemap vs. No Follow Tags
Hi There, I am working on the following website: https://wave.com.au/ I have become aware that there are different pages that are competing for the same keywords. For example, I just started to update a core, category page - Anaesthetics (https://wave.com.au/job-specialties/anaesthetics/) to focus mainly around the keywords ‘Anaesthetist Jobs’. But I have recognized that there are ongoing landing pages that contain pretty similar content: https://wave.com.au/anaesthetists/ https://wave.com.au/asa/ We want to direct organic traffic to our core pages e.g. (https://wave.com.au/job-specialties/anaesthetics/). This then leads me to have to deal with the duplicate pages with either a canonical link (content manageable) or maybe alternatively adding a no-follow tag or updating the robots.txt. Our resident developer also suggested that it might be good to use Google Index in the sitemap to tell Google that these are of less value? What is the best approach? Should I add a canonical link to the landing pages pointing it to the category page? Or alternatively, should I use the Google Index? Or even another approach? Any advice would be greatly appreciated. Thanks!
Intermediate & Advanced SEO | | Wavelength_International0 -
Website homepage must be first link from every page of sub folders which are related in content?
Hi, I have seen number of websites where they keep their website homepage as first link from every page of sub folders and even from sub domain pages some times. For example, giving website homepage as top navigational menu from every page in their blog. Will this helps in boosting the ranking of homepage if we link from sub folder or sub domain pages where they employ related content like blogs or help guides. Thanks
Intermediate & Advanced SEO | | vtmoz0 -
Canonical Tag Question Regarding Two State Pages
Hi Moz Fam! Question... we have two pages for each state. Both pages are used in our PPC campaigns. One of the two state pages is fully optimized for organic search. The page that's optimized for organic has 1,500-2,000 words, the other one has thin content. All the content is unique, nothing duplicate. We call one set of state pages that I optimized my SEO state pages, then the other ones are our PPC state pages. Should I be setting a canonical tag to one of these pages to let Google know which one is the "master" page? (My SEO state page is the master) I've never used them, so I'm not sure what the right answer is for this. Thanks!
Intermediate & Advanced SEO | | LindsayE0 -
Substantial difference between Number of Indexed Pages and Sitemap Pages
Hey there, I am doing a website audit at the moment. I've notices substantial differences in the number of pages indexed (search console), the number of pages in the sitemap and the number I am getting when I crawl the page with screamingfrog (see below). Would those discrepancies concern you? The website and its rankings seems fine otherwise. Total indexed: 2,360 (Search Consule)
Intermediate & Advanced SEO | | Online-Marketing-Guy
About 2,920 results (Google search "site:example.com")
Sitemap: 1,229 URLs
Screemingfrog Spider: 1,352 URLs Cheers,
Jochen0 -
Canonicle & rel=NOINDEX used on the same page?
I have a real estate company: www.company.com with approximately 400 agents. When an agent gets hired we allow them to pick a URL which we then register and manage. For example: www.AGENT1.com We then take this agent domain and 301 redirect it to a subdomain of our main site. For example
Intermediate & Advanced SEO | | EasyStreet
Agent1.com 301’s to agent1.company.com We have each page on the agent subdomain canonicled back to the corresponding page on www.company.com
For example: agent1.company.com canonicles to www.company.com What happened is that google indexed many URLS on the subdomains, and it seemed like Google ignored the canonical in many cases. Although these URLS were being crawled and indexed by google, I never noticed any of them rank in the results. My theory is that Google crawled the subdomain first, indexed the page, and then later Google crawled the main URL. At that point in time, the two pages actually looked quite different from one another so Google did not recognize/honor the canonical. For example:
Agent1.company.com/category1 gets crawled on day 1
Company.com/category1 gets crawled 5 days later The content (recently listed properties for sale) on these category pages changes every day. If Google crawled the pages (both the subdomain and the main domain) on the same day, the content on the subdomain and the main domain would look identical. If the urls are crawled on different days, the content will not match. We had some major issues (duplicate content and site speed) on our www.company.com site that needed immediate attention. We knew we had an issue with the agent subdomains and decided to block the crawling of the subdomains in the robot.txt file until we got the main site “fixed”. We have seen a small decrease in organic traffic from google to our main site since blocking the crawling of the subdomains. Whereas with Bing our traffic has dropped almost 80%. After a couple months, we have now got our main site mostly “fixed” and I want to figure out how to handle the subdomains in order to regain the lost organic traffic. My theory is that these subdomains have a some link juice that is basically being wasted with the implementation of the robots.txt file on the subdomains. Here is my question
If we put a ROBOTS rel=NOINDEX on all pages of the subdomains and leave the canonical (to the corresponding page of the company site) in place on each of those pages, will link juice flow to the canonical version? Basically I want the link juice from the subdomains to pass to our main site but do not want the pages to be competing for a spot in the search results with our main site. Another thought I had was to place the NOIndex tag only on the category pages (the ones that seem to change every day) and leave it off the product (property detail pages, pages that rarely ever change). Thank you in advance for any insight.0 -
Any downsides of (permanent)redirecting 404 pages to more generic pages(category page)
Hi, We have a site which is somewhat like e-bay, they have several categories and advertisements posted by customers/ client. These advertisements disappear over time and turn into 404 pages. We have the option to redirect the user to the corresponding category page, but we're afraid of any negative impact of this change. Are there any downsides, and is this really the best option we have? Thanks in advance!
Intermediate & Advanced SEO | | vhendriks0 -
How would you handle 12,000 "tag" pages on Wordpress site?
We have a Wordpress site where /tag/ pages were not set to "noindex" and they are driving 25% of site's traffic (roughly 100,000 visits year to date). We can't simply "noindex" them all now, or we'll lose a massive amount of traffic. We can't possibly write unique descriptions for all of them. We can't just do nothing or a Panda update will come by and ding us for duplicate content one day (surprised it hasn't already). What would you do?
Intermediate & Advanced SEO | | M_D_Golden_Peak1 -
Could adding canonical tags to large Ecommerce site ever hurt rankings? Temporarily?
We have a really large site we're working on who's product pages rank well for the most part but also have multiple products listed in different categories with different URL's. I'm assuming there's no downside to adding canonical tags to these right? Its peak season so I don't want to do anything that could, even temporarily, bring down their rankings. Thanks!
Intermediate & Advanced SEO | | iAnalyst.com0