Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is it better to use XXX.com or XXX.com/index.html as canonical page
-
Is it better to use 301 redirects or canonical page? I suspect canonical is easier. The question is, which is the best canonical page, YYY.com or YYY.com/indexhtml? I assume YYY.com, since there will be many other pages such as YYY.com/info.html, YYY.com/services.html, etc.
-
Glad you got it sorted out. If you're 301-redirecting a lot of domains, I'd suggest doing it gradually or maybe holding off on the lowest-quality domains. Google can see a massive set of redirects as a bit of a red flag (too many people have bought up cheap domains and 301-redirected to consolidate the link equity). If the domains are really all closely related or if you're only talking about a handful (<5) then it's probably not a big issue.
-
I think things may be sorted out, but I am not sure. I actually put in 301-redirects from a bunch of domains that I own to this new domain, the content of which will eventually replace my main domain. But, I need to get the domain properly set up and optimized before I move it to my primary domain to replace the ancient web site. At that time, I will also redirect this site to the new, old site.
I used to have Google ad-words tied to some of the domains that I 301-redirected to the new web site that I am building. Those were just a waste of money, however, so I put them on hold. I also had a lot of problems with semel and buttons for web bouncing off those pages that I re-directed. I put in .htaccess commands to stop those spam sites and that seems to work.
-
Google seems to be indexing 30-ish pages, but when I look at the cached home-page, I'm actually seeing the home-page of http://rfprototype.com/. Did you recently change domains or 301-redirect the old site? The cache data is around Christmas (after the original question was posted), so I think we're missing part of the puzzle here.
-
So, I think I may have had things wrong. For one thing, it seems like moz and Google are only indexing 2 pages, while the site index shows something like 80 pages. (I suspect an image is a page, and there are a lot of images. But, there are about 10 or 12 distinct pages at the moment. Also, Google and moz do not seem to show the correct key words in any sense like they should, leading me to think that they were just spidering 2 pages. I don't know why. I added the following to my index.html header:
and
I assume I put them in the correct place. I also believe I don't need canonical pages anywhere else.
Should these changes to my index.html make the proper changes?
-
Yeah, I'd have to concur - all the evidence and case studies I've seen suggest that rel=canonical almost always passes authority (link equity). There are exceptions, but honestly, there are exceptions with 301s, too.
I think the biggest difference, practically, is the impact on human visitors. 301-redirects take people to a new page, whereas canonical tags don't.
-
In terms of rel=canonical that will pass value the same as a 301 redirect - for evidence have a look here:
http://moz.com/learn/seo/canonicalization
"Another option for dealing with duplicate content is to utilize the rel=canonical tag. The rel=canonical tag passes the same amount of link juice (ranking power) as a 301 redirect, and often takes much less development time to implement."
See DR Pete's response in this Moz Q&A:
http://moz.com/community/q/do-canonical-tags-pass-all-of-the-link-juice-onto-the-url-they-point-to
http://googlewebmastercentral.blogspot.co.uk/2009/02/specify-your-canonical.html
https://support.google.com/webmasters/answer/139066?rd=1
http://searchenginewatch.com/sew/how-to/2288690/how-and-when-to-use-301-redirects-vs-canonical
Matts Cutts stated there is not a whole lot of difference between the 301 and the canonical - they will both lose "just a tiny little amount bit, not very much at all" of credit from the referring page.
-
Ok, this is how I look at the situation.
So you have two URLs and the question is either to redirect301 or use canonical? In my opinion 301 is a better solution and this is because it will not only redirect people to the preferred version but the link value as well.
Whereas, with canonicals only search engines will know what is the preferred page but it will not transfer the link value which can help you with organic rankings.
Hope this helps!
-
You would put the canonical link in the index file and I would point that at the xxx.com version rather than the xxx.com/index.html version as people visiting your sites homepage are going to enter the domain and not the specific page so xxx.com rather than xxx.com/index.html.
There are some great articles on Moz explaining all this which I would suggest that you read -
http://moz.com/learn/seo/canonicalization
Dr Pete also did this post answering common questions on rel=canonical.
http://moz.com/blog/rel-confused-answers-to-your-rel-canonical-questions
In terms of 301 redirects and canonicalization both pass the same amount of authority gained by different pages. If you are trying to keep it as clean as possible you need to be careful you don't create an issue redirecting your index file to your domain - here is an old post explaining how moz solved this 301 redirect on an Apache server
http://moz.com/blog/apache-redirect-an-index-file-to-your-domain-without-looping
I personally find that if all your links on your site reference your preferred(canonical) URL for the homepage so in this case xxx.com and you redirect the www version to this or vice versa depending on your preference then you add a canonical in the index.html file pointing at xxx.com in this case unless you prefer to do it the other way round with www.xxx.com for both you will be fine.
Hope this helps
-
I forgot. Of course, there is no xxx.com page, per se. It is actually xxx.com/index.html so if you needed to put the canonical reference on xxx.com, how would you do it?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Does a no-indexed parent page impact its child pages?
If I have a page* in WordPress that is set as private and is no-indexed with Yoast, will that negatively affect the visibility of other pages that are set as children of that first page? *The context is that I want to organize some of the pages on a business's WordPress site into silos/directories. For example, if the business was a home remodeling company, it'd be convenient to keep all the pages about bathrooms, kitchens, additions, basements, etc. bundled together under a "services" parent page (/services/kitchens/, /services/bathrooms/, etc.). The thing is that the child pages will all be directly accessible from the menus, so there doesn't need to be anything on the parent /services/ page itself. Another such parent page/directory/category might be used to keep different photo gallery pages together (/galleries/kitchen-photos/, /galleries/bathroom-photos/, etc.). So again, would it be safe for pages like /services/kitchens/ and /galleries/addition-photos/ if the /services/ and /galleries/ pages (but not /galleries/* or anything like that) are no-indexed? Thanks!
Technical SEO | | BrianAlpert781 -
Blog Page Titles - Page 1, Page 2 etc.
Hi All, I have a couple of crawl errors coming up in MOZ that I am trying to fix. They are duplicate page title issues with my blog area. For example we have a URL of www.ourwebsite.com/blog/page/1 and as we have quite a few blog posts they get put onto another page, example www.ourwebsite.com/blog/page/2 both of these urls have the same heading, title, meta description etc. I was just wondering if this was an actual SEO problem or not and if there is a way to fix it. I am using Wordpress for reference but I can't see anywhere to access the settings of these pages. Thanks
Technical SEO | | O2C0 -
Canonical issues using Screaming Frog and other tools?
In the Directives tab within Screaming Frog, can anyone tell me what the difference between "canonicalised", "canonical", and "no canonical" means? They're found in the filter box. I see the data but am not sure how to interpret them. Which one of these would I check to find canonical issues within a website? Are there any other easy ways to identify canonical issues?
Technical SEO | | Flock.Media0 -
How to stop google from indexing specific sections of a page?
I'm currently trying to find a way to stop googlebot from indexing specific areas of a page, long ago Yahoo search created this tag class=”robots-nocontent” and I'm trying to see if there is a similar manner for google or if they have adopted the same tag? Any help would be much appreciated.
Technical SEO | | Iamfaramon0 -
Investigating a huge spike in indexed pages
I've noticed an enormous spike in pages indexed through WMT in the last week. Now I know WMT can be a bit (OK, a lot) off base in its reporting but this was pretty hard to explain. See, we're in the middle of a huge campaign against dupe content and we've put a number of measures in place to fight it. For example: Implemented a strong canonicalization effort NOINDEX'd content we know to be duplicate programatically Are currently fixing true duplicate content issues through rewriting titles, desc etc. So I was pretty surprised to see the blow-up. Any ideas as to what else might cause such a counter intuitive trend? Has anyone else see Google do something that suddenly gloms onto a bunch of phantom pages?
Technical SEO | | farbeseo0 -
How Does Google's "index" find the location of pages in the "page directory" to return?
This is my understanding of how Google's search works, and I am unsure about one thing in specific: Google continuously crawls websites and stores each page it finds (let's call it "page directory") Google's "page directory" is a cache so it isn't the "live" version of the page Google has separate storage called "the index" which contains all the keywords searched. These keywords in "the index" point to the pages in the "page directory" that contain the same keywords. When someone searches a keyword, that keyword is accessed in the "index" and returns all relevant pages in the "page directory" These returned pages are given ranks based on the algorithm The one part I'm unsure of is how Google's "index" knows the location of relevant pages in the "page directory". The keyword entries in the "index" point to the "page directory" somehow. I'm thinking each page has a url in the "page directory", and the entries in the "index" contain these urls. Since Google's "page directory" is a cache, would the urls be the same as the live website (and would the keywords in the "index" point to these urls)? For example if webpage is found at wwww.website.com/page1, would the "page directory" store this page under that url in Google's cache? The reason I want to discuss this is to know the effects of changing a pages url by understanding how the search process works better.
Technical SEO | | reidsteven750 -
De-indexing millions of pages - would this work?
Hi all, We run an e-commerce site with a catalogue of around 5 million products. Unfortunately, we have let Googlebot crawl and index tens of millions of search URLs, the majority of which are very thin of content or duplicates of other URLs. In short: we are in deep. Our bloated Google-index is hampering our real content to rank; Googlebot does not bother crawling our real content (product pages specifically) and hammers the life out of our servers. Since having Googlebot crawl and de-index tens of millions of old URLs would probably take years (?), my plan is this: 301 redirect all old SERP URLs to a new SERP URL. If new URL should not be indexed, add meta robots noindex tag on new URL. When it is evident that Google has indexed most "high quality" new URLs, robots.txt disallow crawling of old SERP URLs. Then directory style remove all old SERP URLs in GWT URL Removal Tool This would be an example of an old URL:
Technical SEO | | TalkInThePark
www.site.com/cgi-bin/weirdapplicationname.cgi?word=bmw&what=1.2&how=2 This would be an example of a new URL:
www.site.com/search?q=bmw&category=cars&color=blue I have to specific questions: Would Google both de-index the old URL and not index the new URL after 301 redirecting the old URL to the new URL (which is noindexed) as described in point 2 above? What risks are associated with removing tens of millions of URLs directory style in GWT URL Removal Tool? I have done this before but then I removed "only" some useless 50 000 "add to cart"-URLs.Google says themselves that you should not remove duplicate/thin content this way and that using this tool tools this way "may cause problems for your site". And yes, these tens of millions of SERP URLs is a result of a faceted navigation/search function let loose all to long.
And no, we cannot wait for Googlebot to crawl all these millions of URLs in order to discover the 301. By then we would be out of business. Best regards,
TalkInThePark0 -
Sitmap Page - HTML and XML
Hi there I have a domain which has a sitemap in html for regular users and a sitemap in xml for the spiders. I have a warning via seomoz saying that i have too many links on the html version. What do i do here? regards Stef
Technical SEO | | stefanok0