Both links with ".html" and without are working , Is that a problem ?
-
Default format of my url ending with ".html" , I know it's not a problem .. But both links with ".html" and without are working , Is that critical problem or not ? and how to solve it ?
-
If the canonical tag is on the page, despite the fact the page loads in either version only the one that you have in your canonical tag will be indexed.
As far the Moz reports these are not updated minute by minute, to see if the canonical tag has fixed the problem in Moz's reports you'll have to wait till they do their next crawl (once a week). You may also use Google's Webmaster and ask for a re-crawl after you made changes.
Hope that helps,
Don
-
Thanks a lot ^_^ , But i need some help to apply this since i'm not developer
I already have canonical tag made by our developers but i don't know if it work properly , How could i know ?
I'm still getting duplicate content reports
I'll state the situation for you
www.example.com/somepage.html (site main format)
rel="canonical" for this page is = http://www.example.com/somepage.html
www.example.com/somepage (duplicate)
rel="canonical" for this page is = http://www.example.com/somepage
www.example.com/somepage/ (another duplicate)
rel="canonical" for this page is = http://www.example.com/somepage/
Is that right or not ?
and how can i measure if google see that as duplicate or not ?
Thanks
-
Hi Mohamed,
This is an important issue. By leaving both versions of the pages up, you run the risk of having it indexed twice and possibly suffer from duplicate content penalties.
There are a couple ways to fix it, but the easiest would be add a canonical meta tag to the page that points to the version your site should be using
**for no extension**
or
**with the extension**
The other way would be to modify your .htaccess file in your root directory to forward all traffic to whichever version you are using.
To force removal of the extension you could do this
RewriteEngine On
RewriteCond %{REQUEST_FILENAME} !-f
RewriteRule ^([^.]+)$ $1.html [L,NC,R=301]To force extension you could do this
RewriteEngine On
RewriteCond %{REQUEST_URI} !^..html$
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule ^(.)$ $1.html [L,R=301]The htaccess should work on most host but if I remember correctly GoDaddy has some special requirements. Check with them if you are using them as a host.
Hope this helps,
Don
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Spam link? Links from linguee
Hi Everyone My site received a notification of unnatural links in Webmaster Tools and the site has had a penalty applied. I can see there are a lot of links from a site : linguee.com .de. nl. ect ..more than 30k of them! I am not sure where did those links come from! The suddenly appeared over the weekend. Does anyone has similar experience before and any suggestion? Thanks Ricky
Technical SEO | | SEO-SMB0 -
How Does Google's "index" find the location of pages in the "page directory" to return?
This is my understanding of how Google's search works, and I am unsure about one thing in specific: Google continuously crawls websites and stores each page it finds (let's call it "page directory") Google's "page directory" is a cache so it isn't the "live" version of the page Google has separate storage called "the index" which contains all the keywords searched. These keywords in "the index" point to the pages in the "page directory" that contain the same keywords. When someone searches a keyword, that keyword is accessed in the "index" and returns all relevant pages in the "page directory" These returned pages are given ranks based on the algorithm The one part I'm unsure of is how Google's "index" knows the location of relevant pages in the "page directory". The keyword entries in the "index" point to the "page directory" somehow. I'm thinking each page has a url in the "page directory", and the entries in the "index" contain these urls. Since Google's "page directory" is a cache, would the urls be the same as the live website (and would the keywords in the "index" point to these urls)? For example if webpage is found at wwww.website.com/page1, would the "page directory" store this page under that url in Google's cache? The reason I want to discuss this is to know the effects of changing a pages url by understanding how the search process works better.
Technical SEO | | reidsteven750 -
Best use of robots.txt for "garbage" links from Joomla!
I recently started out on Seomoz and is trying to make some cleanup according to the campaign report i received. One of my biggest gripes is the point of "Dublicate Page Content". Right now im having over 200 pages with dublicate page content. Now.. This is triggerede because Seomoz have snagged up auto generated links from my site. My site has a "send to freind" feature, and every time someone wants to send a article or a product to a friend via email a pop-up appears. Now it seems like the pop-up pages has been snagged by the seomoz spider,however these pages is something i would never want to index in Google. So i just want to get rid of them. Now to my question I guess the best solution is to make a general rule via robots.txt, so that these pages is not indexed and considered by google at all. But, how do i do this? what should my syntax be? A lof of the links looks like this, but has different id numbers according to the product that is being send: http://mywebshop.dk/index.php?option=com_redshop&view=send_friend&pid=39&tmpl=component&Itemid=167 I guess i need a rule that grabs the following and makes google ignore links that contains this: view=send_friend
Technical SEO | | teleman0 -
Page for Link Building
Hello guys, My question is about link building and reciprocal links. Since many directories request a reciprocal link, makes me wonder if is not better to create a unique page in the website only for this kind of links. What do you guys recommend? Thanks in advance, PP
Technical SEO | | PedroM0 -
Would this be considered "thin content?"
I share a lot of images via twitter and over the last year I've used several different tools to do this; mainly twitpic, and now instagram. Last year I wanted to try to find a way to host those images on my site so I could get the viewers of the picture back to my site instead a 3rd party (twitpic, etc.) I found a few plugins that worked "sort of" well, and so I used that for a while. (I have since stopped doing that in favor of using instagram.) But my question is do all of these image posts hurt my site you think? I had all of these images under a category called "twitter" but have since moved them to an uncategorized category until I figure out what I want to do with them. I wanted to see if anyone could chime in and give me some advice. Since the posts are just images with no content (other than the image) and the title isn't really "optimized" for anything do these posts do me more harm than good. Do I delete them all? Leave them as is? Or do something else? Also in hindsight I'm assuming this was a bad idea since the bounce rate for people clicking on a link just to see an image was probably very high, and may have caused the opposite result of what I was looking for. If I knew than what I know now I would have tracked the bounce rate of those links, how many people who viewed one of those images actually went to another page on the site, etc. But hindsight's 20/20. 🙂
Technical SEO | | NoahsDad0 -
International Websites: rel="alternate" hreflang="x"
Hi people, I keep on reading and reading , but I won't get it... 😉 I mean this page: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=189077&topic=2370587&ctx=topic On the bottom of the page they say: Step 2: Use rel="alternate" hreflang="x" Update the HTML of each URL in the set by adding a set of rel="alternate" hreflang="x" link elements. Include a rel="alternate" hreflang="x" link for every URL in the set, like this: This markup tells Google's algorithm to consider all of these pages as alternate versions of each other. OK! Each URL needs this markup. BUT: Do i need it exactly as written above, or do I have to put in the complete URL of the site, like: The next question is, what happens exactly in the SERPS when I do it like this (an also with Step1 that I haven't copied here)? Google will display the "canonical"-version of the page, but wehen a user from US clicks he will get on http://en-us.example.com/**page.htm **??? I tried to find other sites which use this method, but I haven't found one. Can someone give me an example.website??? Thank you, thank you very much! André
Technical SEO | | waynestock0 -
Will Google display the "@" Symbol in a SERP Title?
In our page title's, we'd like to include the "@" symbol. Will google display that symbol in the search results if we include it in the page's title?
Technical SEO | | sftravel0 -
Website Structure - What works best for passing the Link Juice?
As most people know it's usual that the main menu is after the top right small links in html. My questions are: What is more important for google or how does google tell which link is more important than the other to pass juice on? If the top right links are in front of the main menu in html would they get more link juice than the main menu? Should i focus in working on better html structure but still keeping the same look ( reverse the html code but keep the same look through css)? Any suggestions?
Technical SEO | | mosaicpro0