Multiple doamin with same content?
-
I have multiple websites with same content such as
http://www.example.org and so on. My primary url is http://www.infoniagara.com and I also placed a 301 on .org.
Is that enough to keep away my exampl.org site from indexing on google and other search engines?
the eaxmple.org also has lots of link to my old html pages (now removed). Should i change that links too? or will 301 redirection solve all such issues (page not found/crawl error) of my old webpages?
i would welcome good seo practices regarding maintaining multiple domains
thanks and regards
-
You want your redirect rules on the server, not client site. In Apache you can do this with mod_write and the .htaccess file like so.
To add the www
RewriteEngine on
RewriteCond %{HTTP_HOST} !^www.
RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]To remove the www:
RewriteEngine On
RewriteCond %{HTTP_HOST} ^www.(.+)$ [NC]
RewriteRule ^(.*)$ http://%1/$1 [R=301,L]In IIS they have a rewrite command too. I've not used it myself but this should help: http://www.petermoss.com/post/How-to-redirect-non-www-domain-to-www-domain-requests-in-IIS-7.aspx
-
Anyway I am lucky being with a group of average team mates.
-
I would recommend either asking a new Q&A as to how an IIS redirect works, or checking Google. I lack experience working in that environment. I was spoiled by working with a very talented team who had performed all those changes and I never needed to learn any aspects of IIS.
-
Hi,
Thanks. As you said i was checking server side redirection. My site server is II7. I tried to make a server side rediretion but couldn't. I found a java script redirection and created a redirection. See the page; http://www.infoniagara.com/d-bed-roses.html
I think this too is not a correct redirection, is it?
thanks
-
Glad to be of help. You are always free to reach out here at the SEOmoz Q&A. If you feel a need to reach me specifically my contact information is in my user profile.
-
Ohh thanks so much Ryan. Let me learn the server side redirection and other aspects related to it.
Thanks once again for your consideration and time. I hope I can approach you in future too.
best regards
-
The javascript code you shared is not a proper redirect.
A proper redirect happens on the server. Instead of loading the original target page, the server instead will load the redirect page instantly along with a header response code of 301 which tells search engines the content has moved to a new URL.
If you use javascript in the manner you shared, the original page will load with a 200 "all ok" header code, and then 3 seconds later the javascript will trigger and load the new page with a 200 header code. All the backlinks will still be applied to the original page and not the redirected page.
The exact method of performing a redirect varies based on your server setup. If you have a LAMP server with cPanel, there is a Redirect tool which you can use.
-
hi,
It's really helpful and now I understand it.
I know that you are one of the true masters of SEO and I think you can clarify one more doubt. It is also regarding the issue of redirection. I want to know that a script i use for redirectining old pages to newpages is right or not?
This script is working properly and I want to know that what type of redirection is this and is it a proper redirection to get the backlink juice? (I add this script to the body just after the
header.)
thanks in advance
-
A 301 redirect is the proper solution and superior to the canonical. It is fine to have the canonical too, but add the redirect.
When you ask "should I do it for each page", understand a single redirect can forward all non-www traffic on your site to it's www equivalent. If you are unsure how to perform the redirect, simply ask your host. Most sites are on managed hosting and it is a very common and easy request.
-
hi dear Ryan,
thanks so much for your time and valuable suggestions. As you said, i am aware of this problem and therefore i added a canonical url to the homepage. should i make a 301 from non-www to www url? should i do it for each page?
thanks & regards
-
In short, you should not use duplicate content across various domains. Doing such will likely negatively affect rankings for either site. Try doing a search that would naturally return the results for a duplicated web page. You will likely find one page ranks well, while the other page ranks significantly lower due to the duplication.
I checked your .org site and it's pages are properly 301 redirected to the .com site. This change would cause any valid pages listed on the .org site to disappear from Google's index. It may take a month from the date the 301 was implemented for Google to crawl and update the entire site.
One point I would add is I suggest you perform a Google site:http://www.infoniagara.org search. Notice you do have a lot of search results for the .org site still present. Those pages are properly redirected to the .org site but they receive 404 errors. If the pages are really gone and there is no equivalent, that is fine and these results should disappear from Google's index over time. If there are similar pages on your site, you should 301 redirect these pages to them.
Another issue, your .com site appears in both the www and non-www form. If you take a URL, remove the "www" then the page appears normally showing the non-www URL. This is a problem which needs to be fixed as it is dividing your backlink juice. Pick one version of your URL, the www or non-www version, and 301 the other version.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do I make a content calendar to increase my rank for a key word?
I've watched more than a few seminars on having a content calendar. Now I'm curious as to what I would need to do to increase ranking for a specific keyword in local SEO. Let's say I wanted to help them increase their rank for used trucks in buffalo, NY. Would I regularly publish blog posts about used trucks? Thanks!
White Hat / Black Hat SEO | | oomdomarketing0 -
20-30% of our ecommerce categories contain no extra content, could this be a problem
Hello, About 20-30% of our ecommerce categories have no content beyond the products that are in them. Could this be a problem with Panda? Thanks!
White Hat / Black Hat SEO | | BobGW0 -
Are multiple links on the same domain worth pursuing?
Let's say you get a guest post on a blog that links to your site for $100. How much is the link from another guest post on the same domain worth? $90? $50? Does each additional link from the same domain lose its SEO impact? What if one blogger loves your content and is willing to post 10+ of your blog posts with links to your site in each - is that worth pursuing just from an SEO standpoint (I know it can be a good branding opportunity if the readership is right)?
White Hat / Black Hat SEO | | pbhatt0 -
Looking for a Way to Standardize Content for Thousands of Pages w/o Getting Duplicate Content Penalties
Hi All, I'll premise this by saying that we like to engage in as much white hat SEO as possible. I'm certainly not asking for any shady advice, but we have a lot of local pages to optimize :). So, we are an IT and management training course provider. We have 34 locations across the US and each of our 34 locations offers the same courses. Each of our locations has its own page on our website. However, in order to really hone the local SEO game by course topic area and city, we are creating dynamic custom pages that list our course offerings/dates for each individual topic and city. Right now, our pages are dynamic and being crawled and ranking well within Google. We conducted a very small scale test on this in our Washington Dc and New York areas with our SharePoint course offerings and it was a great success. We are ranking well on "sharepoint training in new york/dc" etc for two custom pages. So, with 34 locations across the states and 21 course topic areas, that's well over 700 pages of content to maintain - A LOT more than just the two we tested. Our engineers have offered to create a standard title tag, meta description, h1, h2, etc, but with some varying components. This is from our engineer specifically: "Regarding pages with the specific topic areas, do you have a specific format for the Meta Description and the Custom Paragraph? Since these are dynamic pages, it would work better and be a lot easier to maintain if we could standardize a format that all the pages would use for the Meta and Paragraph. For example, if we made the Paragraph: “Our [Topic Area] training is easy to find in the [City, State] area.” As a note, other content such as directions and course dates will always vary from city to city so content won't be the same everywhere, just slightly the same. It works better this way because HTFU is actually a single page, and we are just passing the venue code to the page to dynamically build the page based on that venue code. So they aren’t technically individual pages, although they seem like that on the web. If we don’t standardize the text, then someone will have to maintain custom text for all active venue codes for all cities for all topics. So you could be talking about over a thousand records to maintain depending on what you want customized. Another option is to have several standardized paragraphs, such as: “Our [Topic Area] training is easy to find in the [City, State] area. Followed by other content specific to the location
White Hat / Black Hat SEO | | CSawatzky
“Find your [Topic Area] training course in [City, State] with ease.” Followed by other content specific to the location Then we could randomize what is displayed. The key is to have a standardized format so additional work doesn’t have to be done to maintain custom formats/text for individual pages. So, mozzers, my question to you all is, can we standardize with slight variations specific to that location and topic area w/o getting getting dinged for spam or duplicate content. Often times I ask myself "if Matt Cutts was standing here, would he approve?" For this, I am leaning towards "yes," but I always need a gut check. Sorry for the long message. Hopefully someone can help. Thank you! Pedram1 -
Is this Duplicate content?
Hi all, This is now popping up in Moz after using this for over 6 months.
White Hat / Black Hat SEO | | TomLondon
It is saying this is now duplicate site content. What do we think? Is this a bad strategy, it works well on the SERPS but could be damaging the root domain page ranking? I guess this is a little shady. http://www.tomlondonmagic.com/area/close-up-magician-in-crowborough/ http://www.tomlondonmagic.com/area/close-up-magician-in-desborough/ http://www.tomlondonmagic.com/area/close-up-magician-in-didcot/ Thanks.0 -
Is it still valuable to place content in subdirectories to represent hierarchy or is it better to have every URL off the root?
Is it still valuable to place content in subdirectories to represent hierarchy on the site or is it better to have every URL off the root? I have seen websites structured both ways. It seems having everything off the root would dilute the value associated with pages closest to the homepage. Also, from a user perspective, I see the value in a visual hierarchy in the URL.
White Hat / Black Hat SEO | | belcaro19860 -
Content ideas?
We run a printing company and we are struggling to come up with unique content people will actually want to know, is there any way of getting the ball rolling? We were thinking of ideas such as exhibition guide but this seems to have been overdone. Any help would be appreciated.
White Hat / Black Hat SEO | | BobAnderson0 -
How does Google decide what content is "similar" or "duplicate"?
Hello all, I have a massive duplicate content issue at the moment with a load of old employer detail pages on my site. We have 18,000 pages that look like this: http://www.eteach.com/Employer.aspx?EmpNo=26626 http://www.eteach.com/Employer.aspx?EmpNo=36986 and Google is classing all of these pages as similar content which may result in a bunch of these pages being de-indexed. Now although they all look rubbish, some of them are ranking on search engines, and looking at the traffic on a couple of these, it's clear that people who find these pages are wanting to find out more information on the school (because everyone seems to click on the local information tab on the page). So I don't want to just get rid of all these pages, I want to add content to them. But my question is... If I were to make up say 5 templates of generic content with different fields being replaced with the schools name, location, headteachers name so that they vary with other pages, will this be enough for Google to realise that they are not similar pages and will no longer class them as duplicate pages? e.g. [School name] is a busy and dynamic school led by [headteachers name] who achieve excellence every year from ofsted. Located in [location], [school name] offers a wide range of experiences both in the classroom and through extra-curricular activities, we encourage all of our pupils to “Aim Higher". We value all our teachers and support staff and work hard to keep [school name]'s reputation to the highest standards. Something like that... Anyone know if Google would slap me if I did that across 18,000 pages (with 4 other templates to choose from)?
White Hat / Black Hat SEO | | Eteach_Marketing0