Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Merge 2 websites into one, using a non-existing, new domain.
-
I need to merge https://www.WebsiteA.com and https://www.WebsiteB.com to a fresh new domain (with no content) https://www.WebsiteC.com. I want to do it the best way to keep existing SEO juice.
Website A is the companies home page and built with Wordpress
Website B is the company product page and built with Wordpress
Website C will be the new site containing both website A and B, utilizing Wordpress also.
What is the best way to do this? I have research a lot and keep hitting walls on how to do it.
It's a little trickier because it's two different domains going to a brand new domain.
Thanks
-
Thanks Andy!
I'm still a little confused on how I will be copying data over from Site A to Site C. Is it just a matter of creating an empty Wordpress shell with no template and copying the data from Site A to it? (And how is that usually done?) Then would I redirect all Site A's pages to Site C's pages, just with the URL changed? I.E www.SiteA.com/large-dogs redirect to www.SiteC.com/large-dogs
I don't want all of Site B either, just about 10 pages from it, would I manually copy them over also, and how would that be done?
Thanks
-
I did just this type of thing a little over a year ago and organic traffic is up over 300% now. We made the change mainly to improve the structure of the website(s), with more logical organization and better internal linking. We did do the move all at once (thousands of pages) but it took a lot of behind-the scenes planning to be ready for that.
First came the decisions about what sections and categories made sense for our site. (Using the URL structure to guide users around the site makes it easier for them to find what they are looking for and interlinking between related posts as appropriate is also good—and this helps a lot with search engines.)
Then came the organization of posts into their new categories. To make things easier, we kept the individual path names the same (so www.siteA.com/old-category/old-post-string became www.siteC.com/new category/old-post-string) and uploaded them into their new categories when the time came.
We also used this time to do a limited content review (posts with the most traffic) and we updated a lot of these. We made the choice to keep most of our old posts, even though in our market they can get outdated quickly, to conserve any links we may have acquired. (The main site that we were directing to the new site was pretty old and had picked up a lot of links over time.)
We could have done a more complete content review before the changeover, but in part we wanted to see how these posts did under the new structure—we did get renewed life out of some of them, and we further updated and optimized those.
In conjunction with the export of the old sites to the new one, we made sure to 301 redirect all of the old posts to their counterparts on the new site. For the posts we chose not to bring over, we 301 redirected them to a related post in the same category.
We still occasionally come across things that need to be fixed—old posts that need redirecting/updating or 404 errors that need to be tracked down (one big issue we found was a lot of old pages had old links with hard paths to the old website root domains, causing a bunch of nasty internal not found errors—not good!) but overall we are happy with the change. (Up 308%!)
-
Hi,
One way to do this is to decide which site is going to be the main site (site A) sat on C and copy this data over. If you are bringing in site B, then this can sit at another structure level - you will end up with this...
Site A --> Site C, main pages
Site B --> Site C, product pagesThat then brings in everything from both sites to the new domain.
You then want to redirect both of the old sites to the new one, but don't redirect everything to the root. That isn't a good use of 301 mapping. You need to be mapping on a page level so that you will see...
www.sitea.com/about us -301- www.sitec.com/aboutus
www.siteb.com/newproducts/hammers -301- www.sitec.com/newproducts/hammersThere can be differences in the URL's - you don't need to stick with the same structure as the other sites if it doesn't make sense, but always map pages to something very similar.
Page level is the only way to go if you want to maintain a seamless transition for users as well.
Also, don't expect to hit a switch and do this all at once. You can do this over a period of time because to the user, they will just be redirected to the new pages. You will retain more link juice like this.
This is quite a lengthy process and I am sure I have missed the in-between bits, but this is the basis of what you want to be doing.
Others might chip in with other suggestions for you.
-Andy
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
GoogleBot still crawling HTTP/1.1 years after website moved to HTTP/2
Whole website moved to https://www. HTTP/2 version 3 years ago. When we review log files, it is clear that - for the home page - GoogleBot continues to only access via HTTP/1.1 protocol Robots file is correct (simply allowing all and referring to https://www. sitemap Sitemap is referencing https://www. pages including homepage Hosting provider has confirmed server is correctly configured to support HTTP/2 and provided evidence of accessing via HTTP/2 working 301 redirects set up for non-secure and non-www versions of website all to https://www. version Not using a CDN or proxy GSC reports home page as correctly indexed (with https://www. version canonicalised) but does still have the non-secure version of website as the referring page in the Discovery section. GSC also reports homepage as being crawled every day or so. Totally understand it can take time to update index, but we are at a complete loss to understand why GoogleBot continues to only go through HTTP/1.1 version not 2 Possibly related issue - and of course what is causing concern - is that new pages of site seem to index and perform well in SERP ... except home page. This never makes it to page 1 (other than for brand name) despite rating multiples higher in terms of content, speed etc than other pages which still get indexed in preference to home page. Any thoughts, further tests, ideas, direction or anything will be much appreciated!
Technical SEO | | AKCAC1 -
In writing the url, it is better to use the language used by the people of my country or English?
We speak Persian and all people search in Persian on Google. But I read in some sources that the url should be in English. Please tell me which language to use for url writing?
Technical SEO | | ghesta
For example, I brought down two models: 1fb0e134-10dc-4737-904f-bfdf07143a98-image.png https://ghesta.ir/blog/how-to-become-rich/
2)https://ghesta.ir/blog/چگونه-پولدار-شویم/0 -
My WP website got attack by malware & now my website site:www.example.ca shows about 43000 indexed page in google.
Hi All My wordpress website got attack by malware last week. It affected my index page in google badly. my typical site:example.ca shows about 130 indexed pages on google. Now it shows about 43000 indexed pages. I had my server company tech support scan my site and clean the malware yesterday. But it still shows the same number of indexed page on google.
Technical SEO | | ChophelDoes anybody had ever experience such situation and how did you fixed it. Looking for help. Thanks FILE HIT LIST:
{YARA}Spam_PHP_WPVCD_ContentInjection : /home/example/public_html/wp-includes/wp-tmp.php
{YARA}Backdoor_PHP_WPVCD_Deployer : /home/example/public_html/wp-includes/wp-vcd.php
{YARA}Backdoor_PHP_WPVCD_Deployer : /home/example/public_html/wp-content/themes/oceanwp.zip
{YARA}webshell_webshell_cnseay02_1 : /home/example2/public_html/content.php
{YARA}eval_post : /home/example2/public_html/wp-includes/63292236.php
{YARA}webshell_webshell_cnseay02_1 : /home/example3/public_html/content.php
{YARA}eval_post : /home/example4/public_html/wp-admin/28855846.php
{HEX}php.generic.malware.442 : /home/example5/public_html/wp-22.php
{HEX}php.generic.cav7.421 : /home/example5/public_html/SEUN.php
{HEX}php.generic.malware.442 : /home/example5/public_html/Webhook.php0 -
Google tries to index non existing language URLs. Why?
Hi, I am working for a SAAS client. He uses two different language versions by using two different subdomains.
Technical SEO | | TheHecksler
de.domain.com/company for german and en.domain.com for english. Many thousands URLs has been indexed correctly. But Google Search Console tries to index URLs which were never existing before and are still not existing. de.domain.com**/en/company
en.domain.com/de/**company ... and an thousand more using the /en/ or /de/ in between. We never use this variant and calling these URLs will throw up a 404 Page correctly (but with wrong respond code - we`re fixing that 😉 ). But Google tries to index these kind of URLs again and again. And, I couldnt find any source of these URLs. No Website is using this as an out going link, etc.
We do see in our logfiles, that a Screaming Frog Installation and moz.com w opensiteexplorer were trying to access this earlier. My Question: How does Google comes up with that? From where did they get these URLs, that (to our knowledge) never existed? Any ideas? Thanks 🙂0 -
We switched the domain from www.blog.domain.com to domain.com/blog.
We switched the domain from www.blog.domain.com to domain.com/blog. This was done with the purpose of gaining backlinks to our main website as well along with to our blog. This set us very low in organic traffic and not to mention, lost the backlinks. For anything, they are being redirected to 301 code. Kindly suggest changes to bring back all the traffic.
Technical SEO | | arun.negi0 -
Using the word "FREE" in domain name
Hi, This may seem like a simple question but a new client of mine wishes to use a domain name with the word "free" in it. The website will offer free activity vouchers. I couldn't see this being a problem as there a lot of websites that do this although he was told it may present a problem with the search engines thinking the site was spammy. It won't be and will be offering information and vouchers on local sporting activities. I was wondering if anybody could clarify this please so I can give him a more definitive answer to his question. Thanks for your help.
Technical SEO | | malinkymedia0 -
Multilingual Website - Sub-domain VS Sub-directory
Hi Folks - Need your advice on the pros and cons of going with a sub-domain vs a sub-directory approach for a multi lingual website. The best would be a ccTLD but that is not possible now, so I would be more interested in knowing your take on these 2 options. Though, I have gone through http://www.stateofsearch.com/international-multilingual-sites-criteria-to-establish-seo-friendly-structure/ and this somewhat vouches for a sub-directory, but what would you say'?
Technical SEO | | RanjeetP0 -
SEO Benefit from Redirecting New Exact Match Domains?
Hi, All! This is a question asked in the old Q & A section, but the answer was a little ambiguous and it was about 3 years ago, so I decided to repost and let the knowledgeable SEO public answer... From David LaFerney: It’s clear that it’s much easier to get high rankings for a term if your domain is an exact match for the query. If you own several such domains that are very related such as – investmentrealestate.com, positivecashflow.com, and rentalproperty.com – would you be able to benefit from those by 301ing them to a single site, or would you have to maintain separate sites to help capture those targeted phrases? In a nutshell – SEO wise, is it worth owning multiple domains to exactly match valuable search phrases? Or do you lose the exact match benefit when you redirect?>> To clarify: redirecting an old domain with lots of history and links to a new exact match domain seems to contain SEO benefit. (You get links+exact match domain, approximately.) But the other way around? Redirecting a new exact match domain to an older domain with links? Does that do anything for the ranking of the old domain for the exact match keyword? Or absolutely nothing? (My impression has been that it's nothing, but the question came up for a client and I just wanted to make sure I wasn't missing something.) Thanks in advance!
Technical SEO | | debi_zyx0