Mapping and tracking old and new information architecture
-
Howdy.
So I'm working on "example.com", which has thousands of URLs. The site is going to be redesigned, with some changes to the information architecture.
I'm trying to think of a good way to organize and account for similarities and differences between the original information architecture and the new one. This should help with building 301s.
I've downloaded a list of URLs from example.com from Open Site Explorer. What I would love to do is generate a visual "tree" of the site based on the output from Open Site Explorer. It would basically look like a pyramid with all of the subfolders branching out.
Does anybody know of a tool out there that will do this for me? Or am I going to have a long day in Excel?
Any other thoughts on working through this process are welcome.
Thank you!
-
I wouldn't use OSE for this, given that they may not crawl all your urls
I suggest using a specifc cralwer that you can set to crawl the whole site. Give Xenu, Screaming frog or IIS toolkit a go to get a good idea of your urls. After you go live, make sure you have mapped all your old urls across to new ones
In my experience, a site never actually looks like a tree, people just like to describe it that way because its simple
S
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Do Google Penalties Always Follow a Redirects to New Domains?
I have a couple sites that were penalized by Google for hosting content that made Google look bad. After a major newspaper showcased what was going on they suddenly took a major hit as if someone at Google flipped a switch and told their system not to rank the content for anything other than their brand names. The article made Google look bad because the newspaper highlighted a lot of unverified user generated accusations the reporters assumed not to be true in the context of "these accusations are mostly false, but they still show up on the first page when people search Google." I was thinking one way to fight this would simply be to host the content at a different domain, but I am concerned about the new domain being penalized as well. I don't want to completely shut down all of the original sites because some of them have brand recognition. The oldest domain is 12 years old with backlinks from several news outlets which is why the content ranked so well, but after the penalty that is only the case on Bing. I've read various articles about this tactic. Some say that you will almost always pass the penalty to the new domain if you do a 301 redirect, but the penalties at issue in those articles were for things like buying links or other black hat tactics. This is somewhat different in that I wasn't doing anything black hat, they just decided not to let the site rank for political reasons. I was hoping that maybe that type of penalty wouldn't follow it, but right now I am leaning towards simply creating a second site to syndicate articles. It will need to attribute the articles to their sources though, so they will need either no followed links or possibly a redirection script that bots cannot follow. I would really like it if I could simply change the first site to its .net or .org equivalent and 301 everything though.
Technical SEO | | PostAlmostAnything0 -
Moving content to a new domain
I need to move a lot of content with podcasts and show notes to a new domain. Instead of doing redirects, we want to keep some content on the current domain to retain the link value. There are business reason to keep content on both websites but the new website will primarily be used for SEO moving forward.If we keep the audio portion of the podcast on the old website and move the show notes and the audio portion of the podcast to the new website, is there any issues with duplicate content?Long-term, I presume Google will re-index the old and the new pages, thus no duplicate content, but I want to make sure I'm not missing anything. I was planning to fetch pages in Search Console as we migrate content.Thanks for your help!
Technical SEO | | JimmyFritz0 -
What do you think about my new site?
Hi everyone, I'm looking for a review for my new site www.interlive.it Could you please let me know what do you think about the work that I did for my site. I'll be very happy to receive your suggestions. Regards, Mike
Technical SEO | | salvyy0 -
What if an old site goes into PENDINGDELETE status
Hi, I have an old domain which accidentally was set as PENDINGDELETE by the registry. It's now not resolving to any ip address any more. Actually I was relocating from the old domain to a new domain. Just one month before it become PENDINGDELETE, I have submitted a "Chang of Address" in Google Webmasters Tools as well as setup the web server to 301 redirect all urls on old domain to the new domain. I have some sub-questions for this case: 1. What will happen to the effectiveness of the "Change of Address" in Google Webmasters Tool after old domain is dropped. As a domain is deleted, I have no way to maintain the verified ownership of the it in case Google asks me to reverify. 2. Suppose during last month before it's deleted, Googlebot had crawled 50% of urls on old domains, detected the 301 redirects and save them to its index. When Googlebot crawls those 50% urls again after the old domain is deleted, as those urls are not resolving to any web server, will Googlebot retain the last 301 redirects or drop the 301 redirects as well? 3. After a domain is deleted, how soon will Google purge urls on old domain from its index? Thank you. Best regards Jack Zhao
Technical SEO | | Bull1350 -
Old documents online as link juice
Each month I upload my auction catalog in different formats (word, pdf and rtf). I have about 9 years of catalogues online that have all been indexed by Google. In each catalog there is a link to my terms and conditions page (which has made the page authority for that page quite high in some unusual, but desired keywords), there is also many, many mentions of non-desired keywords in each of those documents and links to my domain. Is it worth updating all these old, previously indexed catalogues with better keyword juice and more relevant links ? Would they even get re-visited by google ? I suppose that leads to the next question... is it worth adding each of these pages to my sitemap ? To this point I have only added my major pages, not any of the subordinate pages etc.
Technical SEO | | blinkybill0 -
New Site Search Critique
Hi I am a huge fan of the SEOMOZ site and this great community which has helped me learn the current SEO skills I have now which are still very basic compared to the pros on the forum. I have tried to follow best practice regarding onsite and technical seo when developing my new site www.cheapfindergames.com and I would really appreciate it if experts on the forum could spare a minute to critique the site from a search perspective please This will give any elements of what onsite and technical SEO I done well and what aspects still need work. I am currently trying to build quality links and social mentions into the site which will take time, and the site has been designed around usability and conversions. Many Thanks Ian
Technical SEO | | ocelot0 -
Is use of javascript to simplify information architecture considered cloaking?
We are considering using javascript to format URLs to simplify the navigation of the googlebot through our site, whilst presenting a larger number of links for the user to ensure content is accessible and easy to navigate from all parts of the site. In other words, the user will see all internal links, but the search engine will see only those links that form our information hierarchy. We are therefore showing the search engine different content to the user only in so far as the search engine will have a more hierarchical information architecture by virture of the fact that there will be fewer links visible to the search engine to ensure that our content is well structured and discoverable. Would this be considered cloaking by google and would we be penalised?
Technical SEO | | JohnHillman0 -
New URL structure caused a HUGE drop?
I have started working with a client who did an upgrade on their e-commerce sive in May of last year. It totally changed the URL structure and they didn't redirect old URLs or do any of the things they should have. Not unexpectedly they they went from about 300 visitors a day to 0 for then rose up to maybe 50 and have remained there ever since. There were some major onsite issues including about 15000 internal links that 302 back to the site. In any case I have fixed most of the onsite problems and worked on a little better categorization + content optimization, etc. We have only been working on this for about 30 days and organic traffic is up and they are ranking for much better keywords, but I expected a little quicker rise. Here is a screenshot out of GA of their descent. Its pretty rapid. I dont think it makes sense to redirect their old URLs at this point since most of them have been deindexed for 10+ months. Anyone have any suggestions on how to get back to their previous level. The domain actually has decent authority and link profile, etc. Is this just going to be a slow climb back? Any thoughts? Fxz9Y.png
Technical SEO | | BlinkWeb0