Impact of simplifying website and removing 80% of site's content
-
We're thinking of simplifying our website which has grown to a very large size by removing all the content which hardly ever gets visited.
The plan is to remove this content / make changes over time in small chunks so that we can monitor the impact on SEO. My gut feeling is that this is okay if we make sure to redirect old pages and make sure that the pages we remove aren't getting any traffic. From my research online it seems that more content is not necessarily a good thing if that content is ineffective and that simplifying a site can improve conversions and usability.
Could I get people's thoughts on this please? Are there are risks that we should look out for or any alternatives to this approach? At the moment I'm struggling to combine the needs of SEO with making the website more effective.
-
I have to agree with you on making this move. Content that doesn't contribute to the quality of your site and receives minimal traffic should be removed. Besides ensuring the redirects are set properly, you can evaluate if these old content do actually make good material for future writing. It would be a waste to just delete them without any second thoughts. Some snippets of these old content can still prove useful and be spinned into new articles once you elaborate on them.
-
Great answers guys - thanks. It's good to know that my gut feeling was close to the mark!
-
Quality over quantity is definitely the order of the day, but before you drop some content completely, take a look at it and see if there is some useful info contained in it which could be consolidated into some of the content that you are actually retaining. Overall though a good content audit can be a good thing even if it means dropping some pages. Here's a useful article regarding content audits which is well worth taking a look at.
-
Sounds like a good idea to me. Make sure you have all the redirects in place to make sure when people want to visit the old content they're redirected to the new content. Also make sure you monitor the rest of your sites SEO traffic to make sure you don't fall in a hidden trap.
-
I think this pruning process makes sense. Although this will potentially decrease key words it will streamline the navigation for the content that is actually getting traffic. This will provide a better flow and potentially a lower bounce rate. Staging these cuts and monitoring the changes seems like a good way to manage your risk.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google's Knowledge Panel
Hi Moz Community. Has anyone noticed a pattern in the websites that Google pulls in to populate knowledge Panels? For example, for a lot of queries Google keeps pulling data from a specific source over and over again, and the data shown in the Knowledge Panel isn't on the target page. Is it possible that Google simply favors some sites over others and no matter what you do, you'll never make it into the Knowledge box? Thanks.
Intermediate & Advanced SEO | | yaelslater0 -
Do I eventually 301 a page on our site that "expires," to a page that's related, but never expires, just to utilize the inbound link juice?
Our company gets inbound links from news websites that write stories about upcoming sporting events. The links we get are pointing to our event / ticket inventory pages on our commerce site. Once the event has passed, that event page is basically a dead page that shows no ticket inventory, and has no content. Also, each “event” page on our site has a unique url, since it’s an event that will eventually expire, as the game gets played, or the event has passed. Example of a url that a news site would link to: mysite.com/tickets/soldier-field/t7493325/nfc-divisional-home-game-chicago bears-vs-tbd-tickets.aspx Would there be any negative ramifications if I set up a 301 from the dead event page to another page on our site, one that is still somewhat related to the product in question, a landing page with content related to the team that just played, or venue they play in all season. Example, I would 301 to: mysite.com/venue/soldier-field tickets.aspx (This would be a live page that never expires.) I don’t know if that’s manipulating things a bit too much.
Intermediate & Advanced SEO | | Ticket_King1 -
Blocking Certain Site Parameters from Google's Index - Please Help
Hello, So we recently used Google Webmaster Tools in an attempt to block certain parameters on our site from showing up in Google's index. One of our site parameters is essentially for user location and accounts for over 500,000 URLs. This parameter does not change page content in any way, and there is no need for Google to index it. We edited the parameter in GWT to tell Google that it does not change site content and to not index it. However, after two weeks, all of these URLs are still definitely getting indexed. Why? Maybe there's something we're missing here. Perhaps there is another way to do this more effectively. Has anyone else ran into this problem? The path we used to implement this action:
Intermediate & Advanced SEO | | Jbake
Google Webmaster Tools > Crawl > URL Parameters Thank you in advance for your help!0 -
E Commerce site - removing discontinued items
We have been hit with a Panda penalty and the site has slowly been losing rankings since January, I've now realised that we have 4000+ page indexed in Google, but only 2000 live products. We have never deleted any of the pages with discontinued items, most of which were created when keyword stuffing and thin content reigned supreme - which explains the Panda penalty. But which is the best and quickest way to delete them from Google? We have already implemented a 'noindex' across all these pages, but as they are no longer in the 'crawlable' site, how will Google find them to know this? Would a 404 work any better - I'm not concerned about any link juice etc to/from these pages, I just want rid. I'm not sure if we can move all these pages into a dedicated directory which would allow us to use Google's Removal Tool - using it with the individual urls would be a mammoth task. Any advice would be most greatly appreciated.
Intermediate & Advanced SEO | | ElaineAllkids0 -
How to find affiliate sites linking to a competitor website?
Hello here, I am trying to understand the best way to find sites that are affiliate of a competitor, through link research. Typically our competitor's affiliates link to our competitor website via any of the following links: http://www.musicnotes.com/sheetmusic/ard.asp?SID=[aff_id]&LID=[link_id] http://click.linksynergy.com/link?id=[aff+id]&offerid=[off_id]&type=2&murl=http%3A%2F%2Fwww.musicnotes.com%2Fsheetmusic%2Fmtd.asp%3Fppn%3D[item_id] The first link looks much easier to find, so I have tried to find the first kind of links with Google by using the "link:" clause as follows: link:http://www.musicnotes.com/sheetmusic/ard.asp Or, similarly, by using Open Site Explorer. But I always get 0 results! It is weird because I know there are thousands of affiliates out there with the same tracking code. How's that possible? Why does it look impossible to find the sites I am looking for? Would you suggest any different approach? Any ideas, suggestions and thoughts are very welcome! Thank you in advance. Fab.
Intermediate & Advanced SEO | | fablau0 -
What's the best internal linking strategy for articles and on-site resources?
We recently added an education center to our site with articles and information about our products and industry. What is the best way to link to and from that content? There are two options I'm considering: Link to articles from category and subcategory pages under a section called "related articles" and link back to these category and subcategory pages from the articles: category page <<--------->> education center article education center article <<---------->> subcategory page Only link from the articles to the category and subcategory pages: education center article ---------->> category page education center article ---------->> subcategory page Would #1 dilute the SEO value of the category and subcategory pages? I want to offer shoppers links to more information if they need it, but this may also take them away from the products. Has anyone tested this? Thanks!
Intermediate & Advanced SEO | | pbhatt0 -
Best solution to get mass URl's out the SE's index
Hi, I've got an issue where our web developers have made a mistake on our website by messing up some URL's . Because our site works dynamically IE the URL's generated on a page are relevant to the current URL it ment the problem URL linked out to more problem URL's - effectively replicating an entire website directory under problem URL's - this has caused tens of thousands of URL's in SE's indexes which shouldn't be there. So say for example the problem URL's are like www.mysite.com/incorrect-directory/folder1/page1/ It seems I can correct this by doing the following: 1/. Use Robots.txt to disallow access to /incorrect-directory/* 2/. 301 the urls like this:
Intermediate & Advanced SEO | | James77
www.mysite.com/incorrect-directory/folder1/page1/
301 to:
www.mysite.com/correct-directory/folder1/page1/ 3/. 301 URL's to the root correct directory like this:
www.mysite.com/incorrect-directory/folder1/page1/
www.mysite.com/incorrect-directory/folder1/page2/
www.mysite.com/incorrect-directory/folder2/ 301 to:
www.mysite.com/correct-directory/ Which method do you think is the best solution? - I doubt there is any link juice benifit from 301'ing URL's as there shouldn't be any external links pointing to the wrong URL's.0 -
Should I robots block site directories with primarily duplicate content?
Our site, CareerBliss.com, primarily offers unique content in the form of company reviews and exclusive salary information. As a means of driving revenue, we also have a lot of job listings in ouir /jobs/ directory, as well as educational resources (/career-tools/education/) in our. The bulk of this information are feeds, which exist on other websites (duplicate). Does it make sense to go ahead and robots block these portions of our site? My thinking is in doing so, it will help reallocate our site authority helping the /salary/ and /company-reviews/ pages rank higher, and this is where most of the people are finding our site via search anyways. ie. http://www.careerbliss.com/jobs/cisco-systems-jobs-812156/ http://www.careerbliss.com/jobs/jobs-near-you/?l=irvine%2c+ca&landing=true http://www.careerbliss.com/career-tools/education/education-teaching-category-5/
Intermediate & Advanced SEO | | CareerBliss0