What do you do about links to constantly moving pages?
-
One of the sites I work for is an employment site, they have a job database and the job pages tend to get links. The problem is that every time one of these jobs is filled, the job page goes away. What should I do to keep the value from these links?
-
I'd have to agree with this more! 301 to the category, that way once a new article/post/page/job etc appears under that article page, it will instantly have a boost from the PA/DA passed
-
Another thing you might want to consider is the use of rel="canonical". If you use the canonical tag on each job page to point back to the appropriate category it will help those category pages rank better rather than spreading the juice out among the individual job postings.
Matt Cutts recently did a video about this practice. He talks about product pages, but it should be the same in principle. You can find the video here: Canonical all product review pages as a single url.
You will still want to handle the missing pages with a 301 or 404, but there will be less concern about losing juice every time a job is filled. And as the video says, this is something to consider but it isn't a solution for everybody.
-
I would keep the pages but put a big red job taken accross the page, or if needed change the content completly.
this would give you more pages to play with when link sculpting also -
That's a good practice for small ads sites. As every of your jobs should be in a category, you should redirect the user to the category browsing page. Best page for the user and for googlebot too.
-
Hang on !
I would definitely avoid "301 back to the root page for jobs" or even a category page.
Over time, you are going to be creating a massive index of empty pages linking to a home page; that looks too spammy to me. If you want to be honest : 404 these pages- the job offer no longer exists, the page no longer exists --- you can personalise your 404 page to send the user to a relevant page
Honesty doesn't always pay though! To leverage the SEO benefits from these pages I would consider archiving the job listing, keeping the same url and just adding a message indicating that the post has been filled (an image will do)
That way, you’re keeping lots of unique content on your site and over time creating a log of pages.
To make these archived job pages useful to the user and to the search engines, dynamically add links to fresher job offers in the same category, company and town.
- Neil
PS Does this new SeoMoz feature now mean I'm now paying to give free advise ?
-
At some level they are user generated, but then they are put into the database and handled from there.
-
I was imagining that the vast majority of their pages would be user generated job listings. But I think I was incorrect.
-
It's actually surprising how many of the links are long term links, while they do sink off of front pages and whatnot, they are still there and even the mild value of them shouldn't go to waste.
-
Given the nature of Spencer's site, I wouldn't imagine that the incoming links to current job offers would have that long a life. So I wouldn't think that there'd be a mazzive pile up of incoming links getting 301'd.
-
Sure, I would 301 to .com/jobs/ or .com/[category]/ or whatever the main page is that will never go away. Depending on what you are doing, you may 301 to the root of your domain.
This really is a structural decision.
-
I definitely am not discounting your way of handling it... I think it's fantastic, especially because it's scalable. Where do you 301 the pages back to, the main category page?
-
Well I would hope that new data would be posted often so you would not have a bad ratio of old data to knew. Google is smart enough to know that some things date out such as products, events, job post, etc.
I have not noticed a penalty, but perhaps others can add comments to this.
-
Eventually, wouldn't a large ratio of your inbound links be pointed to pages that are 301'd to another page?
It just seems to me, that Google wouldn't think that is very 'natural', and perhaps would just feel that the majority of the content on the site is old/ outdated since most of the inbound links point to pages that don't exist anymore. (even if they are 301'd)
-
Yeah, I am starting to use this quite a bit with products moving off the site. No need to spill the juice
No because the 301 is dynamic. Not like adding to the .htaccess file. Also, make sure someone coding PHP does this as you need to make sure there are no white spaces before doing a header location or you will bomb the page.
Check your header to make sure you did the 301 correctly.
http://www.seoconsultants.com/tools/headers
Cheers
-
Hey Richard,
That's a useful script! Thanks!
Do you think in the case of running an employment site, those 301's would begin to rack-up frequently enough to get flagged?
[edit: I meant to add this below Richard Getz script]
-
Hey Spencer,
Is there a way you can dynamically pull the information (for the job) into the page.... so that once the job goes away, you can then change the informatino to be a new job?
The only catch to that, would be the URL structure, becuase obviously you would need to make the URL's generic, such as "/bay-county-seo-job" or something instead of mentioning the company.
On Distilled's recent conference call / webinar, Will discuess their project hiremarshall.com (I think that webinar would be of some help to you- and anyone else reading this).
Specifically, you could develop a model which keeps those pages live, so that the company uses that same page for all of their new job openings.
Donnie Cooper.
-
If these pages are database driven, you can check to see if the post is in the database, if not, then 301 back to the root page for jobs.
Run a PHP script that check the database TRUE = loads the page FALSE header redirect to root page (or whatever you want) and 301 the move.
if (!$_GET['post']) {
$location = "http://www.YourSite.com/jobs/";
header("HTTP/1.1 301 Moved Permanently");
header("Location: {$location}");
exit;Your developer will be able to actually write a valid script testing the page and either returning the job post or redirecting the page.
I hope that helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Links On Out Of Stock Product Pages Causing 404
Hi Moz Community! We're doing an audit of our e-commerce site at the moment and have noticed a lot of 404 errors coming from out of stock/discontinued product pages that we've kept 200 in the past. We kept these and added links on them for categories or products that are similar to the discontinued items but many other links of the page like images, blog posts, and even breadcrumbs have broken or are no longer valid causing lots of additional 404s. If the product has been discontinued for a long time and gets no traffic and has no link equity would you recommend adding a noindex robots tag on these pages so we're not wasting time fixing all the broken links on these? Any thoughts?Thanks
Technical SEO | | znotes0 -
3,511 Pages Indexed and 3,331 Pages Blocked by Robots
Morning, So I checked our site's index status on WMT, and I'm being told that Google is indexing 3,511 pages and the robots are blocking 3,331. This seems slightly odd as we're only disallowing 24 pages on the robots.txt file. In light of this, I have the following queries: Do these figures mean that Google is indexing 3,511 pages and blocking 3,331 other pages? Or does it mean that it's blocking 3,331 pages of the 3,511 indexed? As there are only 24 URLs being disallowed on robots.text, why are 3,331 pages being blocked? Will these be variations of the URLs we've submitted? Currently, we don't have a sitemap. I know, I know, it's pretty unforgivable but the old one didn't really work and the developers are working on the new one. Once submitted, will this help? I think I know the answer to this, but is there any way to ascertain which pages are being blocked? Thanks in advance! Lewis
Technical SEO | | PeaSoupDigital0 -
Can up a page
I do my best to optimize the on-page parameters for my page www.lkeria.com/AADL-logement-Algerie.php for the kw "aadl" but i can't understand what Ii'm doing wrong (i desapear 2 mounths ago). The page is optimize (title, description, h1, h2 etc.) few links with different ancers, but google put a spamy site www[dot]aadl[dot]biz in top 3 ratheer my page. Can you give me some advice to fix this issue? What I am doing wrong? Tanks in advance
Technical SEO | | lkeria0 -
Too Many On-Page Links - caused by a drop down menu
Many of our e-com sites we build for customers have drop down menus to help the user easily find products without having to click - Example: http://www.customandcommercial.com/ But this then causes the report to trigger too many on page links We do have a site map and a google site map So should I put code in place not to follow the drop down menu link items or leave in place?
Technical SEO | | spiralsites0 -
SEO Disasters - Links to Stories of Site Moves/Rebuilds Gone Bad?
Hello, everyone.. I am looking for some links to stories, articles, what have you describing medium to large complex sites that have moved, changed CMS, changed domains, etc, and ended up in a total SEO disaster. Really appreciate anything the community here has or can find! Thanks!
Technical SEO | | DuPont0 -
Too many internal links on one page
Hello All, I have just started using SEO moz. I had one quick question i would like answered. Currently SEOmoz is telling me that there are too many internal links. The recommendation is 100 links per page but the majority of my pages have 125+ links Will this effect the page when its crawled? Look forward to your comments. Thanks in advance
Technical SEO | | TWPLC_seo0 -
I have 15,000 pages. How do I have the Google bot crawl all the pages?
I have 15,000 pages. How do I have the Google bot crawl all the pages? My site is 7 years old. But there are only about 3,500 pages being crawled.
Technical SEO | | Ishimoto0