City and state link stuffing in footer
-
A competitor has links to every state in the U.S., every county in our state and nearby states, and every city in those nearby states. All with corresponding link text and titles that lead to pages with thin, duplicate content. They consistently rank high in the SERPS and have for years. What gives--I mean, isn't this something that should get you penalized?
-
Thanks for your response, Will. It's small business (maybe 10 or 12 employees) at a single location. While they don't really impact me directly, it's particularly bothersome because they are in the advertising and marketing business. We tell clients not to do these things, but all around there are agencies that succeed using these tactics.
-
Hi There!
Unfortunately, as both Ben and Pau are mentioning, this absurd practice is still hanging around the web. While it's very unlikely the stuffed footer is actually helping this competitor to achieve high rankings, it is aggravating to think it isn't preventing them, either.
Your post doesn't mention whether this is actually a business model with physical local offices or is fully virtual, but what I have seen in cases like these is that big brands tend to get away with a great deal of stuff I would never recommend to a smaller brand. It begs the question: how can we explain this phenomenon?
In the past, I've seen folks asserting that Google is soft on big brands. There could be some truth in this, but we've all seen Google take a massive whack at big brand practices with various updates, so that really makes this an unsatisfying assertion.
Another guess is that big brands have built enough supporting authority to make them appear immune to the consequences of bad practices. In other words, they've achieved a level of power in the SERPs (via thousands of links, mentions, reviews, reams of content, etc.) that enables them to overcome minor penalties from bad practices. This could be closer to the truth, but again, isn't fully satisfactory.
And, finally, there's the concept of Google being somewhat asleep at the wheel when it comes to enforcing guidelines and standards, and whether or not that's kind of excusable given the size of the Internet. They can't catch everything. I can see this in this light, but at the same time, don't consider Google to have taken a proactive stance on accepting public reporting of bad practices. Rather, they take the approach of releasing periodic updates which are supposed to algorithmically detect foul play and penalize or filter it. Google is very tied to the ideas of big data and machine intelligence. So far, it's been an interesting journey with Google on this, but it is what has lead to cases exactly like the one you're seeing - with something egregiously unhelpful to human users being allowed to sit apparently unpunished on a website that outranks you, even when you are trying to play a fairer game by the rules.
In cases like this, your only real option is to hang onto the hope that your competitor will be the subject of an update, at some point in the future, that will lessen the rewards they are receiving in the face of bad practices. Until then, it's heads down, working hard on what you can do, with a rigorous focus on what you can control.
-
I've seen a lot of websites that do similar things and rank high on SERP's...
Sometimes this can be explained in some part by a good backlink profile, old domain / website, high amount of content (if the content is relatively original and varied), or because the niche is more receptive to this type of content (when it's something relatively common on your niche)... and other times simply makes no sense why things like this are working in Google for years without getting automatically or manual penalyzed.
Iv'e seen webs with so big keyword stuffing repeating a keyword about 500 times in the homepage, and being ranked in the top of Google for that keyword without seeing nothing internal or external of that website appart of this that can explain that awesome ranking. It's so frustrating knowing that this is penalized by Google and some of your competitors are doing it with impunity while you can't or at least you shouldn't...
-
Hi!
Yes, this absolutely should get them penalized. Unfortunately, I have also seen this work very well for different competitors in various niches. Regardless of what Google says, some old black-hat tactics still work wonders and these sites often fly under the radar. For how long is the question though. It still carries a heavy risk. If they are discovered, they can get a serious penalty slapped on them or at the very least get pushed pretty far down the SERPS. It's really just risk vs. reward. If you are like me, I work for a company that has a ton of revenue at stake, so I think of it like this.
It is much easier for me to explain to them why these thin, low-quality sites are ranking because of a loophole than it would be for me to explain why I got our #1 lead generating channel penalized and blasted into purgatory.
Usually, these sites that use these exact-match anchors on local terms look like garbage. So even if they are driving traffic, I often wonder how much of it is actually converting since the majority of their site looks like a collection of crappy doorway pages. It is still very frustrating to watch them succeed in serps though. I have the same issue.
You could always "try" to report them to Google directly. I do not know if this really works or if anchor-text spam would fall under one of their official categories to file it under, but you could try submitting a spam report here: https://www.google.com/webmasters/tools/spamreport.
I have no idea if this works or not though. Also as a side note, I would run their site through a tool like Majestic SEO or AHREFS and really dig on their backlink profile. I have seen a couple of instances where some spammy sites pulled off some nice links, so their success could also be attributed to those as well.
Hopefully, this helps, I know your pain.
-Ben
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Keyword stuffing as per the on-page grader
Hi Moz Community, I've recently become a Moz Pro user and I very impressed with the insights that it has to offer. However, I have been using the on-page grader to evaluate this page and it suggests that I am using the keyword "kiln dried logs" too many times and not to use more than 15 times. I have a slight dilemma because my product titles all contain this keyword and I wanted to get somebody's take on where the "15 repetitions" comes from and if it is better for me to strip this keyword out of my product titles to fall within the guidelines? Should I optimize just my main category page for this keyword at the expense of potentially losing traffic for my product pages? Any input would be much appreciated.
On-Page Optimization | | RicharCampbell0 -
Footer Section
As we have service oriented website, so what all links we need to add in our footer section?
On-Page Optimization | | Obbserv0 -
Impact of nofollow links
Does anyone know what the impact of a nofollowed link is on the ranking value any given page has to distribute? For example, if I have 2 links on a page, both followed, I know those links each distribute nearly 50% of the total ranking value the current page has to offer. However, if one of those links is nofollowed, does that automatically mean the other link gets the ranking value cast off by the nofollowed link? In other words, the single followed link now distributes nearly 100% of the ranking value the page has to offer? It seems to me I remember hearing this was not the case and that the ranking value a nofollowed link would have if it were followed just evaporates. This would mean the single followed link still only passes on around 50%...not 100%. Is the effect different if the links are internal vs. external? If any citations are available to justify knowledge here, that would be great. I know a lot of people have opinions about this subject, but I'm not sure anyone knows Google's position. Thanks!
On-Page Optimization | | RyanOD0 -
Are there to many internal links here?
This is our main Motorcycle Parts page. In my opinion there are to many links on this page. They have put links to all the top brands after products. Do you think having all these links could be hurting our rankings/ ability for google to crawl the site efficiently? The links are good for navigation, I just think there are way to many there like over 200. Here is the page you can see what i am talking about http://www.rockymountainatvmc.com/s/49/61/Motorcycle-Parts
On-Page Optimization | | DoRM0 -
Footer copyright year statement. good or bad
Hi, I see a lot of sites with a year copyright statment in the footer like Copyright 2011 - DomainName.com or Copyright 2002 - 2012 - Domainname.com since new year a lot of sites (founded before 2011) still have 2011 instead of 2012 in the footer. Do you think the date gives any signals to google? Should someone update the date or remove it completely? I would tend to remove it completely since the page date for google is submitted in the HTTP header. But maybe the info could be of any use for the user. Any best practices?
On-Page Optimization | | Autoschieber0 -
Anchor text on outbound links on a blog, relevancy detrimental or positive?
We have a blog related to computer support, and we have been using guest posts and promotion of those posts to boost freshness and rankings of the blog. We have been restricting outbound links to prevent words such as 'computer repair, 'computer support' etc, because we were under the impression that if we want to rank for those words, we should only allow INCOMING links with that anchor text, and that outbound links from the page, would rob the other parts of the site of the link juice this page provides. My question is, is this wrong? Should I freely allow outbound links on my blog page that contain anchor text that I my self am trying to rank for? Or was I correct initially? Current the anchor text is in 'related' industries, such as mobile apps, technology news, etc...things that google might think are 'related', but not exactly what the site is about.
On-Page Optimization | | ilyaelbert0 -
Does a www.domain.com/# count as a link?
Hi I thinking about consolidation some info from 5 pages onto 1 by using the hashtag at the end of the link to send people to the rigt section of the page. Does each link to the www.domain.com/# count as link so that I wont really gain any linkvalue by doing it?
On-Page Optimization | | home1110 -
Footers... Good or Bad?
I know in the past footers have been seen as a bad idea. I need to get up to date. Does having footer links that point to your own site with anchor texts keywords create problems? or is it safe to have? Will it do anything to our google rankings? Is it going to penalize us? As well, how many is acceptable, do I want all 77 links or have 15 with sub-categories?
On-Page Optimization | | HCGDiet0