How/why is this page allowed to get away with this?
-
I was doing some research on a competitor's backlinks in Open Site Explorer and I noticed that their most powerful link was coming from this page: http://nytm.org/made-in-nyc. I visited that page and found that this page, carrying a PageRank of 7, is just a long list of followed links. That's literally all that's on the entire page - 618 links. Zero nofollow tags. PR7. On top of that, there's a link at the top right corner that says "Want to Join?" which shows requirements to get your link on that page. One of these is to create a reciprocal link from your site back to theirs.
I'm one of those white-hat SEOs who actually listens to Matt Cutts, and the more recent stuff from Moz. This entire page basically goes against everything I've been reading over the past couple years about how reciprocal links are bad, and if you're gonna do it, use a nofollow tag. I've read that pages, or directories, such as these are being penalized by Google, and possible the websites with links to the page could be penalized as well. I've read that exact websites such as these are getting deindexed by the bunches over the past couple years.
My real question is how is this page allowed to get away with this? And how are they rewarded with such high PageRank? There's zero content aside from 618 links, all followed. Is this just a case of "Google just hasn't gotten around to finding and penalizing this site yet" or am I just naive enough to actually listen and believe anything that comes out of Matt Cutts videos?
-
I guess you're right, but does that mean that Google wouldn't consider this a black-hat technique just because the link juice is divided by so many links? I thought it would actually be the opposite, that having only 5 or 10 links passing juice on a page would be okay, but something like 600 would be considered spam. I don't know, but perhaps Matt Cutts has said something about this specifically.
Regardless, have you, or has anyone here heard the phrase, "If your intention is to gain rankings in Google, then it's black-hat" Basically anything you do, such as listing a bunch of links like this without a nofollow link and asking to trade links, based on what I've gathered from Matt Cutts, is considered black-hat. If I'm wrong, please let me know.
But let's assume that everything you're saying is correct. How can we make the most of this situation? For me, for example, I actually went to Open Site Explorer and checked into followed external links and sorted them based on Page Authority. This was actually the most powerful link going to the site (I believe I was researching Quirky.com) based on what Moz was telling me. If what you're saying is true, then shouldn't Moz's algorithm be updated to take into consideration the amount of links on that page, then perhaps also they can take that and divide it with the Page Authority for the page and give us a new number based on that? That would probably be a much more accurate way of ranking pages based on how powerful they are, or how much link juice is going to them. Maybe there's a way to do that now and I'm just not aware. Do you have any strategies you use for this sort of thing, dividing link juice between the number of pages on the site?
-
I would add, at best case scenario its a PR7 page divided by 600+ links, so the actual page authority passed would be very small. Then consider your link would be at the bottom of that list so you would getting even less, if any.
-
Cyto,
Thank you again for another great response. You haven't put me off, quite the contrary. I really enjoy discussions like this because I actually work alone, as a one-man-show, and I don't get the opportunity to discuss SEO or online marketing with anyone really, let alone any experts. So personally, I rely a lot on Matt Cutts, and the info I get here at Moz, and other similar sites that I subscribe to on my RSS feed. Of course I also have a Pro account here at Moz and use it a lot for all of my clients.
I personally feel like Matt Cutts is the only person who knows what they're talking about, and the only person to trust. However, I have heard an SEO say before, "The things Matt Cutts may say is nice and all, but I rely more on the results that I actually find rather than just do what he says blindly." That makes sense, but I feel like that person was referring to doing black-hat stuff, until he gets caught. Regardless, my trust is still with Matt Cutts.
You said in your post (and it may have been a typo, I don't know) this: "My gut feeling is that, Google won't penalise a website who is an internet company made in NYC and listed on a non-profit organization website with a nofollow link. It seems like a natural fit." If, in fact, all of that was true, and they were using noFollow links, I believe this entire discussion is rather pointless because just through using the noFollow links on that page instead of 100% followed links, they would be in the clear as far as I'm concerned. I don't think there is any issues with trading links, anchor text, reciprocal links, etc as long as they are no followed. But in this case they are not no followed. They are all followed links. And they are asking for anchor-text optimized followed links. This is the key for me.
Now you may say to me, hey, it's an internal page with a directory and it's a non-profit .org site. Users may actually gain from this. However, Matt Cutts has said that any time you are doing things for the purpose of gaining rankings, it is considered black-hat. They can have that directory, and keep all the links, and provide this unique benefit to their users. However, in 2013 and 2014 and beyond, Matt Cutts has said this sort of example should have nofollow links because it won't change anything at all for the user experience, but it WILL cut down on spam because people would really be attracted to that page because of the linking opportunity. If the links were nofollow links, I doubt there would be nearly as many people excited about getting on that page. To me, this page is primarily for SEO purposes, in that the page will gain back links from the people who want to be listed, and the user experience is actually secondary. I have gained from Google that the user experience should be first, and the way to do that would be to nofollow all the links.
I am also aware that the algorithm doesn't necessarily take individual pages into account, but rather groups of pages with similar issues. For example, a page with massive links with optimized anchor text from PR1 or PR2 sites will be penalized, as we've seen from past updates. Other things like a text/html ratio should be above at least 15% from what I've seen, and the maximum amount of links per page shouldn't typically be higher than 200. This goes against all that. The craziest part about all of this is that I would expect this page to be somewhere around PR3. But it's PR7. WHY. That is the question. Are they being penalized, and just overpowering the penalization to get there? Has Google in fact placed this website or page on some sort of "white-list" that isn't included in typical algorithm roundups?
I'm actually to the point where I think I'm going to send Matt Cutts and e-mail and let's see if he responds. In the mean time, I would love to keep this discussion going! Cyto, I would love to hear another response, and if anyone else has anything else to add, or any other thoughts or theories (OR EXPERIENCE WITH THIS EXACT SORT OF THING) kindly add to this discussion! Thanks!
-
I don't think there ever is an individual who knows the right answers to everything when it comes to SEO. We're all exploring ideas, learning and sharing knowledge of our own findings and research.
Let's step outside the SEO world, throw away our knowledge and look at the website. Would you say, it is a website your client should be on? If your client is indeed an internet company made in NYC, shouldn't they be mentioned in NY Tech Meetup? From this perspective, I would say yes.
My gut feeling is that, Google won't penalise a website who is an internet company made in NYC and listed on a non-profit organization website with a nofollow link. It seems like a natural fit.
Second, I'm looking at opensiteexplorer, the page has a page authority of 78/100 and 174 root domains including some big power houses like the guardian, bloomberg, forbes. (didn't see any nofollow) I definitely think these are helping.I remember working on a client's webpage once, we optimised the page with rich content, clear call to actions and it was ranking on page 2, got 2 hyperlinks from the BBC and another high authority website and two weeks later, "boom", we were ranking on page 1, position 4.
Now let's explore the "black hat" technique. The core one would be the requesting of reciprocal links with the anchor text "Made in NYC" hyperlinked
You are right, that is "black hat" if I saw someone else do it, but in this scenario, I would go "that's fine". It all depends on the situation.
- A non-profit organisation focused on supporting new york technology community
- The webpage is specific to one and one thing only, listing internet companies made in NYC
- Their selection criteria focuses on active sites with 10k+ visits and ones solely made in NYC
You see, if I were to move this whole concept to a real world scenario, where nytm was a shop and they had a book listing other shops built in NYC, would you penalize them? "you sir, should not list such shops nor should other shops say you have a list of NYC build shops!"
In all honest, I don't see what they are doing as a big no no. I think things should be looked at as a case by case scenario, not to cluster everyone as a single group.
You mentioned why can't you create a .org page similar to this.. I say why not? Note their directory page isn't the core of what they are. This is just a single page of their entire entity and I think that plays a lot in their strength in reaffirming their presence in the web.
I realize you are frustrated, and all of us have our own thoughts. My thinking has always been to compare things to a real life scenario and focus more on creating great content that others will link to, rather than chase it myself. Sure, they might all use nofollow but they clicked through wanting to see my page and I'll let my rich content, site design and clear call to action turn them to a returnee.
Don't fret my friend. In a weird way, this is the perfect board to vent out and hear everyones thoughts and ideas. I hope my thoughts haven't put you off
-
Cyto,
I like your thinking on this one. This is where I was trying to go with it. But still, you asked many of the same questions that I asked. I realize we won't have a solid answer unless Matt Cutts himself speaks on this specific issue. However, I'm still left with unanswered questions. Here's a few points that are left standing:
- I realize there are billions, if not trillions, of websites and pages in existence. However, there are not billions of pages who are at a PageRank of 7. You can try to disregard their PageRank and tell me how it's going to be deprecated soon, or it's not accurate, or whatever. But regardless, they got that page to a PR7. If you think that doesn't matter, and it's not important, I'd like to see you try to get your page to PR7 and tell me how long it took you to do that. What I'm saying is I don't think they magically got to PR7 overnight, and I don't think that Google has missed this site. There's only so many PR9, PR8, and PR7's out there. What are the chances that they completely missed AND messed up on the PageRank for this site? The only other explanation I have for the PageRank is that they were white-hat for a long time, and then when they got to PR7, they flipped to this black-hat type of page. But I doubt that's the case. They're either still benefitting from black-hat techniques, OR we are misjudging this site and Google actually does think it deserves a PR7.
- Try thinking about it like this: yes, this page is practicing many things that are straight-up black-hat, things that Matt Cutts has publicly and openly said is considered spam. Just simple things like a text/html ratio, or a certain number of links per page, or asking to trade links, or having massive links without nofollowing any of them. What if Google saw this page and said, wow this is a black-hat page, let's penalize them. And let's assume that this page is penalized. But what if all the sites on there are linking back to this page, and therefore all the link juice from the other pages pointing back to this page is basically that much more powerful than the penalization that it's basically overpowering the penalization with more back links, thus bringing them to a net PR7? The question here is: can you overpower Google's penalization with more bad back links?
- Looking deeper into the whole .org/non-profit/maybe Google likes these types of pages, perhaps they do and we're all just wrongly assuming things. In this case, I agree with Cyto, this page could be unique and it does benefit the user. However, isn't this the exact scenario that Matt Cutts has told us to implement a noFollow tag? I believe he has said repeatedly, if you must link to another site and you're not sure about it, just put on a noFollow tag. If you have reciprocal links, no need to get rid of them, just simply nofollow the links. It's this sort of thing that is giving me trouble fully accepting that this is a good page and Google likes them. And IF Googles does like this page, and the PR7 is deserved, and the followed links are fine, then I SHOULD try to get my client a link on this page. But I suppose there is a risk because we won't 100% know for sure unless Matt Cutts says so.
- Diving deeper into the "Google may like this sort of page" for the reasons you stated, it sort of contradicts what has already been said from Matt Cutts. For example, if I put a link in a press release back to my homepage, there is some value in that link to the user because it makes it easier for the user to visit that page simply by clicking instead of typing in the URL. In this case, all PR links have been nofollowed across the internet. You can use this same excuse to use a link, and say it creates value, but Google is telling us to noFollow these links. Especially when talking about a "directory" specifically, I have read that Google is shutting these sites down completely. However, we are left wondering if this specific site is on some sort of "white-list". In that case, the first-person to create a "directory of white-listed directories of followed links" I'm sure will be quite successful with that page.
- What is stopping me from creating a .org page similar to this? Why can't I build a page up to PR7 and openly exchange links with people? The biggest thing stopping me from even thinking about something like that is because I am assuming this only worked 5+ years ago. Regardless, I have a client who sells a few unique products, and one of their competitors is Quirky.com who led me to finding this page because they have a back link from this page. The problem I'm seeing is that Quirky.com is benefitting from a link on this page, and I'm worried about joining it due to a potential penalization. In this case, Quirky doesn't really have to worry about anything because they have so many links, and they're established. But if I wanted to get the same link as them, I have to worry. This is the sort of thing that makes it hard to compete with the big players. Not that I think this client is on par with them, but I just get the feeling that they're allowed to do more than we are. Perhaps I'm wrong, but it's the feeling I have.
- It's getting harder and harder for me to find white-hat followed link opportunities. It seems like everywhere I go, the link is going to be nofollowed. Other people's websites, they want to noFollow the link. Guest blog posts, they want to noFollow the link. Press releases are all nofollowed now. The case is either the link is noFollowed, or you risk penalization on a followed link. This is the corner I feel I'm getting pushed into.
- I learned a while back from an SEO that links are the most powerful form of currency in the SEO world. A link is the number one most powerful way to get higher up in the rankings, for the reason that it is basically a sign of saying "this site is trustworthy and worthwhile to check out" and Google puts those things together to say they are worthy of higher rankings. And it all makes sense to me, and I haven't seen anything to tell me otherwise. If I'm wrong and I missed something, let me know. I mean, it's great to put out unique content and all that, but what is the point of the guest post or the press release if there's no indication that you wrote it or that it has anything to do with your site? What is it worth at that point if there is no link included? I understand the organic side that some people may literally read it and visit your site off that, but that's an inefficient way of doing things. I'm down with "link-earning" but only if I can actually earn a followed link. What's the point of a link-earning process if you don't earn the link, know what I mean? It just seems like everything is going this way of noFollowing links, or you have to worry about a penalization. And before you say it, I am aware that it's less than 20% of all links that are noFollowed, but still, this is the feeling I'm getting. (That number may be higher now that all Press Release links are no followed, not sure)
- I'm really not trying to do anything black-hat. I'm trying to do white-hat stuff here, but with the purpose of accelerating my client's process of getting higher in rankings. Listen, I'm doing all the other stuff well, it's just this whole link-building/earning aspect is tough and it seems like 2014 is going to be much harder than previous years.
What are your thoughts on these points?
-
Cyto, that is one of the best analyses that I have seen in a long time.
Thumbs up!
-
Maybe it has all to do with the site itself and the "human" approach Google is taking.
- The site is a non-profit organisation supporting the New York technology community. Domain is .org
- Let's review the page, in all reality such a page is useful - it is showing internet companies made in NYC from a non-profit organisation. What other format would you take? Sure it only has links and looks spammy but it isn't a spam page. It has a purpose and exists in a format that is acceptable by its users - plus, it isn't asking for money. If you were to look at this from a human eye, you would go "this isn't spam"
- Taking the human approach again, if my friend was looking for a collection of internet companies in NYC and he had a research pile with 10 documents and one of them was a handwritten version of this page, shouldn't it be there? Or should it be filed under his research pile of 500 documents? Maybe, we need to see Google differently, it's more complex than a "if statement".
- Maybe reciprocal links to non-profit organisations are viewed differently. A good Samaritan would go "yes" and maybe Google is taking a human approach and going "let me help you as you are a good non-profit organisation"
- Now the nofollow aspect, this is a technical element and I agree, shouldn't it exist? But again, maybe just maybe Google is seeing the site and going "I support such organisations" and the support this can give to other sites isn't so bad.
- Let's take Wikipedia. Would Google punish wikipedia if they didn't have a nofollow? Was the introduction of nofollow done due to Google or a decision Wikipedia made? Maybe it was Wikipedia who noticed users abuse their site and thus introduced the nofollow.
What I'm trying to say is that, maybe Google is evolving to be more complex/human like. It's doing a moral mind on decisions.
On the flip-side, it's just a software that forgot to notice the site amongst the millions of websites out there and in time, it will capture them.
-
I agree with Tim, Moz's PA/DA is based purely on links and does not consider spam pages, google's PR (that is shows to the public) is very unreliable.
Google may have missed those links , or maybe have simply devalued them, we just don't know. But what we do know its the type of practice that google is tying to stop, so someday they might do a big penguin update an start penalizing sites with back links like that. My own rule would be if is easy for humans to spot then it either easy for the algo to spot or some day the algo will spot it.
Yahoo is directory is different as you pay to be considered to get on their list, so its not a direct pay for a link. But I think because yahoo is a trusted site they get away with it, I don't think a no name directory would get away with the same trick. Also I would question how good a yahoo directly really is (some people this its worth the money, others think its not)
-
One thing to consider is that just because Open Site Explorer found the link and assigned it a high PA/DA and the PageRank toolbar shows a high PR score doesn't mean that Google is actually passing link strength through those links. I wouldn't look at site's PageRank as a definitive answer to Google's thoughts on the site's quality.
Page Authority and PageRank are essentially just equations that consider the quantity and strength of links pointing to a page, but they don't say anything about the spamminess of a page's content. It's possible that Google has already noticed the hundreds of followed links coming out of this site, marked the page as spam and devalued all those links. All of that could be done behind the scenes without any changes to the site's PageRank and without any public notice in regards to the site's spamminess.
With that said, it's also possible that Google hasn't caught on to the spam tactics of this site and those links do still have value. The internet is a huge place and as we all know, Google is definitely capable of letting spammy sites slip through the cracks. Unfortunately, Google will probably never pull back the curtain for us, so we won't know how they treat cases like this.
As for whether you should get the link for your client. I would lean towards no, but wouldn't completely rule it out. There is some risk involved, but if your client already has a decent link profile, one link from a spammy site probably won't hurt them and if Google hasn't devalued those links, it could potentially help them.
If you do choose to go after the link, I would first have an open discussion with your client about the potential risks and rewards. If they want to be really aggressive, they might want you to go for it. If they'd rather play it safe, probably better to pass.
Tim
-
So what I've gathered so far is that you're saying that indeed this page is considered "black-hat" for the reasons mentioned above, and that eventually the site, as well as possibly the sites that are listed on this page, could all receive a penalty, and that I should stick to white-hat strategies.
But let's take this a step further. We are simply assuming it is "black-hat" page based on things we've heard and have accepted as fact. However, what if Google and Matt Cutts actually see this page as something different and actually perhaps "white-hat"? I may ask then, how could they see it as white-hat if it's breaking all these other rules? At that point I would look at something like the Yahoo directory. The one where you pay $300 per year to get one backlink. I feel like these two sites are fairly similar and breaking similar rules.
So for some reason, Google likes the Yahoo directory and lets them do what they want. Perhaps they are also putting this webpage in the same boat as the Yahoo directory. Matt Cutts' excuse may be something like, "Well, this page has been around for a while, and it's actually quite a unique page in that there are no other directories like this on the internet, and therefore is actually providing a benefit to the user in the form of a directory of startups in New York." But then I would ask, why not nofollow all the links, because isn't that the whole point of the nofollow tag?
He also may say something like, they have strict guidelines and rules to allow people into the directory. After all, it's not open to anyone, only startups in NYC who get >10k visitors per month to your site. But still, why not the nofollow links? They are also blatantly asking to trade links. And after all that they still have a PR7?
The next step I ask is, should I try to get my client on this list to, so that he may benefit from the directory and the PR7 backlink? I'm thinking twice about it because I don't want to wake up one day and be penalized because I have a link on that directory, and I have a reciprocal link going back to that site. If you ask me, this page is a perfect example of what Google doesn't want, and yet Google is rewarding them. So I'm not sure if perhaps I'm the one who is wrong and perhaps Google actually likes this site because it may be unique to some extent.
I would love to hear some sort of official response on this from someone at Moz, as well as any other people here who are familiar with this sort of situation, and any success on pages like this that we are assuming are black-hat. I would love to have someone from Moz actually visit that page and give me their analysis on if the page is breaking any rules, and why it has the PageRank it does.
-
Thanks for the laugh.. Gagan. That is a really funny quote from Buffet.
I am going to go make a page like this just so my competitors will get their panties in a wad.
-
Hello Trenton,
There are several Million sites in web space as of now and think of hundred multi trillion pages
If a site doing wrong and you doing correct - patience rewards. Sooner or later - you will see rewards coming in your favor only
Warren Buffet said a famous quote (though all his quotes are famous) - "No mater how great are your efforts in business - some things just take time. You cant produce a baby in 1 month time by making 9 women pregnant"
So - my advise is to ignore that poor strategy adopted by competition or even that site - sooner or later it will get on to a dead end. If you still want it to happen fast
you may submit an anti-spam report to Google about that web page :- https://www.google.com/webmasters/tools/spamreportform?hl=en&spamurl=https%3A%2F%2Fwww.google.co.in%2Fwebhp%3Fsourceid%3Dchrome-instant%26espv%3D210%26ie%3DUTF-8
-
Yo have started to learn SEO and you will find such frustrating things again and again where People are doing things against rules of Google and getting better results than those who are doing legitimate things!
But dont get frustrated, stick with things your are doing and learning. Websites that do tricky things will not going to last for long.
You will see such websites tanked very very shortly by Penguin (3 or 3.1 :-))!!!!!!
Regards
-
I am totally flustered by this example and more like this.
on most of my client's projects I see gray and black SEO sites that are leading the serps.
I try to follow the best practice advices and keep a clean shop.
cannot wait to get more info on this.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Will pillar posts create a duplication content issue, if we un-gate ebook/guides and use exact copy from blogs?
Hi there! With the rise of pillar posts, I have a question on the duplicate content issue it may present. If we are un-gating ebook/guides and using (at times) exact copy from our blog posts, will this harm our SEO efforts? This would go against the goal of our post and is mission-critical to understand before we implement pillar posts for our clients.
White Hat / Black Hat SEO | | Olivia9540 -
Robots.txt file in Shopify - Collection and Product Page Crawling Issue
Hi, I am working on one big eCommerce store which have more then 1000 Product. we just moved platform WP to Shopify getting noindex issue. when i check robots.txt i found below code which is very confusing for me. **I am not getting meaning of below tags.** Disallow: /collections/+ Disallow: /collections/%2B Disallow: /collections/%2b Disallow: /blogs/+ Disallow: /blogs/%2B Disallow: /blogs/%2b I can understand that my robots.txt disallows SEs to crawling and indexing my all product pages. ( collection/*+* ) Is this the query which is affecting the indexing product pages? Please explain me how this robots.txt work in shopify and once my page crawl and index by google.com then what is use of Disallow: Thanks.
White Hat / Black Hat SEO | | HuptechWebseo0 -
Mobile Redirect - Cloaking/Sneaky?
Question since Google is somewhat vague on what they consider mobile "equivalent" content. This is the hand we're dealt with due to budget, no m.dot, etc, responsive/dynamic is on the roadmap but still a couple quarters away but, for now, here's the situation. We have two sets of content and experiences, one for desktop and one for mobile. The problem is that desktop content does not = mobile content. The layout, user experience, images and copy aren't the same across both versions - they are not dramatically different but not identical. In many cases, no mobile equivalent exists. Dev wants to redirect visitors who find the desktop version in mobile search to the equivalent mobile experience, when it exists, when it doesn't they want to redirect to the mobile homepage - which really isn't a homepage it's an unfiltered view of the content. Yeah we have push state in place for the mobile version etc. My concern is that Google will look at this as cloaking, maybe not in the cases where there's a near equivalent piece of content, but definitely when we're redirecting to the "homepage". Not to mention this isn't a great user experience and will impact conversion/engagement metrics which are likely factors Google's algorithm considers. What's the MOZ Community say about this? Cloaking or Not and Why? Thanks!
White Hat / Black Hat SEO | | Jose_R0 -
Recovering from Google Penguin/algorithm penalty?
Anyone think recovery is possible? My site has been in Google limbo for the past 8 months to around a year or so. Like a lot of sites we had seo work done a while sgo and had tons of links that Google now looks down on. I worked with an seo company for a few months now and they seem to agree Penguin is the likely culprit, we are on page 8-10 for keywords that we used to be on page 1 for. Our site is informative and has everything in tact. We deleted whatever links possible and some sites are even hard to find contact information for and some sites want money, I paid a few a couple bucks in hopes maybe it could help the process. Anyway we now have around 600 something domains on disavow file we out up in March-April, with around 100 or 200 added recently as well. If need be a new site could be an option as well but will wait and see if the site can improve on Google with a refresh. Anyone think recovery is possible in a situation like this? Thanks
White Hat / Black Hat SEO | | xelaetaks0 -
Should we remove our "index" pages (alphabetical link list to all of the products on the site)?
We run an e-commerce site with a large number of product families, with each family having a number of products within it. We have a set of pages (26 - one for each letter A-Z) that are lists of links to the product family pages. We originally created these pages thinking it would aid in discoverability of these pages to search engines, of course as time has gone on, techniques like this have fallen out of favor with Google as it provides negligible value to the user. Should we consider removing these pages from the site overall? Is it possible that it could be viewed by Panda as resembling a link farm? Thanks in advance!
White Hat / Black Hat SEO | | ChrisRoberts-MTI1 -
Local Doorway Pages
Based on what I've read, setting up localized landing pages ie: /web-design-atlanta, web-design-nyc, /web-design-chicago, etc especially with duplicate content is a big no-no. Remarkably, 2 of our competitors are doing it, (they are just swapping out the locations), and it's working. They don't even have office addresses or local phone numbers listed. They are on the first page for multiple location based searches ("web design nyc", "web design atlanta", etc.). I thought Google penalized for this, or at least didn't index the content. What gives? Am I misinterpreting Google's AUP? Can I report them? If it's legal, we should be doing it as well.
White Hat / Black Hat SEO | | CsmBill0 -
Page not being indexed or crawled and no idea why!
Hi everyone, There are a few pages on our website that aren't being indexed right now on Google and I'm not quite sure why. A little background: We are an IT training and management training company and we have locations/classrooms around the US. To better our search rankings and overall visibility, we made some changes to the on page content, URL structure, etc. Let's take our Washington DC location for example. The old address was: http://www2.learningtree.com/htfu/location.aspx?id=uswd44 And the new one is: http://www2.learningtree.com/htfu/uswd44/reston/it-and-management-training All of the SEO changes aren't live yet, so just bear with me. My question really regards why the first URL is still being indexed and crawled and showing fine in the search results and the second one (which we want to show) is not. Changes have been live for around a month now - plenty of time to at least be indexed. In fact, we don't want the first URL to be showing anymore, we'd like the second URL type to be showing across the board. Also, when I type into Google site:http://www2.learningtree.com/htfu/uswd44/reston/it-and-management-training I'm getting a message that Google can't read the page because of the robots.txt file. But, we have no robots.txt file. I've been told by our web guys that the two pages are exactly the same. I was also told that we've put in an order to have all those old links 301 redirected to the new ones. But still, I'm perplexed as to why these pages are not being indexed or crawled - even manually submitted it into Webmaster tools. So, why is Google still recognizing the old URLs and why are they still showing in the index/search results? And, why is Google saying "A description for this result is not available because of this site's robots.txt" Thanks in advance! Pedram
White Hat / Black Hat SEO | | CSawatzky0 -
How to Handle Sketchy Inbound Links to Forum Profile Pages
Hey Everyone, we recently discovered that one of our craft-related websites has a bunch of spam profiles with very sketchy backlink profiles. I just discovered this by looking at the Top Pages report in OpenSiteExplorer.org for our site, and noticed that a good chunk of our top pages are viagra/levitra/etc. type forum profile pages with loads of backlinks from sketchy websites (porn sites, sketchy link farms, etc.). So, some spambot has been building profiles on our site and then building backlinks to those profiles. Now, my question is...we can delete all these profiles, but how should we handle all of these sketchy inbound links? If all of the spam forum profile pages produce true 404 Error pages (when we delete them), will that evaporate the link equity? Or, could we still get penalized by Google? Do we need to use the Link Disavow tool? Also note that these forum profile pages have all been set to "noindex,nofollow" months ago. Not sure how that affects things. This is going to be a time waster for me, but I want to ensure that we don't get penalized. Thanks for your advice!
White Hat / Black Hat SEO | | M_D_Golden_Peak0