Link analysis task
-
Hi mozzers,
I am currently working on a phd, and one of the professors asked me for help.
He would like to know how many Danish school websites (n=1500) links to a certain section of a government website (the relevant section has around 1600 pages). The problem is, that the government website is coded very poorly from an seo perspective with lots of strange URL variables, entailing OSE can't give valid data.
So, what would be the best way to check how many of the school websites link? Throw all 1500 website through Xenu, or is there a smarter solution? Maybe the link out feature on Bing?
Any suggestions will be greatly appreciate.
Thanks!
-
Exactely.
Sometimes I am not aware of an answer or question due to to much traffic here in the Q&A section ... I kind of loosing the overview
-
Thank you all for the answers - Incredible helpful!
And yes, this open q&a is great - I am sure it will dominate serious SEO discussions by the end of this year
-
Hey Thomas: I just sat through the video for the tool and it looks amazing! They say it uses 12 different sources for their link data(SEOmoz Linkscape, Majestic SEO, Google data, Yahoo, Technorati, etc.) THen it sounds like they give you a live verification. So to answer your question, it looks like it's doing both.
-
Sweet Petra! This site was completely off my radar. I really like the collective intelligence this Q&A is bringing to SEOmoz.
-
Thank you, Dejan.
Correct me if I am wrong - bur Majestic has a pretty deep crawl of the web right? We are examining websites with potential very few ingoing links.
-
I would suggest Majestic SEO to get solid backlinks and Excel to filter out URLs that do not belong to a desired section.
-
One more question: Is Cemper's tools based on an a priori crrawling like SEOmoz or a live crawling like Xenu?
-
Cool didn't knoe the tool could do it. ALways avoided the website as I do with most other keyword rich domain names
-
E.g. buying a 72 hours day pass - http://www.linkresearchtools.com/products-overview/ for Euro 20,- - worth the money - even Ann Smarty recommended the tool.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
A page will not be indexed if published without linking from anywhere?
Hi all, I have noticed one page from our competitors' website which has been hardly linked from one internal page. I just would like to know if the page not linked anywhere get indexed by Google or not? Will it be found by Google? What if a page not linked internally but go some backlinks from other websites? Thanks
Algorithm Updates | | vtmoz0 -
Footer menu links: Header tags or list items?
Hi, I would like to know header tags (h5 or h6) or list items ( ) works better for footer menu links for the best linking structure. Thanks
Algorithm Updates | | vtmoz1 -
Link reclamation and many 301 redirect to one URL
We have many incoming links to a non existing pages of a sub-domain, which we are planning to take down or redirect to a sub-directory. But we are not ready to loose pagerank or link juice as many links of this sub-domain are referred from different external links. It's going to be double redirect obviously. What is the best thing we can go to reclaim these links without loss of link juice or PR? Can we redirect all these links to same sub-domain and redirect the same sub-domain to sub-directory? Will this double redirect works? Or Can we redirect all these links to same sub-domain and ask visitors to visit sub-directory, manual redirection? How fair to manually redirect visitors? Any other options? Thanks, Satish
Algorithm Updates | | vtmoz0 -
After penguin 2.0, 20-25% drop sitewide, no google unatural links message, What could be causing it?
Hi,Since Penguin 2.0 we've taken a 20-25% knock but not recieved an unatural link message from Google. After sending a bunch of removal requests, I decided to submit a disavow file anyway two weeks ago and tried to make sure I rooted out some links that were built way back when our site started and link building best practice was a bit shadier. Analysis of our backlink profile points to about 40-50% links coming from general directories, wondering if perhaps their weight has been adjusted and this is why the drop occured? Having said that we have some high quality links from government sources and highly trusted sites so not too spammy. Can anyone shed some light or offer suggestions? Thanx
Algorithm Updates | | Mulith0 -
Content Caching Memory & Removal of 301 Redirect for Relieving Links Penalty
Hi, A client site has had very poor link legacy, stretching for over 5 years. I started the campaign a year ago, providing valuable good quality links. Link removals and creating a disavow to Google have been done, however after months and months of waiting nothing has happened. If anything, after the recent penguin update, results have been further affected. A 301 redirect was undertaken last year, consequently associating those bad links with the new site structure. I have since removed the 301 redirect in an attempt to detach this legacy, however with little success. I have read up on this and not many people appear to agree whether this will work. Therefore, my new decision is to start a fresh using a new domain, switching from the .com to .co.uk version, helping remove all legacy and all association with the spam ridden .com. However, my main concern with this is whether Google will forever cach content from the spammy .com and remember it, because the content on the new .co.uk site will be exactly the same (content of great quality, receiving hundreds of visitors each month from the blog section along) The problem is definitely link related and NOT content as I imagine people may first query. This could then cause duplicate content, knowing that this content pre-existed on another domain - I will implement a robots.txt file removing all of the .com site , as well as a no index no follow - and I understand you can present a site removal to Google within webmaster tools to help fast track the deindexation of the spammy .com - then once it has been deindexed, the new .co.uk site will go live with the exact same content. So my question is whether Google will then completely forget that this content has ever existed, allowing me to use exactly the same content on the new .co.uk domain without the threat of a duplicate content issue? Also, any insights or experience in the removal of a 301 redirect, detaching legacy and its success would also be very helpful! Thank you, Denver
Algorithm Updates | | ProdoDigital0 -
Does the page title keyword count in anchor text when link is web address?
If someone links to my plumbing site with this link as the anchor text: http://www.plumbers.com/austin-plumbers.html does the key phrase "austin plumbers" get counted in the anchor text by google or is this a sample of anchor text that google ignores? Thanks mozzers! Ron
Algorithm Updates | | Ron100 -
Difference between Google's link: operator and GWT's links to your sites
I haven't used the Google operator link: for a while, and I noticed that there is a big disparity between the operator "link:" and the GWT's links to your site. I compared these results on a number of websites, my own and competitors, and the difference seem to be the same across the board. Has Google made a recent change with how they display link results via the operator? Could this be an indication that they are clean out backlinks?
Algorithm Updates | | tdawson090 -
Since authorship markup requires a domain email, how can a community website allow users to link their Google+ profile?
It seems that Google now requires authors to have a valid email on the domain. This is easy for the traditional web publication. But what about community websites like SEOmoz? How can a community website allow users to link their Google+ profile? Will community websites like SEOmoz be required to 1. Give all users a domain email 2. Ask users to validate the email address with Google? Seems overly complicated.
Algorithm Updates | | designquotes0