How effective is OSE in crawling press release links?
-
How effective is OSE in crawling press release links?
We have released a few press releases recently (over the last couple of months) and OSE doesn't seem to have found them.
-
Hey There,
That could be a possibility. It is hard to say definitively given the nature of web crawlers. They just crawl links as they see them in random succession, so a lot of factors come into play.
Best,
Nick
SEOmoz -
Our releases have appeared on big sites like the Financial Post.
Is it possible they just get buried under other news so OSE can't find them? I know Google indexes these pages, we get the alerts.
-
Hey there,
Just so you know, here's how we compile our index: - We grab the most recent index. - We take the top 10 billion URLs with the highest MozRank (with a fixed limit on some of the larger domains). - We start crawling from the top down until we've crawled 59,000,000,000 pages (which is about 25% the amount in Google's index).
Therefore, if the site is not linked to by one of these seed URLs (or one of the URLs linked to by them in the next update) then it won't show up in our index. Sorry!
We update our Linkscape Index every 4 weeks. Crawling the entire Internet to look for links takes 2-3 weeks, but our crawlers are always in motion. When we need to start processing, we grab all the data they have collected and start processing which can take up to 3 weeks to determine which of those links are the most important. You can see our most recently updated schedule here: http://seomoz.zendesk.com/entries/345964-linkscape-update-schedule
Linkscape focuses on a breadth-first approach. Therefore we almost always have content from the homepage of websites, externally linked-to pages, and pages higher up in a site's information hierarchy. However, deep pages that are buried beneath many layers of navigation are sometimes missed and it may be several index updates before we catch all of these.
If our crawlers or data sources are blocked from reaching those URLs, they may not be included in our index (though links that point to those pages will still be available). Finally, the URLs seen by Linkscape must be linked-to by other documents on the web or our index will not include them.
For now, the best thing you can do to help your domain become indexed is to work on link building for links from sites with high mozrank.
Best,
Nick
SEOmoz -
This all depends on where the press releases have been posted.
If you've got the urls of the sites they're on it may be worth looking at these in OSE to see if SEOmoz has them indexed. However, don't forget that the SEOmoz index is not the same as google's. Just because it's not showing on OSE doesn't mean that G hasn't seen it.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
The best way to track internal links in Google Analytics
Hi there, we are a retail business and we have invested in quality editorial content which sits in our Blog at ourwebsite.co.uk/blog/ The Blog links to the main site (an online store) and I want to track the 'value' of the blog by how many clicks the blog content generates back to the main store. At the moment we're using this code on the end of every link in the Blog: ?utm_source=Blog&utm_medium=Widget&utm_campaign=FromBlog Does this affect SEO and is there a better way of doing it? Thanks.
Reporting & Analytics | | Bee1590 -
URL open with double domain names when click on visit URL link in Google Analytics
I have configured Advance Filter to track the sub-domains traffic as follow : Filter
Reporting & Analytics | | gamesecure
Type: Custom filter > Advanced Field A: Hostname Extract A: (.*) Field B: Request URI Extract B: (.*) Output To: Request URI Constructor: $A1$B1 After that, I am able to see sub-domains record and View Full Page URL In Reports. But when I check reports in All page (e.g. Behavior >> All Pages) or selecting Landing Page as a Primary Dimension. Further I click on Icon given next to displayed Full URL to visit to same domain page, in browser the page
opened but the double domain name comes so page not open successfully. For example : In landing page list following URL given : www.sitegeek.com/compareHosting/arvixe_vs_hostgator If I click on icon given next the displayed URL, in browser following URL will
open https://sitegeek.comwww.sitegeek.com/compareHosting/arvixe_vs_hostgatorIs this First Domain with HTTPs, coming from Google Analytic 'View' where this is taken ? How Can I remove double domains? Thanks, Rajiv
0 -
Linking My Google analytics with visual.ly
Hello, I have a quick question: Is it secure and safe to link my Google Analytics account with Visual.ly here's the link: https://create.visual.ly/graphic/google-analytics I want to create easy-to-understand report from Google Analytics and if you have any experience with this service can you share it with me please. Thanks.
Reporting & Analytics | | JonsonSwartz0 -
Is Webmaster Tools Useless as a broken Link Detector?
Buongiorno from yes we still have free parking Wetherby UK!
Reporting & Analytics | | Nightwing
Ok when it comes to detecting broken links I'm getting really frustrated with webmaster tools. Now I'm probably going to end up with egg on my face with this one but here is an example of webmaster tools reporting a broken link which i cant find. http://i216.photobucket.com/albums/cc53/zymurgy_bucket/phantom-broken-links_zpsb74e1246.jpg Having trawled through the code i just cant see the knackered link? Is it a phantom report or is something usefull being detected here? Grazie tanto,
David1 -
My GWT tells me that verification has failed numerous occasions - will this stop my site being crawled?
I launched www.over50choices.co.uk 6 weeks ago and have had trouble with google indexing and crawling all pages. It tells me 143 submitted & 129 Indexed, but the site has 166 pages? It still shows the old home page image in GWT - which is v annoying! Whilst the site is verified by GA & HTML Tag, it tells me in the Verification section that "reverification failed" on numerous occasions - they seem correspond with when google trys to process the site map. Is this a coincidence ie verification fails when its trying to process the site map, which in turn is leaving me with an out of date site map and therefore not all my pages submitted or crawled? Or will this not effect the googles ability to crawl the site? Your help please. Ash
Reporting & Analytics | | AshShep10 -
Google Webmaster Tools - When will the links go away!?
About 9 months back we thought having an extremely reputable company build our client some local citations would be a good idea. You definitely know this citation company, but I'll leave names out. Regardless, it's our mistake to cut corners. Google Webmaster Tools quickly picked up these new citations and added them to the links section. One of these citation spawned a complete mess of about 60K+ links on their network of sites through ridiculous subdomains of every state in the country and so many other domain variations. We immediately went into remove mode and had the site's webmaster take down the bad links from their site. This process took about a month for outreach. The bad links (60K+) have not been on the spam site for well over 6 months but GWT still shows them in the "links to your site" section. Majestic, Bing, and OSE only displayed the bad links for a brief time. Why is webmaster tools still showing these links after 6+ months? We typically see GWT update about every 2 weeks, a month tops. Any ideas? Could a changed robots.txt on the bad site prevent Google from updating the links displayed in GWT? We have submitted to disavow, but Google replied with "no manual penalty". We even blasted the bad site with Fiverr links, in hopes that Google would re-crawl them. No luck with anything we do. We have patiently waited for way too long. The rankings for this site got crushed on Google after these citations. How do we fix this? Should we worry about this? Any advice would really help. Thanks so much in advance.
Reporting & Analytics | | zadro0 -
Open internal links in a new tab increase bonus rate?
Hello! This week I used a simple method to reduce my blog Google Analytics bounce rate. My blog all the posts are guides, in order to follow them, user need to download a zip file (same zip file). Otherwise they can't. Therefore I added a separate blog post to download all the necessary files. As a result of that I can reduce my bounce rate from 62-70% to 45-50% level. Now I'm thinking to open that zip file download page in a new tab. If I open my blog zip file download page, in a new tab. It will again increase my bounce rate? I reduced my bounce rate using that zip file download page. Thanks!
Reporting & Analytics | | Godad0 -
Posting on blog comments with anchor text on high ranked pages effective?
So i've identified some blogs which have a fairly high ranking and lots of traffic. They also allow anchor text in the name field. Does it make sense for me to comment on these blogs, or does google treat these with less authority that true page links? Any advice is greatly appreciated! TIA
Reporting & Analytics | | symbolphoto0