Why does OSE show old data (Previous update results)?
-
Moz api started to show July 13 update results for my website. I checked it 2 days ago and saw all new established links and updated DA PA for July 27 update. But last 2 days both Moz Api and OSE main page show July 13 update results. Is there a maintenance or mismatching error between old and new databases?
-
It seems fixed now. I have just checked again and fortunately new update results appears.
-
We can certainly check on it if you provide the domain! I would recommend sending in the information to help@moz.com where you particular case can be discussed outside the public forum.
-
As i said before after 27 July update i checked several times and saw all new established links and increasing DA 5-6 points-up for my websites. But yesterday evening suddenly Moz Api and Ose started to show old data. Problem still persists.
-
We do not have a penalty system, but it's certainly possible that your metrics did not meaningfully change from one index release to another. It's all calculated based off the backlink profile we have for your site.
-
No way! Still it shows July 13 data for my all websites except one. That's very interesting. Is there any Moz penalty for some domains? I mean does moz have a penalty system?
-
Sorry for any trouble! While there can be a small time period after an index update in which the API is updated but the tools and apps built from it have not yet switched over to the new data, at this point everything from the July 27 update should be available from all sources. If you're seeing incorrect or inconsistent data, feel free to reach out to us at help@moz.com with the domains and URLs that you're seeing issues with, and we can look into it. Thanks!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Regarding Moz API token password update
Hi, In March we have updated password for MOZ API and used in our application it worked, but currently the updated password is not working and in the MOZ site the old password is shown and its active. We are using Legacy username and password.
API | | NickAndrews
We see that 5 tokens can be added for API, if we add 2 tokens both will be active.
We are currently using free services. Please help us resolve this issue.0 -
Moz Api Integration with Google Data Studio
I want to automate my competition research process. Right now my competition research is completely manual which is ok if working in a small project. Usually, start looking at Google for the best sites for a specific niche. In an excel sheet insert the URL in a column and start to put information in next columns, information such as DA, PA, backlinks, page speed, schemas validation and stuff like that. As I mentioned for a small project I usually take 10 websites as a reference to analyze how difficult can be rank a website. The problem starts when I have to scale when I have to evaluate 100 or 200 or even more websites in different niches or target location or industry. Does not make sense keep doing manually. So my question is there any way to connect the Moz API to Google Data Studio or Google Sheet
API | | Roman-Delcarmen2 -
Why the Feb 2018 update was so early?
Hi There! We are using Moz to compare our metrics to increase our SEO / SERP penetration. According to MOZ API Updates, it was mentioned that the next update will be on 26th Feb. But the update was early, could you please let us know the reason for the same. Why is there a discrepancy between the date mentioned for the Moz Update and the date of release? Thanks Malik Zakaria
API | | mzakaria0 -
API ACCESS to COMPETITOR DATA
Is there access with any of the available API's that I may be able to show my client's keyword competitive data for them vs their competitors (The ones they specify?). Perhaps in JSON format?
API | | Digitlead0 -
Two days since the supposed update
And still no update. Less a question, more a comment. Feeling a bit of deja vu here.
API | | pfrance3 -
Oct 14 2015 MOZScape update: none of DA are changed?
Hello, everybody. Today I noticed that finally the latest release of mozscape update has been posted. Now, I noticed that NONE of 25 campaigns websites NOR any of their competitors DA has changed. I do understand that DA can stay the same, but 25+25*3=100 websites domain authorities hasn't changed at all since August 4th (the date of previous mozscape release)? Or is this happening only to me? Please advise.
API | | DmitriiK2 -
September's Mozscape Update Broke; We're Building a New Index
Hey gang, I hate to write to you all again with more bad news, but such is life. Our big data team produced an index this week but, upon analysis, found that our crawlers had encountered a massive number of non-200 URLs, which meant this index was not only smaller, but also weirdly biased. PA and DA scores were way off, coverage of the right URLs went haywire, and our metrics that we use to gauge quality told us this index simply was not good enough to launch. Thus, we're in the process of rebuilding an index as fast as possible, but this takes, at minimum 19-20 days, and may take as long as 30 days. This sucks. There's no excuse. We need to do better and we owe all of you and all of the folks who use Mozscape better, more reliable updates. I'm embarassed and so is the team. We all want to deliver the best product, but continue to find problems we didn't account for, and have to go back and build systems in our software to look for them. In the spirit of transparency (not as an excuse), the problem appears to be a large number of new subdomains that found their way into our crawlers and exposed us to issues fetching robots.txt files that timed out and stalled our crawlers. In addition, some new portions of the link graph we crawled exposed us to websites/pages that we need to find ways to exclude, as these abuse our metrics for prioritizing crawls (aka PageRank, much like Google, but they're obviously much more sophisticated and experienced with this) and bias us to junky stuff which keeps us from getting to the good stuff we need. We have dozens of ideas to fix this, and we've managed to fix problems like this in the past (prior issues like .cn domains overwhelming our index, link wheels and webspam holes, etc plagued us and have been addressed, but every couple indices it seems we face a new challenge like this). Our biggest issue is one of monitoring and processing times. We don't see what's in a web index until it's finished processing, which means we don't know if we're building a good index until it's done. It's a lot of work to re-build the processing system so there can be visibility at checkpoints, but that appears to be necessary right now. Unfortunately, it takes time away from building the new, realtime version of our index (which is what we really want to finish and launch!). Such is the frustration of trying to tweak an old system while simultaneously working on a new, better one. Tradeoffs have to be made. For now, we're prioritizing fixing the old Mozscape system, getting a new index out as soon as possible, and then working to improve visibility and our crawl rules. I'm happy to answer any and all questions, and you have my deep, regretful apologies for once again letting you down. We will continue to do everything in our power to improve and fix these ongoing problems.
API | | randfish11 -
Suggestion - How to improve OSE metrics for DA & PA
I am sure everyone is aware at Moz, that although the Moz link metrics ( primarily I am talking about DA & PA) are good, there is a lot of room for improvement, and that there are a lot of areas where the metric values given to some types of site are well out of whack with what their "real" values should be. Some examples
API | | James77
www.somuch.com (Link Directory) - DA 72
www.articlesbase.com (Article Directory) - DA 89
www.ezinearticles.com (Article Directory) - DA 91 I'm sure everyone would agree that links from these domains are not as powerful (if of any value at all), as their DA would suggest, and therefore by definition of how moz metrics work, the sites these have links from such sites are also inflated - thus they throw the whole link graph out of whack. I have 2 suggestions which could be used to singularly or in conjunction (and obviously with other factors that Moz use to calculate DA and PA) which could help move these values to what they should more realistically be. 1/. Incorporate rank values.
This is effectively using rank values to reverse engine what google (or other engines) as a "value" on a website. This could be achieved (if moz were not to build the data gathering system itself), by intergrating with a company that already provides this data - eg searchmetrics, semrush etc. As an example you would take a domian and pull in some rank values eg http://www.semrush.com/info/somuch.com?db=us - where you could use traffic, traffic price, traffic history as a metric as part of the overall Moz scoring alogrithm. As you can see from my example according to SEMRush the amount of traffic and traffic price is extreamly low for what you would expect of a website that has a DA of 72. Likewise you will find this for the other two sites and similarly to pretty much any other site you will test. This is essentially because your tapping into Googles own ranking factors, and thereby more inline with what real values (according to Google) are with respect to the quality of a website. Therefore if you were to incorporate these values, I believe you could improve the Moz metrics. 2/. Social Sharing Value
Another strong indicator of quality the amount of social sharing of a document or website as a whole, and again you will find as with my examples, that pages on these sites have low social metrics in comparison to what you would normally associate with sites of these DA values. Obviously to do this you would need to pull social metrics of all the pages in your link DB. Or if this we to tech intense to achieve, again work with a partner such as searchmetrics, which provide "Total Social Interations" on a domain level basis. Divide this value by the number of Moz crawled pages and you would have a crude value of the overall average social scorability of a webpage on a given site. Obviously both the above, do have their flaws if you looked at them in complete isolation, however in combination they could provide a robust metric to use in any alogrithm, and in combination with current moz values used in the alogrithm I believe you could make big strides into improving overall Moz metrics.1