At augmentcode.com, we've been evaluating Haiku for some time, it's actually a very good model. We found out it's 90% as good as Sonnet and is ~34% faster than sonnet!
Where it doesn't shine much is on very large coding task. but it is a phenomenal model for small coding tasks and the speed improvement is much welcome
90% as good as Sonnet 4 or 4.5?
Openrouter just started reporting, and it's saying Haiku is 2x as fast (60tps vs 125tps) and 2-3x less latent (2-3s vs 1s)
Fivetran acquired Census (reverse-etl) & Tobiko (dbt alternative).
I wonder who's next to really consolidate their platform play and compete with the old legacy MDM provider like Informatica. Data Observability or Catalog like Monte Carlo and Atlan. The whole Modern Data Stack has either died, acquired or merged by now. Wonder what's missing for Fivetran to IPO too.
I also wonder what this merge means for Airbyte who raised 150m at 1.5b in 2023.
Observability is a good guess, but I'd venture to guess that the conversations going on internally are about how to capture value across the entire stack. I wouldn't be surprised if we hear about them acquiring either a database/warehouse company and/or an analytics solution. Or vice versa, them getting acquired by a bigger player that wants to offer more connectors and data modeling functionality.
the MTR study is a joke. it surveyed only 16 devs. in the era of Sonnet 3.5
Can we stop citing this study
I'm not saying the DORA study is more accurate, but at least it surveyed 5000 developers, globally and more recently (between June 13 and July 21, 2025) which means using the most recent SOTA models
> I'm not saying the DORA study is more accurate, but at least it surveyed 5000 developers, globally and more recently
It's asking a completely different question; it is a survey of peoples' _perceptions of their own productivity_. That's basically useless; people are notoriously bad at self-evaluating things like that.
It didn't "survey" devs. It paid them to complete real tasks while they were randomly assigned to use AI or not, and measured the actual time taken to complete the tasks vs. just the perception. It is much higher quality evidence than a convenience sample of developers who just report their perceptions.
does that mean that Windsurf will only support OpenAI models going forward? I doubt OpenAI will pay to have users use Gemini/Claude? Especially as all of these Ai coding tools (Windsurf, Augment, Cursor) are heavily subsidizing the users.
I wonder what Anthropic makes of this. Windsurf was like a top 3 customers of them, might be a big revenue blow too?
Then Cursor takes on GitHub for the control of the repo.