Quality matters as well as speed though: reworking comes at a cost, so you really need to be tracking more than one metric. A lot of problems are caused by optimising for one metric above all else.
> Looks like Elon's finally giving up on XAI and just selling the compute
I don't think that's certain yet, but I do think that the open-source models like Gemma and Qwen are getting so good so fast that even Anthropic has real risk around the long-term value of their models and tooling.
Basically, if I'm Anthropic or xAI, I try to get revenue whenever and wherever possible and see what sticks. There's no value in playing for monopolistic control when everything is so volatile.
I don't know if it relates to the same data centers, but this also comes hours after several still recent Grok models were deprecated at short notice. Grok 4.1 Fast is the cheapest way to do research on X (cheaper than the X API!) and it's gone on May 15: https://docs.x.ai/developers/models - freeing up compute to sell?
The details are secret. It very well could be wasted GPU time but Anthropic could have made a killer offering as well.
I'm just speculating, but a particularly killer offering Elon wouldnt be able to refuse would be if Anthropic agreed to give them some training data / technology.
Giving Musk the benefit of the doubt, here's a thought experiment: It doesn't seem like any of the big labs in the US can keep a lead for more than 3 months. The Chinese models are closing in. Even if xAI comes up with the best model, so what?
On the other hand, power and compute are limited. Ridiculous as orbital compute sounds, land/power on earth is not easily scalable. There are too many limiting factors, chief among which in the US is regulation. But in space, if you make one satellite work, you just get more resources and launch more. This also leads naturally to Tesla's plan for a chip fab.
I don't think this is giving up. He's getting inside information on how Claude works, and a huge stream of Claude usage data. This will all inform future grok development, IMO.
And to continue with the analogy, he neither replaces the coach, nor the actual team players.
He just sits on the bench, paid for his - additional - role. Exactly the contrary of the Coinbase manager-IC, which is supposed to replace 2 jobs in 1.
This seems like an advertisement for an open source package
>Scale Python across 1,000 CPUs or GPUs in 1 second.
Burla is a high-performance parallel processing library with an extremely fast developer experience. Scale batch processing, vector embeddings, inference, or build pipelines with dynamic hardware.
Edit: Author comment was flagged dead. They work at burla which is a managed cloud service for parallelizing python
If effective AI enhanced SWEs can ship features in a week, the guys who ship 1 feature a quarter stand out more?
reply