Please be specific because outside of anecdotal blog posts by people who don’t know what they’re talking about it’s not true. Look at scaling laws, composite benchmarks from the epoch capability index, nothing at all suggests “model progress is slowing down”
Qwen3.6 9B is as good as GPT-4o and runs on my M2 MacBook Air. Models are getting stronger and less costly at the same time, but these are somewhat separate branches of research. Frontier labs are spending more because they are still getting marginal returns and there is more capacity to spend than there was a year ago.
Please be specific because outside of anecdotal blog posts by people who don’t know what they’re talking about it’s not true. Look at scaling laws, composite benchmarks from the epoch capability index, nothing at all suggests “model progress is slowing down”
Which indications are that?
The cost factors on the new models compared to the old models.
Qwen3.6 9B is as good as GPT-4o and runs on my M2 MacBook Air. Models are getting stronger and less costly at the same time, but these are somewhat separate branches of research. Frontier labs are spending more because they are still getting marginal returns and there is more capacity to spend than there was a year ago.
You are mixing cost and progress. It’s not because it’s more and more expensive that progress is slowing down by itself.
1 reply →
Investment dollars.
Source for that claim?
Nobody is releasing NEW models
…not only is this not true but it also doesn’t matter. Why would this indicate performance saturating?
What constitutes a NEW model for the purposes of calculating progress?
What? DeepSeekV3 just came out and is incredible for the price. Mythos is also half-released.
The standard networking connection has been called “Ethernet” for more than thirty years, so networking has stagnated, right?
4 replies →