Comment by Sol-
3 days ago
Is that not just how scaling goes? It generally feels like the top models are mostly interchangeable and the one that came out at time t+1 will be better than earlier models from time t.
Grok 4 has probably been training when O3 was released, and now that Grok 4 is released, OpenAI is probably preparing O4, Google is preparing Gemini 3 and soon new SOTA benchmark scores will appear.
So it is impressive but not surprising, no? Whoever releases the latest model and has sufficient compute will be SOTA.
Meta had enough compute I think. No SOTA though.