← Back to context

Comment by Sol-

3 days ago

Is that not just how scaling goes? It generally feels like the top models are mostly interchangeable and the one that came out at time t+1 will be better than earlier models from time t.

Grok 4 has probably been training when O3 was released, and now that Grok 4 is released, OpenAI is probably preparing O4, Google is preparing Gemini 3 and soon new SOTA benchmark scores will appear.

So it is impressive but not surprising, no? Whoever releases the latest model and has sufficient compute will be SOTA.