Comment by airstrike
8 hours ago
I mean, yes, but also because it's not as good as Claude today. Bit of a self fulfilling prophecy and they seem to be measuring the wrong thing.
4% of their tokens or total tokens in the market?
8 hours ago
I mean, yes, but also because it's not as good as Claude today. Bit of a self fulfilling prophecy and they seem to be measuring the wrong thing.
4% of their tokens or total tokens in the market?
> I mean, yes, but also because it's not as good as Claude today.
I'm not sure, sometimes GPT-5 Codex (or even the regular GPT-5 with Medium/High reasoning) can do things Sonnet 4.5 would mess up (most recently, figuring out why some wrappers around PrimeVue DataTable components wouldn't let the paginator show up and work correctly; alongside other such debugging) and vice versa, sometimes Gemini 2.5 Pro is also pretty okay (especially when it comes to multilingual stuff), there's a lot of randomness/inconsistency/nuance there but most of the SOTA models are generally quite capable. I kinda thought GPT-5 wasn't very good a while ago but then used it a bunch more and my views of it improved.
Out of curiosity, did you try asking Opus 4.1 as well?
> it's not as good as Claude today
In my experience this is not true anymore. Of course, mine is just one data point.
You're underestimating the amount of general population that's using ChatGPT. Us, people using it for codegen, are extreme minority.
Their tokens, they released a report a few months ago.
However, I can only imagine that OpenAI outputs the most intentionally produced tokens (i.e. the user intentionally went to the app/website) out of all the labs.