Past performance is not a predictor of future performance.
While it's possible the gap between GPT5 and 4 is as big as between 4 and 3, it's unlikely. The gap between 2 and 3 was much larger as the one between 3 and 4 (and similarly between 1 and 2).
Also, it's not clear that GPT5 will do this in an *economical way* once the spigot of investor money stops.
Not for most human language, or anything that requires business-specific context where what's publicly available lags behind the state of the art your business cares about.
And of course, not if you care about token throughput more than fancy abilities. Or price for that matter.
So for many if not most businesses needs GPT-4 isn't the best tool out there, and GPT-5 is the canonical example of a vaporware right now.
I think the real power of customized small models will be running things on local hardware, except that we're in an awkward phase where the local hardware isn't quite beefy enough to run anything really useful yet. Maybe Apple will do something interesting in that space at WWDC.
Also not feasible. A network request to groq type machines will outperform your local hardware by such a huge amount that it wont make sense other than some very niche tasks
Except nobody but groq has such type of machines, and the economics of cloud AI is very hard to make it works in practice. Offloading the capital cost (which is the hardest kind of cost to swallow for a company) to customers is very compelling business-wise.
Past performance is not a predictor of future performance.
While it's possible the gap between GPT5 and 4 is as big as between 4 and 3, it's unlikely. The gap between 2 and 3 was much larger as the one between 3 and 4 (and similarly between 1 and 2).
Also, it's not clear that GPT5 will do this in an *economical way* once the spigot of investor money stops.
Not for most human language, or anything that requires business-specific context where what's publicly available lags behind the state of the art your business cares about.
And of course, not if you care about token throughput more than fancy abilities. Or price for that matter.
So for many if not most businesses needs GPT-4 isn't the best tool out there, and GPT-5 is the canonical example of a vaporware right now.
I think the real power of customized small models will be running things on local hardware, except that we're in an awkward phase where the local hardware isn't quite beefy enough to run anything really useful yet. Maybe Apple will do something interesting in that space at WWDC.
Also not feasible. A network request to groq type machines will outperform your local hardware by such a huge amount that it wont make sense other than some very niche tasks
Except nobody but groq has such type of machines, and the economics of cloud AI is very hard to make it works in practice. Offloading the capital cost (which is the hardest kind of cost to swallow for a company) to customers is very compelling business-wise.
3 replies →
Network availability, latency, privacy, etc. many qualities to consider beyond model size and performance for applications.
1 reply →