Comment by brainless
1 day ago
I know this will sound strange, but SOTA model companies will eventually allow subscription based usage through third-party tools. For any usage whatsoever.
Models are pretty much democratized. I use Claude Code and opencode and I get more work done these days with GLM or Grok Code (using opencode). Z.ai (GLM) subscription is so worth it.
Also, mixing models, small and large ones, is the way to go. Different models from different providers. This is not like cloud infra where you need to plan the infra use. Models are pretty much text in, text out (let's say for text only models). The minor differences in API are easy to work with.
Wouldn't this mean SOTA model companies are incentivized not to allow subscriptions through third parties?
If all the models are interchangeable at the API layer, wouldn't they be incentivized to add value at the next level up and lock people in there to prevent customers from moving to competitors on a whim.
> If all the models are interchangeable at the API layer, wouldn't they be incentivized to add value at the next level up
Just the other day, a 2016 article was reposted here [https://news.ycombinator.com/item?id=46514816] on the 'stack fallacy', where companies who are experts in their domain repeatedly try and fail to 'move up the value chain' by offering higher-level products or services. The fallacy is that these companies underestimate the essential compexities of the higher-level and approach the problem with arrogance.
That would seem to apply here. Why should a model-building company have any unique skill at building higher-level integration?
If their edge comes from having the best model, they should commoditize the complement and make it as easy as possible for everyone to use (and pay for) their model. The standard API allows them to do just this, offering 'free' benefits from community integrations and multi-domain tasks.
If their edge does not come from the model – if the models are interchangeable in performance and not just API – then the company will have deeper problems justifying its existing investments and securing more funding. A moat of high-level features might help plug a few leaks, but this entire field is too new to have the kind of legacy clients that keep old firms like IBM around.
I do not know what that next level is to be honest. Web search, crawler, code execution, etc. can all be easily added on the agent side. And some of the small models are so good when the context is small that being locked into one provider makes no sense. I would rather build a heavy multi-agent solution, using Gemini, GLM, Sonnet, Haiku, GPT, and even use BERT, GlinER and other models for specific tasks. Low cost, no lock-in, still get high quality output.
AI labs are not charities and there is no way to make money offering unlimited access to SOTA LLMs. Even as costs drop, that will continue to be true for the best models in 2027, 2028 etc. - as demonstrated by the fact that CPU time still costs money. The current offerings are propped up by a VC bubble and not sustainable.
I agree but that is not the issue. See the really "large" models are great at a few things but they are not needed for daily tasks, including most coding tasks. Claude Code itself uses Haiku for a lot of tasks.
The non-SOTA companies will eat more of this pie and squeeze more value out of the SOTA companies.
They already do it’s called the API