Comment by htrp

10 months ago

> We firmly believe the moat for AI application is in the data and the data engineering today. At some point, the process of building custom LLMs might get so fast and easy that we’ll all return to building our own models. That simply isn’t the case today.

Customized small models will outperform larger general models for your specific use case.

I did expect that too, but it isn't happening reliably.

But small customized models seem to perform close to as well as large general ones.

No they wont. Any model you train now will be beat by GPT5 easily

  • Past performance is not a predictor of future performance.

    While it's possible the gap between GPT5 and 4 is as big as between 4 and 3, it's unlikely. The gap between 2 and 3 was much larger as the one between 3 and 4 (and similarly between 1 and 2).

    Also, it's not clear that GPT5 will do this in an *economical way* once the spigot of investor money stops.

  • Not for most human language, or anything that requires business-specific context where what's publicly available lags behind the state of the art your business cares about.

    And of course, not if you care about token throughput more than fancy abilities. Or price for that matter.

    So for many if not most businesses needs GPT-4 isn't the best tool out there, and GPT-5 is the canonical example of a vaporware right now.

  • I think the real power of customized small models will be running things on local hardware, except that we're in an awkward phase where the local hardware isn't quite beefy enough to run anything really useful yet. Maybe Apple will do something interesting in that space at WWDC.

    • Also not feasible. A network request to groq type machines will outperform your local hardware by such a huge amount that it wont make sense other than some very niche tasks

      6 replies →