Comment by gen220

4 days ago

Yea I mean this is the first publishable draft of a startup cooking on this.

I'm confident there are at least 1-2 OOMs of improvement to come here in terms of the (intelligence : wattage) ratio.

I really thought we were going to need to see a couple of dramatic OOM-improvement changes to the model composition / software layer, in order to get models of Opus 3.7's capability running on our laptops.

This release tells me that eventual breakthrough won't even be strictly necessary, imo.

The way I imagine it in 2-4 years we're going to be hit with a triple glut of better architecture, massive oversupply of hardware and potentially one or two hardware efforts like this really taking off. It's pretty crazy we're already 4 years in and outside of very niche / low availability solutions, it's still either GPU or bust

  • That's interesting! How do you see "oversupply of hardware" playing out?

    Is it because we stop doing ~2024-style, large-scale training (marginal returns aren't worth it)? Or because supply way outpaces the training+inference demand?

    AFAIU if the trend lines /S-curves keep chugging along as they are, we won't hit hardware oversupply for a long, long time without some sort of AI training winter.