Comment by api

10 months ago

Hmm... in that case the analogy with AI is even better. This sounds like neural networks before things like deep learning and the transformer architecture -- before we figured out how to scale them. Turns out this did require some innovations. It wasn't just a matter of making a bigger model.