Comment by Animats
3 days ago
Embedding the model at chip fab time ought to be useful for robotics, driving, vision, and audio applications, at least. The training sets are good for years.
So they use 3 bit values. Is that current thinking? LLMs started at 32-bit floats, and have gradually shrunk. 8-bit floats seem to work. Is 3 bits pushing it?
If you're making your own chip, you might as well explore analog computation.