Comment by karpathy
8 hours ago
It will work great with 40GB GPU, probably a bit less than twice slower. These are micro models of a few B param at most and fit easily during both training and inference.
8 hours ago
It will work great with 40GB GPU, probably a bit less than twice slower. These are micro models of a few B param at most and fit easily during both training and inference.
How low can this go? Can this run on a 5090 card (32GiB)?