Comment by dragonwriter
3 months ago
The RAM requirements for storing the parameters are set by the total, not active, parameters. Llama4 Scout is 109B model, so, at Int4 quantization, it will require ~55GB for the model. With 64GB, you could probably run it, but I would imagine not with a very large context size.
No comments yet
Contribute on Hacker News ↗