Comment by dragonwriter
14 days ago
The RAM requirements for storing the parameters are set by the total, not active, parameters. Llama4 Scout is 109B model, so, at Int4 quantization, it will require ~55GB for the model. With 64GB, you could probably run it, but I would imagine not with a very large context size.
No comments yet
Contribute on Hacker News ↗