Comment by NitpickLawyer
11 hours ago
FYI they also released FP8 quants, and those should be faster on your setup (we have the same). As long as you keep kv at 16bit, FP8 should be close-to-lossless compared to 16bit, but with more context available and faster inference speed.
No comments yet
Contribute on Hacker News ↗