Comment by falense
8 hours ago
I disagree. There is in fact a non-zero chance that we will get good enough models that are MOE optimized for desktop size hardware that can do a lot of the same things as the SOTA models. Im certainly crossing my fingers that the open-weights models continue improving. Engram from Deepseek for instance seems very interesting for a compute to memory offloading perspective.
https://www.reddit.com/r/LocalLLaMA/comments/1s0czc4/round_2...
No comments yet
Contribute on Hacker News ↗