Comment by EagnaIonat
2 days ago
The more recent LLMs work fine on an M1 mac. Can't speak for Windows/Linux.
There was even a recent release of Granite4 that runs on a Raspberry Pi.
https://github.com/Jewelzufo/granitepi-4-nano
For my local work I use Ollama. (M4 Max 128GB)
- gpt-oss. 20b or 120b depending on complexity of use cases.
- granite4 for speed and lower complexity (around the same as gpt20b).
No comments yet
Contribute on Hacker News ↗