Comment by cyberax
13 hours ago
Uhmm... I have a local Ollama setup on Linux+AMD, and it was only a bit more involved than this sample. And only because I wanted to run everything in a container.
If you mean that you can't just run the largest unquantized models, then it's indeed true.
No comments yet
Contribute on Hacker News ↗