Comment by iforgotpassword
19 hours ago
But why do you even have to do this fucking bullshit that you randomly stumble upon, while googling error message after error message, ending up in random github repos and issues?
And no, just because the three random cards you have work doesn't mean "everything works". Just tried an MI300A a few months ago... I just wanted to test ollama as this is one of the hottest applications for GPU acceleration now, it will surely be well supported right? First, the gfx version listed for it in the ollama docs is wrong - but OK, figured it out. Then Tried some random models with it, the only output it ever generates is GGGGGGGGGGGGG. Apparently only fp16 models work, nothing more quantized. So I pick one explicitly. Then it's slower than running on the cpu in the same system.
Thanks but no thanks; this cost me two days when Nvidia just works first try.
> But why do you even have to do this fucking bullshit
Because it's like 2-4 times cheaper than to go nvidia?..
> the three random cards you have
It's more than 3 random cards. I run 6900 XT, 7900 XTX, W7900 Pro, VII, VII Pro, Vega 56, Vega 64, 6800 XT, 5700 XT plus I've experimented with a 9950 iGPU, a 5xxx series iGPU and the only thing which didn't work was 3400g iGPU.
> Apparently only fp16 models work
fp8 works for me
> Because it's like 2-4 times cheaper than to go nvidia?..
That still makes nvidia way cheaper compared to the time wasted getting that shit to run plus debug an unknown amount of issues down the road. I told my boss "look I wasted two days on this and it's barely working, I could've done actual work here". Two other colleagues poked at it some more without success. going Nvidia was a no-brainer then.