Comment by davely
3 months ago
I think we need to wait for someone to convert it into a GGUF file format.
However, once that happens, you can run it (and any GGUF model) from Hugging Face![0]
3 months ago
I think we need to wait for someone to convert it into a GGUF file format.
However, once that happens, you can run it (and any GGUF model) from Hugging Face![0]
So this?
https://huggingface.co/brittlewis12/s1-32B-GGUF
I ran it, so far it seems like a pretty good model, especially locally.
oh god, this is terrible!
I just said "Hello!" and it went off the rails.
why how what? can you add a sample prompt with output ?
1 reply →
you can load the safetensors with ollama, you just have to provide a modelfile. or wait for someone to do it. It will in theory also quantize it for you, as I guess most ppl cannot load a 129 GB model...