Comment by anentropic 3 months ago and: https://github.com/simplescaling/s1 8 comments anentropic Reply mettamage 3 months ago When you're only used to ollama, how do I go about using this model? davely 3 months ago I think we need to wait for someone to convert it into a GGUF file format.However, once that happens, you can run it (and any GGUF model) from Hugging Face![0][0] https://huggingface.co/docs/hub/en/ollama mettamage 3 months ago So this?https://huggingface.co/brittlewis12/s1-32B-GGUF 4 replies → fl0id 3 months ago you can load the safetensors with ollama, you just have to provide a modelfile. or wait for someone to do it. It will in theory also quantize it for you, as I guess most ppl cannot load a 129 GB model...
mettamage 3 months ago When you're only used to ollama, how do I go about using this model? davely 3 months ago I think we need to wait for someone to convert it into a GGUF file format.However, once that happens, you can run it (and any GGUF model) from Hugging Face![0][0] https://huggingface.co/docs/hub/en/ollama mettamage 3 months ago So this?https://huggingface.co/brittlewis12/s1-32B-GGUF 4 replies → fl0id 3 months ago you can load the safetensors with ollama, you just have to provide a modelfile. or wait for someone to do it. It will in theory also quantize it for you, as I guess most ppl cannot load a 129 GB model...
davely 3 months ago I think we need to wait for someone to convert it into a GGUF file format.However, once that happens, you can run it (and any GGUF model) from Hugging Face![0][0] https://huggingface.co/docs/hub/en/ollama mettamage 3 months ago So this?https://huggingface.co/brittlewis12/s1-32B-GGUF 4 replies → fl0id 3 months ago you can load the safetensors with ollama, you just have to provide a modelfile. or wait for someone to do it. It will in theory also quantize it for you, as I guess most ppl cannot load a 129 GB model...
fl0id 3 months ago you can load the safetensors with ollama, you just have to provide a modelfile. or wait for someone to do it. It will in theory also quantize it for you, as I guess most ppl cannot load a 129 GB model...
When you're only used to ollama, how do I go about using this model?
I think we need to wait for someone to convert it into a GGUF file format.
However, once that happens, you can run it (and any GGUF model) from Hugging Face![0]
[0] https://huggingface.co/docs/hub/en/ollama
So this?
https://huggingface.co/brittlewis12/s1-32B-GGUF
4 replies →
you can load the safetensors with ollama, you just have to provide a modelfile. or wait for someone to do it. It will in theory also quantize it for you, as I guess most ppl cannot load a 129 GB model...