← Back to context

Comment by pmarreck

4 days ago

do you know why ollama hasn't updated its models in over a month while many fantastic models have been released in that time, most recently GLM 4.5? It's forcing me to use LM Studio which I for whatever reason absolutely do not prefer.

thank you guys for all your work on it, regardless

You know that if you go to hugging face and find a gguf page, you can click on Deploy and select ollama. It comes with “run” but whatever—just change to pull. Has a jacked name, but works.

Also, if you search on ollama’s models, you’ll see user ones that you can download too

GLM 4.5 has a new/modified architecture. From what I understand, MLX was really one of the only frameworks that had support for it as of yesterday. LM Studio supports MLX as one backend. Everyone else was/is still developing support for it.

Ollama has the new 235B and 30B Qwen3 models from this week, so it’s not as if they have done nothing for a month.

We work closely with majority of research labs / model creates directly. Most of the times we will support models on release day. There are sometimes where the release window for major models are fairly close - and we just have to elect to support models where we believe will better support a majority of users.

Nothing out of spite, and purely limited by the amount of effort required to support these models.

We are hopeful too -- where users can technically add models to Ollama directly. Although there is definitely some learning curve.

just so you know, you can grab any gguf from huggingface and specify the quant like this:

  ollama pull hf.co/bartowski/nvidia_OpenCodeReasoning-Nemotron-7B-GGUF:IQ4_XS