Comment by GTP
19 hours ago
Well, with the Chinese AI divisions becoming a serious competitor more and more, they should start caring about their reputation. Otherwise people will go to the cheaper competitor.
19 hours ago
Well, with the Chinese AI divisions becoming a serious competitor more and more, they should start caring about their reputation. Otherwise people will go to the cheaper competitor.
Yea I am more or less done with these big providers. I'm running local primarily now. These constant screw ups, not caring about customers, political issues, it's just not worth it for me. I get some people are hooked on vibe coding but the latest wave of small models I'm good for my needs.
What do you use now? How much ram do you have? I am increasingly thinking of doing that
Well about 4 weeks ago I was mostly running small models. Some of my favorites were deepseek r1 8b and qwen 3.5 9b. Those are more or less good for boiler plate super fast responses(what I cared about most).
Now I am still trying out all the models that dropped this month. I am running qwen 3.6 35 a3b on a 16gb vram rtx 4060 ti.
I wish I sprung for a 24gb vram card but I never thought the price difference would matter. It seems like it does and I bet in the future there will be more models at this size because this is crazy.
It's not as good as opus if you are doing completely hands off programming but it's completely fine for me. I mostly use it for auto complete or templating a class. Other people are using it for agentic workflows with success.
Check out /r/localllama for more experiences. My set up is not the best but it is working for me and is saving me money.
2 replies →