Comment by tosh

12 hours ago

afaiu not all of their models are open weight releases, this one so far is not open weight (?)

What would a good coding model to run on an M3 Pro (18GB) to get Codex like workflow and quality? Essentially, I am running out quick when using Codex-High on VSCode on the $20 ChatGPT plan and looking for cheaper / free alternatives (even if a little slower, but same quality). Any pointers?

  • Nothing. This summer I set up a dual 16GB GPU / 64GB RAM system and nothing I could run was even remotely close. Big models that didn't fit on 32gb VRAM had marginally better results but were at least of magnitude slower than what you'd pay for and still much worse in quality.

    I gave one of the GPUs to my kid to play games on.

    • Yup, even with 2x 24gb GPUs, it's impossible to get anywhere close to the big models in terms of quality and speed, for a fraction of the cost.

      2 replies →

  • Short answer: there is none. You can't get frontier-level performance from any open source model, much less one that would work on an M3 Pro.

    If you had more like 200GB ram you might be able to run something like MiniMax M2.1 to get last-gen performance at something resembling usable speed - but it's still a far cry from codex on high.

  • at the moment, I think the best you can do is qwen3-coder:30b -- it works, and it's nice to get some fully-local llm coding up and running, but you'll quickly realize that you've long tasted the sweet forbidden nectar that is hosted llms. unfortunately.

  • They are spending hundreds of billions of dollars on data centers filled with GPUs that cost more than an average car and then months on training models to serve your current $20/mo plan. Do you legitimately think there's a cheaper or free alternative that is of the same quality?

    I guess you could technically run the huge leading open weight models using large disks as RAM and have close to the "same quality" but with "heat death of the universe" speeds.

  • 18gb RAM it is a bit tight

    with 32gb RAM:

    qwen3-coder and glm 4.7 flash are both impressive 30b parameter models

    not on the level of gpt 5.2 codex but small enough to run locally (w/ 32gb RAM 4bit quantized) and quite capable

    but it is just a matter of time I think until we get quite capable coding models that will be able to run with less RAM

  • Z.ai has glm-4.7. Its almost as good for about $8/mo.

    • Not sure if it's me but at least for my use cases (software devl, small-medium projects) Claude Opus + Claude Code beats by quite a margin OpenCode + GLM 4.7. At least for me Claude "gets it" eventually while GLM will get stuck in a loop not understanding what the problem is or what I expect.

      1 reply →

  • A local model with 18GB of ram that has the same quality has codex high? Yeah, nah mate.

    The best could be GLN 4.7 Flash, and I doubt it's close to what you want.

  • "run" as in run locally? There's not much you can do with that little RAM.

    If remote models are ok you could have a look at MiniMax M2.1 (minimax.io) or GLM from z.ai or Qwen3 Coder. You should be able to use all of these with your local openai app.