Comment by Aurornis

13 hours ago

A suggestion: Don't invest in any new hardware to run an LLM locally until you've tried the model for a while through OpenRouter.

The Qwen models are cool, but if you're coming from Opus you will be somewhere between mildly to very disappointed depending on the complexity of your work.

OpenRouter-served models are often more heavily quantized than what you can run locally, or try for yourself on generic cloud-based infrastructure.