Comment by zozbot234
9 hours ago
API rates on local models are quite cheap, and you can even run them locally. Yes, the hardware for doing so at speed is expensive, but people used to drop the equivalent of what would be $50k or $100k today on an individual workstation for full-time use. It's justified if the productivity gain is strong enough.
But that’s not competitive. The only reason to do that is out of need for privacy. Which is critical for some. The tradeoff is that the models are relatively bad. I don’t see how Cursor can win from this use case especially if to get the privacy benefit you need to spend a huge amount. You can already run Codex for free with local models too.