← Back to context

Comment by storus

4 days ago

With local inference on pretty decent local models we have nowadays (Qwen-3.5 and better) it's not much of a concern anymore.

Sure it is - there's still an opportunity cost of spending tokens(time/energy) creating a library from scratch vs using a preexisting well understood API.

what percentage of people is using local models for anything serious? I reckon single digits if even that. And for a corporate work environment, probably close to 0.