Comment by JSR_FDED
2 days ago
Same here, I’d be very interested to learn how others decide what model to use for which tasks.
I find these minutes-long iterations so painful that in practice I always go for the fast non-reasoning models.
2 days ago
Same here, I’d be very interested to learn how others decide what model to use for which tasks.
I find these minutes-long iterations so painful that in practice I always go for the fast non-reasoning models.
Imagine a tricky distributed systems problem where you feed all of the context of your setup to the LLM and it uses the advanced reasoning to diagnose possible avenues. I did that recently with a frontier model to unwrap some very tricky istio related connection pooling issues causing syn/ack floods.
For coding I usually use the fast frontier model like o4minihigh, but I bust out the fancy research models when I want things like general architecture and design feedbacks that require broader advanced reasoning
I don't often have LLMs write a lot of code for me, but when I do, I don't mind waiting a couple more minutes for a result that will waste less of my time in debugging when I try to use it.
Also it's useful to have models review code that I wrote -- in some cases years ago -- to uncover old bugs. Current models are generally far too eager to say "Yup! Looks good! You da man!" when there are actually serious flaws in the code they are reviewing. So again, this is a task that justifies use of the most powerful models currently available, and that doesn't have to run in real time.