← Back to context

Comment by dent9

3 hours ago

This is completely wrong. Codex 5.2 and Claude Sonnet 4.5 don't have any of these issues. They will regularly tell you that you're wrong if you bother to ask them and they will explain why and what a better solution is. They don't make up anything. The code they produce is noticeably more efficient in LoC than previous models. And yes they really will do research, they will search the Internet for docs and articles as needed and cite their references inline with their answers.

You talk as if you haven't used a LLM since 2024. It's now almost 2026 and things have changed a lot.

With apologies, and not GP, but this has been the same feedback I've personally seen on every single model release.

Whenever I discuss the problems that my peers and I have using these things, it's always something along the lines of "but model X.Y solves all that!", so I obediently try again, waste a huge amount of time, and come back to the conclusion that these things aren't great at generation, but they are fantastic at summarization and classification.

When I use them for those tasks, they have real value. For creation? Not so much.

I've stopped getting excited about the "but model X.Y!!" thing. Maybe they are improving? I just personally haven't seen it.

But according to the AI hypers, just like with every other tech hype that's died over the past 30 years, "I must just be doing it wrong".

  • A lot of people are consistently getting their low expectations disproven when it comes to progress in AI tooling. If you read back in my comment history, six months ago I was posting about how AI is over hyped BS. But I kept using it and eventually new releases of models and tools solved most of the problems I had with them. If it has not happened for you yet then I expect it will eventually. Keep up with using the tools and models and follow their advancements and I think you'll eventually get to the point where your needs are met