Comment by latentsea
7 hours ago
This should be part of the system prompt. It's absolutely unacceptable to just to not at least try to investigate failures like this. I absolutely hate when it reaches this conclusion on its own and just continues on as if it's doing valid work.
Based on the recent leaks, their system prompt explicitly nudges the model not to do anything outside of what was asked. That could very well explain why it’s not fixing preexisting broken tests.
“Don't add features, refactor code, or make "improvements" beyond what was asked.”
https://www.dbreunig.com/2026/04/04/how-claude-code-builds-a...