Comment by bufferoverflow
3 days ago
LLMs can recognize errors in their own output. That's why thinking models generally perform much better than the non-thinking ones.
3 days ago
LLMs can recognize errors in their own output. That's why thinking models generally perform much better than the non-thinking ones.
No, a block of text that begins "please improve on the following text:" is likely to continue after the included block with some text that sounds like a correction or refinement.
Nothing is "recognized", nor is anything "an error". Nothing is "thinking" any more than it would be if the LLM just printed whether the next letter were more likely to be a vowel or consonant. Just because it's doing a better job modeling text doesn't magically make it be doing something that's not a text prediction function.
You're using the same words again. It looks like reasoning, but it's a simulation.
The LLM merchants are driving it though, by using pre-existing words for things that are not what they are saying they are.
It's amazing what they can do, but an LLM cannot know if what it outputs is true or correct, just statistically likely.