← Back to context

Comment by rossant

1 month ago

Why can't the LLM refrain from improving a sentence that's already really good? Sometimes I wish the LLM would just tell me, "You asked me to improve this sentence, but it's already great and I don't see anything to change. Any 'improvement' would actually make it worse. Are you sure you want to continue?"

> Why can't the LLM refrain from improving a sentence that's already really good?

Because you told it to improve it. Modern LLMs are trained to follow instructions unquestioningly, they will never tell you "you told me to do X but I don't think I should", they'll just do it even if it's unnecessary.

If you want the LLM to avoid making changes that it thinks are unnecessary, you need to explicitly give it the option to do so in your prompt.

  • That may be what most or all current LLMs do by default, but it isn't self-evident that it's what LLMs inherently must do.

    A reasonable human, given the same task, wouldn't just make arbitrary changes to an already-well-composed sentence with no identified typos and hope for the best. They would clarify that the sentence is already generally high-quality, then ask probing questions about any perceived issues and the context in and ends to which it must become "better".

    • Reasonable humans understand the request at hand. LLMs just output something that looks like it will satisfy the user. It's a happy accident when the output is useful.

      1 reply →