Comment by asadotzler

16 hours ago

> but if it's not sure, it should quit

Can it be sure or not? I've never been able to get LLMs to give confidence measures that match their actual outputs. I'll ask an LLM "Are you sure?" and it'll reply "Absolutely" when it's output is completely wrong, or it'll backtrack on a correct output with "I should not have provided an answer when I was unsure. Here is an answer I am sure of..." and then provide something completely wrong.

If they can't properly and consistently score their confidence, how do they "know" when to quit and give it back to the human?

    > Claude did you just try to completely remove my large file *BEFORE* checking it into git LFS?
    > You're absolutely right! I should not have attempted an 'rm' comment on unstaged data. I guess I got a little frustrated with git, haha!