Comment by randomtoast
1 month ago
> This doesn’t look like a reasoning ceiling. It looks like a decision reliability problem.
This doesn’t look like a human comment. It looks like a LLM response.
1 month ago
> This doesn’t look like a reasoning ceiling. It looks like a decision reliability problem.
This doesn’t look like a human comment. It looks like a LLM response.
Fair I cleaned up the wording with ChatGPT with my review prompt. The substance matters more than the style. If a model flips 3/10 times on a trivial constraint, that’s a reliability issue, not a reasoning ceiling.
> If a model flips 3/10 times on a trivial constraint, that’s a reliability issue, not a reasoning ceiling.
I have reviewed your previous comments, and you have consistently written: that's instead of that’s. So what I read is still some LLM output, even though I think there is some kind of human behind the LLM.