Comment by Archelaos
2 days ago
I never expected LLMs to be like an actual conversation between humans. The model is in some respects more capable and in some respects more limited than a human. I mean, one could strive for an exact replica of a human -- but for what purpose? The whole thing is a huge association machine. It is a surealistic inspiration generator for me. This is how it works at the moment, until the next break through ...
> but for what purpose?
I recently introduced a non-technical person to Claude Code, and this non-human behavior was a big sticking point. They tried to talk to Claude similar as to a human, presenting it one piece of information at a time. With humans this is generally beneficial, and they will either nod for you to continue or ask clarifying questions. With Claude this does not work well, you have to infodump as much as possible in each message
So even from a perspective of "how do we make this automaton into the best tool", a more human-like conversation flow might be beneficial. And that doesn't seem beyond the technological capabilities at all, it's just not what we encourage in today's RLHF
I often find myself in these situations where I'm afraid that if I don't finish infodumping everything in a single message, it'll go in the wrong direction. So what I've been doing is switching it back to Plan Mode (even when I don't need a plan as such), just as a way of telling it "Hold still, we're still having a conversation".
I do this with cursor ai too. I tell, don't change anything, let me hear out what you plan to fix and what you will change
I haven't tried claude, but Codex manages this fine as long as you prompt it correctly to get started.
A lazy example:
"This goal of this project is to do x. Let's prepare a .md file where we spec out the task. Ask me a bunch of questions, one at a time, to help define the task"
Or you could just ask it to be more conversational, instead of just asking questions. It will do that.
also, this is what chat-style interfaces encourage. Anything where the "enter" key sends the message instead of creating a paragraph block is just hell.
I'm prompting Gemini, and I write:
I have the following code, can you help me analyze it? <press return>
<expect to paste the code into my chat window>
but Gemnini is already generating output, usually saying "I'm waiting for you to enter the code"
Yeah, seems like current models might benefit from a more email-like UI, and this'll be more true as they get longer task time horizons.
Maybe we want a smaller model tuned for back and forth to help clarify the "planning doc" email. Makes sense that having it all in a single chat-like interface would create confusion and misbehavior.
Like many chat-style interfaces, it's typically shift-enter to insert a newline.
1 reply →
I usually do the "drip feed" with ChatGPT, but maybe that's not optimal. Hmm, maybe info dump is a good thing to try.
There a recent(ish: May 2025) paper about how drip-feeding information is worse than restarting with a revised prompt once you realize details are missing.[0]
[0] https://arxiv.org/abs/2505.06120
1 reply →
I hate when I accidentally hit return halfway through writing my prompt and it gives me two pages of advice about some nonsense half sentence.
Clarifying ambiguity in questions before dedicating more resources to search and reasoning about the answer seems both essential and almost trivial to elicit via RLHF.
I'd be surprised if you can't already make current models behave like that with an appropriate system prompt.
The disconnect is that companies are trying desperately to frame LLMs as actual entities and not just an inert tech tool. AGI as a concept is the biggest example of this, and the constant push to "achieve AGI" is what's driving a lot of stock prices and investment.
A strictly machinelike tool doesn't begin answers by saying "Great question!"