Comment by Terr_

2 days ago

It means if you want something resembling a self-introspective theory of mind, you need to arrange the overall document to cohere to documents where such things are/appear-to-be happening.

This leads us to new questions: How can we characterize and identify real-world documents which fit? How can we determine what features may be significant, and which of those can be easily transplanted to our use-case?

There are a lot of words but it feels like you have never really used LLM's (apologies for the bluntness).

We see LLM's introspecting all the time[1].

>Notably, DeepSeek-AI et al. report that the average response length and downstreamperformance of DeepSeek-R1-Zero increases as training progresses. They further report an “aha moment” during training, which refers to the “emergence” of the model’s ability to reconsider its previously generated content. As we show in Section 3.2, this reconsideration behaviour is often indicated by the generation of phrases such as ‘wait, ...’ or ‘alternatively, ...’

[1] https://arxiv.org/pdf/2504.07128

  • Unless they show you the Markov chain weights (and I've never seen one that does), that's confabulation, not introspection.

    • Unless you can show the Markov chain weights, I declare all your thoughts confabulation, not introspection.

You are just doubling down on protecting your argument.

I operate LLMs in many conversational modes where it does ask clarifying questions, probing questions, baseline determining questions.

It takes at most one sentence in the prompt to get them to act this way.

  • > It takes at most one sentence in the prompt to get them to act this way.

    What is this one sentence you are using?

    I am struggling to elicite clarification behavior form llms

    • What is your domain and what assumptions are they making that they should be asking you for? Have you tried multiple models?

  • Could you share your prompt to get it to ask clarifying questions? I'm wondering if it would work in custom instructions.

    • It is domain dependent, you really need to play with it. Tell it you are doing pair thinking and either get it to ask questions about things it doesn't understand, or get it to ask you questions to get you to think better. Project the AI into a vantage point in the latent space and then get it to behave in the way that you want it to.

      You can ask it to use the Socratic method, but then it is probing you, not its own understanding. Now have it use the socratic method on itself. You can tell it to have multiple simultaneous minds.

      Play with deepseek in thinking and non-thinking mode, give it nebulous prompts and see if you can get it to ask for clarifications.