← Back to context

Comment by janalsncm

13 hours ago

It is incoherent to ask for a “confession” from an LLM. An LLM is fundamentally predicting a next token, repeatedly. If you ask it “Why did you do X” it will not do the human thing and introspect about latent motives that we are only finding out about now. It will respond in the statistically likely way, which isn’t useful.

All this is to say that if you don’t know what you’re doing with software you can shoot yourself in the foot, and now with AI agents you can shoot yourself in the foot with a machine gun.

Don’t ask the AI agent nicely not to delete your backup databases. That isn’t reliable. Do not give them write permission to a thing you’re not comfortable with them writing to.