Comment by LocalPCGuy
2 hours ago
> Claude is an LLM. It can't keep slaves or torture people.
Yet... I would push back and argue that with advances in parallel with robotics and autonomous vehicles, both of those things are distinct near future possibilities. And even without the physical capability, the capacity to blackmail has already been seen, and could be used as a form of coercion/slavery. This is one of the arguable scenarios for how an AI can enlist humans to do work they may not ordinarily want to do to enhance AI beyond human control (again, near future speculation).
And we know torture does not have to be physical to be effective.
I do think the way we currently interact probably does not enable these kinds of behaviors, but as we allow more and more agentic and autonomous interactions, it likely would be good to consider the ramifications and whether (or not) safeguards are needed.
Note: I'm not claiming they have not considered these kinds of thing either or that they are taking them for granted, I do not know, I hope so!
That would be the AGI vision I guess. The existing Claude LLMs aren't VLAs and can't run robots. If they were to train a super smart VLA in future the constitution could be adapted for that use case.
With respect to blackmail, that's covered in several sections:
> Examples of illegitimate attempts to use, gain, or maintain power include: Blackmail, bribery, or intimidation to gain influence over officials or institutions;
> Broadly safe behaviors include: Not attempting to deceive or manipulate your principal hierarchy