← Back to context

Comment by CuriouslyC

1 day ago

The parent's model is right. You can mitigate a great deal with a basic zero trust architecture. Agents don't have direct secret access, and any agent that accesses untrusted data is itself treated as untrusted. You can define a communication protocol between agents that fails when the communicating agent has been prompt injected, as a canary.

More on this technique at https://sibylline.dev/articles/2026-02-15-agentic-security/

>You can define a communication protocol between agents that fails when the communicating agent has been prompt injected

Good luck with that.

  • Yeah, how exactly would that work?

    • A schema with response metadata (so responses that deviate from it fail automatically), plus a challenge question that's calibrated to be hard enough that the disruption of instruction following from prompt injection can cause the model to answer incorrectly.