Comment by MattPalmer1086

6 days ago

Well, that is my point. There is an inbuilt vulnerability in these systems as they do not (and apparently cannot) separate data and commands.

This is just one vector for this, there will be many, many more.

LLMs are doing what you train them to do. See for example " The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions " by Eric Wallace et al.

  • Interesting. Doesn't solve the problem entirely but seems to be a viable strategy to mitigate it somewhat.