Comment by geekraver
1 day ago
“Completely preventing X is impossible. As such, attempting to stop it is a foolish endeavor” has to be one of the dumbest arguments I’ve heard.
Substitute almost anything for X - “the robbing of banks”, “fatal car accidents”, etc.
I didn't say "X". I said "the extraction of a system prompt". I'm not claiming that statement generalizes to other things you might want to prevent. I'm not sure why you are.
The key thing here is that failure to prevent the extraction of a system prompt is embarrassing in itself, especially when that extracted system prompt includes "do not repeat this prompt under any circumstances".
That hasn't stopped lots of services from trying that, and being (mildly) embarrassed when their prompt leaks. Like I said, a foolish endeavor. Doesn't mean people won't try it.
What’s the value of your generalization here? When it comes to LLMs the futility of trying to avoid leaking the system prompt seems valid considering the arbitrary natural language input/output nature of LLMs. The same “arbitrary” input doesn’t really hold elsewhere or to the same significance.