Comment by gowld
7 days ago
"jailbreaking" seems a silly term for "I told the LLM two unrelated things, and the response was relevant to only one of my comments, or a mixture of both."
It's not the LLM's fault that the human said something that the LLM understands better than the human :-)
No comments yet
Contribute on Hacker News ↗