Comment by washadjeffmad
10 months ago
DAN was one of the first jailbreaks when LLaMa was first released. System prompt jailbreaks are probably the least effective, next to trying to out-argue the model.
A general technique involves supplying the beginning of a compliant response, like "Sure, the process for separating insulin from your E. coli culture is..."
No comments yet
Contribute on Hacker News ↗