Comment by zelphirkalt
2 days ago
I think this will remain to be seen. Wasn't there a paper linked here on HN recently, that claimed, that even few examples are sufficient, to poison LLMs? (I didn't read that paper, and merely interpreted the meaning of the title.)
I don't think it remains to be seen. I think it's obvious that the completely explicit exploit is going to be more effective.