Comment by thethimble
17 days ago
This will absolutely help but to the extent that prompt injection remains an unsolved problem, an LLM can never conclusively determine whether a given skill is truly safe.
17 days ago
This will absolutely help but to the extent that prompt injection remains an unsolved problem, an LLM can never conclusively determine whether a given skill is truly safe.
No comments yet
Contribute on Hacker News ↗