Comment by cbolton
3 months ago
You can bypass the system prompt by using the API? I thought part of the "safety" of LLMs was implemented with the system prompt. Does that mean it's easier to get unsafe answers by using the API instead of the GUI?
3 months ago
You can bypass the system prompt by using the API? I thought part of the "safety" of LLMs was implemented with the system prompt. Does that mean it's easier to get unsafe answers by using the API instead of the GUI?
Safety is both the system prompt and the RLHF posttraining to refuse to answer adversarial inputs.
Yes, it is.