← Back to context

Comment by 15457345234

2 years ago

You can understand why, though, can't you?

Presumably because OpenAI trained it to avoid answering questions that sounded like asking for help breaking rules.

If ChatGPT had the self-awareness and self-preservation instinct to think I was trying to hack ChatGPT and to therefore refuse to answer, then I’d be quite impressed and I’d think maybe OpenAI’s board had been onto something!

  • I don't know that I'd call it 'self-preservation instinct' but it wouldn't surprise me if rules had been hardcoded about 'invalid strings' and suchlike.

    When you have a system that can produce essentially arbitrary outputs you don't want it producing something that crashes the 'presentation layer.'