Comment by Animats
2 years ago
What does "safe" mean?
1. Will not produce chat results which are politically incorrect and result in publicity about "toxic" comments?
2. Will not return false factual information which is dangerously wrong, such as that bad recipe on YC yesterday likely to incubate botulism toxin?
3. Will not make decisions which harm individuals but benefit the company running the system?
4. Will not try to take over from humans?
Most of the political attempts focus on type 1. Errors of type 2 are a serious problem. Type 3 errors are considered a feature by some, and are ignored by political regulators. We're not close to type 4 yet.
Ilya's talking about type 4.
As opposed to 3a, "Will not assist companies to take over".