← Back to context

Comment by contravariant

5 months ago

Could be, but 'AI model says weird shit' has almost never stuck around unless it's public (which won't happen here), really common, or really blatantly wrong. And usually at least 2 of those three.

For something usually hidden the first two don't really apply that well, and the last would have to be really blatant unless you want an article about "Model recovers from mistake" which is just not interesting.

And in that scenario, it would have to mean the CoT contains something like blatant racism or just a general hatred of the human race. And if it turns out that the model is essentially 'evil' but clever enough to keep that hidden then I think we ought to know.

Just no. AI being racist is still a popular meme. "Because the programmers are white males blah blah".

  • Why can't it be, if it were (I'm not saying that it is, mind) trained on racist material?

    • The problem is being kind of right (but not really) for the wrong reasons. Normies think it was told to be a certain way. While kind of true, they think of it more like Eliza.