Comment by NitpickLawyer
17 hours ago
That's an interesting testing case, not for the political aspect, but for the data aspect. One would assume that the totality of "sensitive" data (especially in chinese) that gets thrown into the training dataset is quite limited. Getting a model that wasn't trained on such data (presumably) to actually talk about it would be an interesting exercise. Tho I'd suggest doing it with smaller models first.
No comments yet
Contribute on Hacker News ↗