Comment by togetheragainor
3 days ago
Some people think it’s a feature that when you prompt a computer system to do something, it does that thing, rather than censoring the result or giving you a lecture.
Perhaps you feel that other people shouldn’t be trusted with that much freedom, but as a user, why would you want to shackle yourself to a censored language model?
That’s what the Anthropic models do for me. I suppose I could be biased because I’ve never had a need for a model that spews racist, bigoted or sexist responses. The stuff @grok recently posted about Linda Yaccarino is a good example of why I don’t use it. But you do you.
You probably know better, and I probably should know better than to bother engaging, but...
Why would you conflate giving a computer an objective command with what is essentially someone else giving you access to query a very large database of "information" that was already curated by human beings?
Look. I don't know Elon Musk, but his rhetoric and his behavior over the last several years has made it very clear to me that he has opinions about things and is willing to use his resources to push those opinions. At the end of the day, I simply don't trust him to NOT intentionally bias *any* tool or platform he has influence over.
Would you still see it as "censoring" a LLM if instead of front-loading some context/prompt info, they just chose to exclude certain information they didn't like from the training data? Because Mr. Musk has said, publicly, that he thinks Grok has been trained on too much "mainstream media" and that's why it sometimes provides answers on Twitter that he doesn't like, and that he was "working on it." If Mr. Musk goes in and messes around with the default prompts and/or training data to get the answers that align with his opinions, is that not censorship? Or is it only censorship when the prompt is changed to not repeat racist and antisemitic rhetoric?