Comment by steveBK123
3 days ago
It's gotta be more than that too though. Maybe training data other companies won't touch? Hidden prompt they aren't publishing? Etc.
Clearly Musk has put his hand on the scale in multiple ways.
3 days ago
It's gotta be more than that too though. Maybe training data other companies won't touch? Hidden prompt they aren't publishing? Etc.
Clearly Musk has put his hand on the scale in multiple ways.
I think they just told grok to favor conservative "sources" and it became "mechahitler" as the result.
It was starting N.... chains yesterday along with several other 4chan memes, so its definitely ingested a dataset consisting of at least 4chan posts that any sane company wouldn't touch with a 1000ft pole.
> Maybe training data other companies won't touch
That's a bingo. 3 weeks ago, Musk invited[1] X users to Microsoft-Tay[2] Grok by having them share share "divisive facts", then presumably fed the over 10,000 responses into the training/fine-tuning data set.
1. https://x.com/elonmusk/status/1936493967320953090
2. In 2016, Microsoft decided to let its Tay chatbot interact, and learn from Twitter users, and was praising Hitler in short order. They did it twice too, before shutting it down permanently. https://en.m.wikipedia.org/wiki/Tay_(chatbot)
That tweet seems like the bigger story.
I've seen lots of deflection saying Yaccarino chose to retire prior to Grok/MechaHitler, but the tweet predates that.
Even more deflection about how chatbots are easy to bait into saying weird things, but you don't need to bait when it has been specifically trained on it.
All of this was intentional. Musk is removing more of the mask, and he doesn't need Yaccarino to comfort advertisers any more.
I think it's more so that they push changes quickly without exhaustively testing. Compare that to Google, who sits on a model for years for fear of hurting their reputation, or OpenAI and Anthropic who extensively red teams models
Why does Grok keep "failing" in the same directional way if its just a testing issue?