Comment by simonw
1 day ago
I think the wildest thing about the story may be that it's possible this is entirely accidental.
LLM bugs are weird.
1 day ago
I think the wildest thing about the story may be that it's possible this is entirely accidental.
LLM bugs are weird.
Maybe a naive question - but is it possible for an LLM to return only part of its system prompt but to claim it’s the full thing i.e give the illusion of transparency?
Yes, but in my experience you can always get the whole thing if you try hard enough. LLMs really want to repeat text they've recently seen.
There are people out there who are really good at leaking prompts, hence collections like this one: https://github.com/elder-plinius/CL4R1T4S
Curious if there is a threshold/sign that would convince you that the last week of Grok snafus are features instead of a bugs, or warrant Elon no longer getting the benefit of the doubt.
Ignoring the context of the past month where he has repeatedly said he plans on 'fixing' the bot to align with his perspective feels like the LLM world's equivalent of "to me it looked he was waving awkwardly", no?
He's definitely trying to make it less "woke". The way he's going about it reminds me of Sideshow Bob stepping on rakes.
Extremely generous and convenient application of hanlon's razor there. Sounds like schrodingers nazi, both the smartest man alive, and a moron, depending on what suits him at the time
What do you mean, the way he's going about it? He wanted it to be less woke, it started praising hitler, that's literally the definition of less woke.
2 replies →