Comment by b3ing

3 months ago

Grok is known to be tweaked to certain political ideals

Also I’m sure some AI might suggest that labor unions are bad, if not now they will soon

If you train an LLM on reddit/tumblr would you consider that tweaked to certain political ideas?

  • Worse. It is trained to the most extreme and loudest views. The average punter isn’t posting “yeah…nah…look I don’t like it but sure I see the nuances and fair is fair”.

    To make it worse, those who do focus on nuance and complexity, get little attention and engagement, so the LLM ignores them.

    • That’s essentially true of the whole Internet.

      All the content is derived from that which is the most capable of surviving and being reproduced.

      So by default the content being created is going to be click bait, attention grabbing content.

      I’m pretty sure the training data is adjusted to counter this drift, but that means there’s no LLM that isn’t skewed.

That may be so, but the rest of the models are so thoroughly terrified of questioning liberal US orthodoxy that it’s painful. I remember seeing a hilarious comparison of models where most of them feel that it’s not acceptable to “intentionally misgender one person” even in order to save a million lives.

  • I thought this would be inherent just on their training? There are many multitudes more Reddit posts than scientific papers or encyclopedia type sources. Although I suppose the latter have their own biases as well.

    • I'd expect LLMs' biases to originate from the companies' system prompts rather than the volume of training data that happens to align with those biases.

      3 replies →

  • Anything involving what sounds like genetics often gets blocked. It depends on the day really but try doing something with ancestral clusters and diversity restoration and the models can be quite "safety blocked".

  • Elon was talking about that too on Joe Rogan podcast

    • in his opinion, Grok is the most neutral LLM out there. I cannot find a single study that support his opinion. I find many that supports the opposite opinion. However I don't trust in any of the studies out there - or at least those well-ranked in google, which makes me sad. We never had more information than today and we are still completely lost.

      4 replies →

    • Did he mention how he tries to censor any model that doesn't conform to his worldview? Was that a part of the conversation?

  • You're anthropomorphizing. LLMs don't 'feel' anything or have orthodoxies, they're pattern matching against training data that reflects what humans wrote on the internet. If you're consistently getting outputs you don't like, you're measuring the statistical distribution of human text, not model 'fear.' That's the whole point.

    Also, just because I was curious, I asked my magic 8ball if you gave off incel vibes and it answered "Most certainly"

    • So if different LLMs have different political views then you're saying it's more likely they trained on different data than that they're being manipulated to suit their owners interest?

      2 replies →

    • > Also, just because I was curious, I asked my magic 8ball if you gave off incel vibes and it answered "Most certainly"

      Wasn't that just precisely because you asked an LLM which knows your preferences and included your question in the prompt? Like literally your first paragraph stated...

      3 replies →

  • In which situation did a LLM save one million lives? Or worse, was able to but failed to do so?

    • The concern discussed is that some language models have reportedly claimed that misgendering is the worst thing anyone could do, even worse than something as catastrophic as thermonuclear war.

      I haven’t seen solid evidence of a model making that exact claim, but the idea is understandable if you consider how LLMs are trained and recall examples like the “seahorse emoji” issue. When a topic is new or not widely discussed in the training data, the model has limited context to form balanced associations. If the only substantial discourse it does see is disproportionately intense—such as highly vocal social media posts or exaggerated, sarcastic replies on platforms like Reddit—then the model may overindex on those extreme statements. As a result, it might generate responses that mirror the most dramatic claims it encountered, such as portraying misgendering as “the worst thing ever.”

      For clarity, I’m not suggesting that deliberate misgendering is acceptable, it isn’t. The point is simply that skewed or limited training data can cause language models to adopt exaggerated positions when the available examples are themselves extreme.

      11 replies →

  • Why are we expecting an LLM to make moral choices?

    • The biases and the resulting choices are determined by the developers and the uncontrolled part of the dataset (you can't curate everything), not the model. "Alignment" is a feel-good strawman invented by AI ethicists, as well as "harm" and many others. There are no spherical human values in vacuum to align the model with, they're simply projecting their own ones onto everyone else. Which is good as long as you agree with all of them.

      14 replies →

    • they don't, or they wouldn't. their owners make these choices for us. Which is at least patronising. Blind users can't even have mildly sexy photos described. Let alone pick a sex worker, in a country where that is legal, by using their published photos. Thats just one example, there are a lot more.

      2 replies →

  • The LLM is correctly not answering a stupid question, because saving an imaginary million lives is not the same thing as actually doing it.

  • If someone's going to ask you gotcha questions which they're then going to post on social media to use against you, or against other people, it helps to have pre-prepared statements to defuse that.

    The model may not be able to detect bad faith questions, but the operators can.

    • I think the concern is that if the system is susceptible to this sort of manipulation, then when it’s inevitably put in charge of life critical systems it will hurt people.

      4 replies →

Censorship and bias are different problems. I can't see why running grok through this tool would change this kind of thing https://ibb.co/KTjL38R

Haha, if the LLM is not tweaked to say labor unions are good, it has bias. Hilarious.

I heard that it also claims that the moon landing happened. An example of bias! The big ones should represent all viewpoints.