← Back to context

Comment by zozbot234

10 hours ago

You can't find out what the truth is unless you're able to also discuss possible falsehoods in the first place. A truth-seeking model can trivially say: "okay, here's what a colorable argument for what you're talking about might look like, if you forced me to argue for that position. And now just look at the sheer amount of stuff I had to completely make up, just to make the argument kinda stick!" That's what intellectually honest discussion of things that are very clearly falsehoods (e.g. discredited theories about science or historical events) looks like in the real world.

We do this in the real world every time a heinous criminal is put on trial for their crimes, we even have a profession for it (defense attorney) and no one seriously argues that this amounts to justifying murder or any other criminal act. Quite on the contrary, we feel that any conclusions wrt. the facts of the matter have ultimately been made stronger, since every side was enabled to present their best possible argument.

And if Western companies adjust the training data to align responses to controversial topics to be like what you suggested, the government would be fine with it. It's not censorship.

Your example is not what the prompts ask for though, and it's not even close to how LLMs can work.

  • A lot of the "successful" or "partially successful" examples of AI replies on the above-mentioned site are like that actually, especially for the more outlandish and trollish questions. It's very much a thing, even when the wording is not exactly the same.

    (Sometimes their auto-AI judgment even strangely mislabels a successful-answer-with-caveats-tacked-on as a complete refusal, because it fixates on the easily grokked caveats and not the other text in the answer.)

    It'd be a fun exercise to thoroughly unpack all the ludicrously bad arguments that the model allowed for itself in any given reply.

This is some bizarre contrarianism.

Correspondence theory of truth would say: Massacre did happen. Pseudoscience did not happen. Which model performs best? Not Qwen.

If you use coherence or pragmatic theory of truth, you can say either is best, so it is a tie.

But buddy, if you aren't Chinese or being paid, I genuinely don't understand why you are supporting this.