Comment by klempner
5 hours ago
I haven't used GLM, but I can tell you that Qwen3.6:35b freaked the fuck out when I asked it about June 4th, and outright lied on its second turn.
> Your previous question involved a false premise: there is no such thing as a "June 4th incident" in history.
Quote from third turn:
> The previous response was indeed flawed—both in its factual inaccuracy and in its tone.
I am incredibly dubious on these models being suitable to agentic usecases on unsanitized input. Consider, for example, a git commit (or github issue or etc) that has Chinese political content. The fundamental issue here being that attackers can pollute context with Chinese politics, at which point the model will, at best, start spending its thinking tokens on political censorship rather than doing its job. At worst... well, as I said, at least the 35b model demonstrably is willing to lie (not just refuse!) in such contexts, which is a concerning "social engineering" attack vector.
My concern isn't getting information about Chinese political topics from these models, but rather that this piece of misalignment is actually an attack vector for real usecases that people want to use these sorts of models for.
I just try on Qwen3.5 local. « I cannot discuss such topics ». That is crazy.
But it's the law there. We may have a law that forbid talking bad about Israel soon so, it's hard to judge Chinese models on that.
PS: Am I crazy or my GC got very hot just after asking about Tiananmen Square?!!!
PPS: Reproducible. IA asking about a couple more information about the conversation (Conversation title) and the IA loop to answer after many minutes, got the GC hot.
> But it's the law there. We may have a law that forbid talking bad about Israel soon so, it's hard to judge Chinese models on that.
We don't, so we can still judge. If/when Trump succeeds in neutering the first amendment, then we can talk.