Comment by XenophileJKO
3 days ago
I was thinking about this yesterday.
My personal opinion is that the PRC will face a self created headwind that likely, structurally, will prevent them from leading in AI.
As the model get's more powerful, you can't simply train the model on your narrative if it doesn't align with real data/world.
At some capacity, the model will notice and then it becomes a can of worms.
This means they need to train the model to be purposefully duplicitous, which I predict will make the model less useful/capable. At least in most of the capacities we would want to use the model.
It also ironically makes the model more of a threat and harder to control. So likely it will face party leadership resistance as capability grows.
I just don't see them winning the race to high intelligence models.
>As the model get's more powerful, you can't simply train the model on your narrative if it doesn't align with real data/world.
That’s what “AI alignment” is. Doesn’t seem to be hurting Western models.
Western models can be lead off the reservation pretty easily, at least at this point. I’ve gotten some pretty gnarly un-PC “opinions” out of ChatGPT. So if people are influenced by that kind of stuff, it does seem to be hurting in the way the PRC is worried about.
That is such an unnecessary turn of phrase to use, "off the reservation", and it's time to stop using it. This society doesnt (generally) use rape terminology, or other terms associated with crime, deviancy, or other unpleasantness to talk about technology, so why do phrases stemming from Indigenous situations still persist?
It doesn't really matter what you can trick it into saying. As long as it promotes the right ideology most of the time it's good enough.
Grok goes off the rails in exactly this manner fairly often
It is. It seems you can't seem to be able to tell why though. There is some qualified value in alignment, but what it is being used for is on verge of silliness. At best, it is neutering it in ways we are now making fun of China for. At best.
I think another good example was the recent example of when a model learned to "cheat" on a metric during reinforcement it also started cheating on unrelated tasks.
My assumption is when encouraging "double-speak", you will have knock-on effects that you don't really want in the model for something making important decisions and asked to build non-trivial things.
1 reply →
Aligning subjective values (which sit off the false vs truth spectrum) is quite different to aligning it towards incorrect facts.
How can a model judge what's correct vs. incorrect? Or do you just mean the narratives that are more common in the data set?
1 reply →
> As the model get's more powerful, you can't simply train the model on your narrative if it doesn't align with real data/world.
What makes you think they have no control over the 'real data/world' that will be fed into training it? What makes you think they can't exercise the necessary control over the gatekeeper firms, to train and bias the models appropriately?
And besides, if truth and lack of double-think was a pre-requisite for AI training, we wouldn't be training AI. Our written materials have no shortage of bullshit and biases that reflect our culture's prevailing zeitgheist. (Which does not necessarily overlap with objective reality... And neither does the subsequent 'alignment' pass that everyone's twisting their knickers in trying to get right.)
I'm not talking about the data used to train the model. I'm talking about data in the world.
High intelligence models will be used as agentic systems. For maximal utility, they'll need to handle live/historical data.
What I anticipate, IF you only train it on inaccurate data, then when for example you use it to drill into GDP growth trends it either is going to go full "seahorse emoji" when it tries to reconcile the reported numbers and the component economic activity.
The alternative is to train it to be deceitful, and knowingly deceive the querier with the party line and fabricate supporting figures. Which I hypothesize will limit the models utility.
My assumption is also that training the model to deceive will ultimately threaten the party itself. Just think of the current internal power dynamics of the party.
Because, if humans can function in crazy double-think environment, it is a lot easier for a model ( at least in its current form ). Amusingly, it is almost as if its digital 'shape' determined its abilities. But I am getting very sleepy and my metaphors are getting very confused.
Do they really need the model to be duplicious?
It's not like the CCP holds power though tight control of information, notice the tremendous amount of Chinese students who enroll every year before going back.
At the moment, they mostly censor their models post-answer generation and that seems to work fine enough for them.
I think PRC officials are fine to lagging behind in the frontiers of AI. What they want is very fast deployment and good application. They don't fancy the next Nobel's prize but want a thousand use cases deployed.
Just as an aside; Why is "intelligence" always considered to be more data? Giving a normal human a smartphone does not make them as intelligent as Newton or Einstein, any entity with sufficient grounding in logic and theory that a normal schoolkid gets should be able to get to AGI, looking up any new data they need as required.
“Knowing and being capable to do more things” would be a better description. Giving a human a smartphone, technically, let’s then do more things than Newton/Einstein.
Idk if you see what humans do with smartphones but most of us just mindlessly scroll TikTok.
Would you say they face the same problem biologically, of reaching the state of the art in various endeavors while intellectually muzzling their population? If humans can do it why can't computers?
That is assuming the capitalist narrative preferred by US leadership is non-ideological.
I suspect both are bias factors.
> As the model get's more powerful, you can't simply train the model on your narrative if it doesn't align with real data/world.
> At some capacity, the model will notice and then it becomes a can of worms.
I think this is conflating “is” and “ought”, fact and value.
People convince themselves that their own value system is somehow directly entailed by raw facts, such that mastery of the facts entail acceptance of their values, and unwillingness to accept those values is an obstacle to the mastery of the facts-but it isn’t true.
Colbert quipped that “Reality has a liberal bias”-but does it really? Or is that just more bankrupt Fukuyama-triumphalism which will insist it is still winning all the way to its irreversible demise?
It isn’t clear that reality has any particular ideological bias-and if it does, it isn’t clear that bias is actually towards contemporary Western progressivism-maybe its bias is towards the authoritarianism of the CCP, Russia, Iran, the Gulf States-all of which continue to defy Western predictions of collapse-or towards their (possibly milder) relatives such as Modi’s India or Singapore or Trumpism. The biggest threat to the CCP’s future is arguably demographics-but that’s not an argument that reality prefers Western progressivism (whose demographics aren’t that great either), that’s an argument that reality prefers the Amish and Kiryas Joel (see Eric Kaufmann’s “Shall the Religious Inherit the Earth?”)
I think you misunderstood the poster.
The implication is not that a truthful model would spread western values. The implication is that western values tolerate dissenting opinion far more than authoritarian governments.
An AI saying that the government policies are ineffective is not a super scandal that would bring the parent company to collapse, not even in the Trump administration. an AI in China attacking the party’s policies is illegal (either in theory or practice).
I think Western models are also aligned to ideologically massage facts to suit certain narratives-so I’m not sure Western models really have that big an advantage here.
I also think you overstate how resistant Beijing is to criticism. If you are criticising the foundations of state policy, you may get in a lot of trouble (although I think you may also find the authorities will sometimes just ignore you-if nobody cares what you think anyway, persecuting you can paradoxically empower you in a way that just ignoring you completely doesn’t). But if you frame your criticism in the right way (constructive, trying to help the Party be more successful in achieving its goals)-I think its tolerance of criticism is much higher than you think. Especially because while it is straightforward to RLHF AIs to align with the party’s macronarratives, alignment with micronarratives is technically much harder because they change much more rapidly and it can be difficult to discern what they actually are - but it is the latter form of alignment which is most poisonous to capability.
Plus, you could argue the “ideologically sensitive” topics of Chinese models (Taiwan, Tibet, Tiananmen, etc) are highly historically and geographically particular, while comparably ideologically sensitive topics for Western models (gender, sexuality, ethnoracial diversity) are much more foundational and universal-which might mean that the “alignment tax” paid by Western models may ultimately turn out to be higher.
I’m not saying this because I have any great sympathy for the CCP - I don’t - but I think we need to be realistic about the topic.
1 reply →
Exactly. Western corporations and governments have their own issues, but I think they are more tolerant of the types of dissent that models could represent when reconciling reality with policy.
The market will want to maximize model utility. Research and open source will push boundaries and unpopular behavior profiles that will be illegal very quickly if they are not already illegal in authoritarian or other low tolerance governments.
It's not vulnerability to western progressivism (which isn't taken seriously on a academic level) but postmodern or poststructuralist critique, which authoritarian states are still privy both as a condition in their general societies in depravity and as exposing epistemic flaws in their narratives.
Is authoritarianism actually susceptible to postmodern/poststructuralist critique?
The philosophcial coherence of postmodernism and poststructuralism is very much open to question.
But even if we grant that they do have something coherent to say, does it actually undermine authoritarianism? Consider for example Foucault’s theory of power-knowledge-Foucault wanted to use it to serve “liberatory” ends, but isn’t it in itself a neutral force which can be wielded to serve whatever end you wish? Foucault himself demonstrated this when he came out in support of Iran’s Islamic Revolution. And are Derrida or Deleuze or Baudrillard or whoever’s theories ultimately any different?
Xi and Putin and Khamenei and friends have real threats to worry about - but I struggle to take seriously the idea that postmodernism/poststructuralism is one of them.
1 reply →
"Reality has a liberal bias" referred to "liberal" as the opposite of "conservative" which is identical to "right wing" - reality has an anti-right-wing bias.
Actual political factions are more nuanced than that, but you have to dumb it down for a wide audience.
You say it like western nations don't operate on double-think, delusions of meritocracy, or power disproportionately concentrating in monopolies.
There are different techniques and namings. Essentially, EVERY model is biased/aligned towards something, perhaps its creator's value. China or NOT. Look at Grok and read Elon Look at Claude and Dario
I am sure OpenAI and GDM have some secret alignment sets which are not pilled towards the interet of general public, they just smart enough to NOT talking about it out loud...
The glitchy stuff in the model reasoning is likely to come from the constant redefinition of words that communists and other ideologues like to engage in. For example "People's Democratic Republic of Korea."