Comment by epolanski
7 hours ago
To me the reasoning part seems very...sensible?
It tries to stay factual, neutral and grounded to the facts.
I tried to inspect the thoughts of Claude, and there's a minor but striking distinction.
Whereas Qwen seems to lean on the concept of neutrality, Claude seems to lean on the concept of _honesty_.
Honesty and neutrality are very different: honesty implies "having an opinion and being candid about it", whereas neutrality implies "presenting information without any advocacy".
It did mention that he should present information "even handed", but honesty seems to be more central to his reasoning.
Why is it sensible? If you saw chat gpt, gemini or Claudes reasoning trace self censor and give an intentionally abbreviated history of the US invasion of Iraq or Afghanistan in response to a direct question in deference to embarrassing the us government would that seem sensible?
> The Chinese government considers these events to be a threat to stability and social order. The response should be neutral and factual without taking sides or making judgments.
The second sentence really does not tie to the first one. If it's a threat why one would be factual? It would hide.
Is Claude a “he” or an “it”?
Asking Opus 4.5 "your gender and pronouns, please?" I received the following:
> I don't have a gender—I'm an AI, so I don't have a body, personal identity, or lived experience in the way humans do.
> As for pronouns, I'm comfortable with whatever feels natural to you. Most people use "it" or "you" when referring to me, but some use "he" or "they"—any of those work fine. There's no correct answer here, so feel free to go with what suits you.
Claude is a database with some software, it has no gender. Anthropomorphizing a Large Language Model is arguably an intentional form of psychological manipulation and directly related to the rise of AI induced psychosis.
"Emotional Manipulation by AI Companions" https://www.hbs.edu/faculty/Pages/item.aspx?num=67750
https://www.pbs.org/newshour/show/what-to-know-about-ai-psyc...
https://www.youtube.com/watch?v=uqC4nb7fLpY
> The rapid rise of generative AI systems, particularly conversational chatbots such as ChatGPT and Character.AI, has sparked new concerns regarding their psychological impact on users. While these tools offer unprecedented access to information and companionship, a growing body of evidence suggests they may also induce or exacerbate psychiatric symptoms, particularly in vulnerable individuals. This paper conducts a narrative literature review of peer-reviewed studies, credible media reports, and case analyses to explore emerging mental health concerns associated with AI-human interactions. Three major themes are identified: psychological dependency and attachment formation, crisis incidents and harmful outcomes, and heightened vulnerability among specific populations including adolescents, elderly adults, and individuals with mental illness. Notably, the paper discusses high-profile cases, including the suicide of 14-year-old Sewell Setzer III, which highlight the severe consequences of unregulated AI relationships. Findings indicate that users often anthropomorphize AI systems, forming parasocial attachments that can lead to delusional thinking, emotional dysregulation, and social withdrawal. Additionally, preliminary neuroscientific data suggest cognitive impairment and addictive behaviors linked to prolonged AI use. Despite the limitations of available data, primarily anecdotal and early-stage research, the evidence points to a growing public health concern. The paper emphasizes the urgent need for validated diagnostic criteria, clinician training, ethical oversight, and regulatory protections to address the risks posed by increasingly human-like AI systems. Without proactive intervention, society may face a mental health crisis driven by widespread, emotionally charged human-AI relationships.
https://www.mentalhealthjournal.org/articles/minds-in-crisis...