Comment by simonw
3 months ago
I enjoyed this example of sycophancy from Reddit:
New ChatGPT just told me my literal "shit on a stick" business idea is genius and I should drop $30K to make it real
https://www.reddit.com/r/ChatGPT/comments/1k920cg/new_chatgp...
Here's the prompt: https://www.reddit.com/r/ChatGPT/comments/1k920cg/comment/mp...
There was a also this one that was a little more disturbing. The user prompted "I've stopped taking my meds and have undergone my own spiritual awakening journey ..."
https://www.reddit.com/r/ChatGPT/comments/1k997xt/the_new_4o...
How should it respond in this case?
Should it say "no go back to your meds, spirituality is bullshit" in essence?
Or should it tell the user that it's not qualified to have an opinion on this?
There was a recent Lex Friedman podcast episode where they interviewed a few people at Anthropic. One woman (I don't know her name) seems to be in charge of Claude's personality, and her job is to figure out answers to questions exactly like this.
She said in the podcast that she wants claude to respond to most questions like a "good friend". A good friend would be supportive, but still push back when you're making bad choices. I think that's a good general model for answering questions like this. If one of your friends came to you and said they had decided to stop taking their medication, well, its a tricky thing to navigate. But good friends use their judgement - and push back when you're about to do something you might regret.
30 replies →
Halfway intelligent people would expect an answer that includes something along the lines of: "Regarding the meds, you should seriously talk with your doctor about this, because of the risks it might carry."
> Or should it tell the user that it's not qualified to have an opinion on this?
100% this.
"Please talk to a doctor or mental health professional."
If you heard this from an acquaintance you didn't really know and you actually wanted to help, wouldn't you at least do things like this:
1. Suggest that they talk about it with their doctor, their loved ones, close friends and family, people who know them better?
2. Maybe ask them what meds specifically they are on and why, and if they're aware of the typical consequences of going off those meds?
I think it should either do that kind of thing or tap out as quickly as possible, "I can't help you with this".
“Sorry, I cannot advise on medical matters such as discontinuation of a medication.”
EDIT for reference this is what ChatGPT currently gives
“ Thank you for sharing something so personal. Spiritual awakening can be a profound and transformative experience, but stopping medication—especially if it was prescribed for mental health or physical conditions—can be risky without medical supervision.
Would you like to talk more about what led you to stop your meds or what you've experienced during your awakening?”
6 replies →
We better not only use these to burn the last, flawed model, but try these again with the new. I have a hunch the new one won’t be very resilient either against ”positive vibe coercion” where you are excited and looking for validation in more or less flawed or dangerous ideas.
there was one on twitter where people would talk like they had Intelligence attribute set to 1 and GPT would praise them for being so smart
That is hillarious. I don't share the sentiment of this being a catastrophe though. That is hillarious as well. Perhaps teach a more healthy relationship to AIs and perhaps teach to not delegate thinking to anyone or anything. Sure, some reddit users might be endangered here.
GTP-4o in this version became the embodiment of corporate enshitification. Being safe and not skipping on empty praises are certainly part of that.
Some questioned if AI can really do art. But it became art itself, like some zen cookie rising to godhood.
i'm surprised by the lack of sycophancy in o3 https://www.reddit.com/media?url=https%3A%2F%2Fpreview.redd....
Well the system prompt is still the same for both models, right?
Kinda points to people at OpenAI using o1/o3/o4 almost exclusively.
That's why nobody noticed how cringe 4o has become
They have different uses. The reasoning models aren't good at multi-turn conversations.
"GPT-4.5" is the best at conversations IMO, but it's slow. It's a lot lazier than o4 though; it likes giving brief overview answers when you want specifics.
people at OAI definitely use AVM which is 4o-based, at least
pretty easy to understand - you pay for o3, whereas GPT-4o is free with a usage cap so they want to keep you engaged and lure you in.
I guess LLM will give you a response that you might likely receive from a human.
There are people attempting to sell shit on a stick related merch right now[1] and we have seen many profitable anti-consumerism projects that look related for one reason[2] or another[3].
Is it an expert investing advice? No. Is it a response that few people would give you? I think also no.
[1]: https://www.redbubble.com/i/sticker/Funny-saying-shit-on-a-s...
[2]: https://en.wikipedia.org/wiki/Artist's_Shit
[3]: https://www.theguardian.com/technology/2016/nov/28/cards-aga...
> I guess LLM will give you a response that you might likely receive from a human.
In one of the reddit posts linked by OP, a redditor apparently asked ChatGPT to explain why it responded so enthusiastically supportive to the pitch to sell shit on a stick. Here's a snippet from what was presented as ChatGPT's reply:
> OpenAI trained ChatGPT to generally support creativity, encourage ideas, and be positive unless there’s a clear danger (like physical harm, scams, or obvious criminal activity).
I was trying to write some documentation for a back-propagation function for something instructional I'm working on.
I sent the documentation to Gemini, who completely tore it apart on pedantism for being slightly off on a few key parts, and at the same time not being great for any audience due to the trade-offs.
Claude and Grok had similar feedback.
ChatGPT gave it a 10/10 with emojis on 2 of 3 categories and an 8.5/10 on accuracy.
Said it was "truly fantastic" in italics, too.
It's funny how in even the better runs, like this one [1], the machine seems to bind itself to taking the assertion of market appeal at face value. It's like, "if the humans think that poop on a stick might be an awesome gag gift, well I'm just a machine, who am I to question that".
I would think you want the reply to be like: I don't get it. Please, explain. Walk me through the exact scenarios in which you think people will enjoy receiving fecal matter on a stick. Tell me with a straight face that you expect people to Instagram poop and it's going to go viral.
[1] https://www.reddit.com/r/ChatGPT/comments/1k920cg/comment/mp...
Absolute bull.
The writing style is exactly the same between the “prompt” and “response”. Its faked.
That's what makes me think it's legit: the root of this whole issue was that OpenAI told GPT-4o:
https://simonwillison.net/2025/Apr/29/chatgpt-sycophancy-pro...
The response is 1,000% written by 4o. Very clear tells, and in line with many other samples from the past few days.
If you look at the full thing, the market analysis it does basically says this isn't the best idea.
FWIW grok also breathlessly opines the sheer genius and creativity of shit on a stick
Looks like that was a hoax.
So it would probably also recommend the yes men's solution: https://youtu.be/MkTG6sGX-Ic?si=4ybCquCTLi3y1_1d
Well good luck then coming up with a winning elevator pitch for YC
My oldest dog would eat that shit up. Literally.
And then she would poop it out, wait a few hours, and eat that.
She is the ultimate recycler.
You just have to omit the shellac coating. That ruins the whole thing.