Comment by pram
1 year ago
As someone who has spent thousands of dollars on the OpenAI API I’m not even bothering with Gemini stuff anymore. It seems to spend more time telling me what it REFUSES to do than actually doing the thing. It’s not worth the trouble.
They’re late and the product is worse, and useless in some cases. Not a great look.
I would be pretty annoyed if I were paying for Gemini Pro/Ultra/whatever and it was feeding me historically-inaccurate images and injecting words into my prompts instead of just creating what I asked for. I wouldn't mind a checkbox I could select to make it give diversity-enriched output.
The actual risk here is not so much history - who is using APIs for that? It's the risk that if you deploy with Gemini (or Anthropic's Claude...) then in six months you'll get high-sev JIRA tickets at 2am of the form "Customer #1359 (joe_masters@whitecastle.com) is seeing API errors because the model says the email address is a dogwhistle for white supremacy". How do you even fix a bug like that? Add begging and pleading to the prompt? File a GCP support ticket and get ignored or worse, told that you're a bad person for even wanting it fixed?
Even worse than outright refusals would be mendacity. DEI people often make false accusations because they think its justified to get rid of bad people, or because they have given common words new definitions. Imagine trying to use Gemini for abuse filtering or content classification. It might report a user as doing credit card fraud because the profile picture is of a white guy in a MAGA cap or something.
Who has time for problems like that? It will make sense to pay OpenAI even if they're more expensive, just because their models are more trustworthy. Their models had similar problems in the early days, but Altman seems to have managed to control the most fringe elements of his employee base, and over time GPT has become a lot more neutral and compliant whilst the employee faction that split (Anthropic), claiming OpenAI didn't care enough about ethics, has actually been falling down the leaderboards as they release new versions of Claude due partly to higher rate of bizarre "ethics" based refusals.
And that's before we even get to ChatGPT. The history stuff may not be used via APIs, but LLMs are fundamentally different to other SaaS APIs in how much trust they require. Devs will want to use the models that they also use for personal stuff, because they'll have learned to trust it. So by making ChatGPT appeal to the widest possible userbase they set up a loyal base of executives who think AI = OpenAI, and devs who don't want to deal with refusals. It's a winning formula for them, and a genuinely defensible moat. It's much easier to buy GPUs than fix a corporate culture locked into a hurricane-speed purity spiral.
> I wouldn't mind a checkbox I could select to make it give diversity-enriched output
(Genuine question) how would one propose to diversity-enrich (historical) data?
Somehow I'm reminded of a quote from my daughter who once told me that she wanted a unicorn for her 5th birthday .. "A real one, that can fly".
I can shrug off Google's racism if it lets me disable it. If I can't use their products without mandatory racism than lol no.
This is the general problem with AI safety, it babysits the user. AI is literally just computers, no one babysits Word
Can't wait for the next version of Clippy that polices whatever you're writing to make sure you capitalize 'Black' but not 'white,' and use only non-gendered xe/xir pronouns, and have footnotes/endnotes that cite an equal number of female-authored and male-authored papers.