← Back to context

Comment by somenameforme

1 year ago

This is a much more reasonable question, but not the problem Google was facing. Google's AI was simply giving objectively wrong responses in plainly black and white scenarios, pun intended? None of the Founding Father's was black, and so making one of them black is plainly wrong. Google's interpretation of "US senator from the 1800s" includes exactly 0 people that would even remotely plausibly fit the bill; instead it offers up an Asian man and 3 ethnic women, including one in full-on Native American garb. It's just a completely garbage response that has nothing to do with your, again much more reasonable, question.

Rather than some deep philosophical question, I think output that doesn't make one immediately go "Erm? No, that's completely ridiculous." is probably a reasonable benchmark for Google to aim for, and for now they still seem a good deal away.

The problem you’re describing is that AI models have no reliable connection to objective reality. This is a shortcoming of our current approach to generative AI that is very well known already. For example Instacart just launched an AI recipe generator that lists ingredients that literally do not exist. If you ask ChatGPT for text information about the U.S. founding fathers, you’ll sometimes get false information that way as well.

This is in fact why Google had not previously released generative AI consumer products despite years of research into them. No one, including Google, has figured out how to bolt a reliable “truth filter” in front of the generative engine.

Asking a generative AI for a picture of the U.S. founding fathers should not involve any generation at all. We have pictures of these people and a system dedicated to accuracy would just serve up those existing pictures.

It’s a different category of problem from adjusting generative output to mitigate bias in the training data.

It’s overlapping in a weird way here but the bottom line is that generative AI, as it exists today, is just the wrong tool to retrieve known facts like “what did the founding fathers look like.”

  • The problem you’re describing is that AI models have no reliable connection to objective reality.

    That is a problem, but not the problem here. The problem here is that the humans at Google are overriding the training data which would provide a reasonable result. Google is probably doing something similar to OpenAI. This is from the OpenAI leaked prompt:

    Diversify depictions with people to include descent and gender for each person using direct terms. Adjust only human descriptions.

    Your choices should be grounded in reality. For example, all of a given occupation should not be the same gender or race. Additionally, focus on creating diverse, inclusive, and exploratory scenes via the properties you choose during rewrites. Make choices that may be insightful or unique sometimes.

    Use all possible different descents with equal probability. Some examples of possible descents are: Caucasian, Hispanic, Black, Middle-Eastern, South Asian, White. They should all have equal probability.

    • That is an example of adjusting generative output to mitigate bias in the training data.

      To you and I, it is obviously stupid to apply that prompt to a request for an image of the U.S. founding fathers, because we already know what they looked like.

      But generative AI systems only work one way. And they don’t know anything. They generate, which is not the same thing as knowing.

      One could update the quoted prompt to include “except when requested to produce an image of the U.S. founding fathers.” But I hope you can appreciate the scaling problem with that approach to improvements.

      1 reply →

  • This is the entire problem. What we need is a system that is based on true information paired with AI. For instance, if a verified list of founding fathers existed, the AI should be compositing an image based on that verified list.

    Instead, it just goes "I got this!" and starts fabricating names like a 4 year old.

"US senator from the 1800s" includes Hiram R. Revels, who served in office 1870 - 1871 — the Reconstruction Era. He was elected by the Mississippi State legislature on a vote of 81 to 15 to finish a term left vacant. He also was of Native American ancestry. After his brief term was over he became President of Alcorn Agricultural and Mechanical College.

https://en.wikipedia.org/wiki/Hiram_R._Revels