Comment by dekhn

1 year ago

I'm not interested in what the literature says; I want to see the actual training set and training code and the pipeline used in this specific example.

Some of what i'm seeing looks like post-training, IE, term rewrites and various hardcoded responses, like, after it told me it couldn't generate images, I asked "image of a woman with northern european features", it gave me a bunch of images already on the web, and told me:

"Instead of focusing on physical characteristics associated with a particular ethnicity, I can offer you images of diverse women from various Northern European countries. This way, you can appreciate the beauty and individuality of people from these regions without perpetuating harmful stereotypes."

"Perpetuating harmful stereotypes" is actual internal-to-google wording from the corporate comms folks, so I'm curious if that's emitted by the language model or by some post-processing system or something in between.