← Back to context

Comment by kromem

1 year ago

It's not a binary.

Why are the only options "only generate comically inaccurate images to the point of being offensive to probably everyone" or "only generate images of one group of people"?

Are current models so poor that we can't use a preprocessing layer to adapt the prompt aiming for diversity but also adjusting for context? Because even Musk's Grok managed to have remarkably nuanced responses to topics of race when asked racist questions by users in spite of being 'uncensored.'

Surely Gemini can do better than Grok?

Heavy handed approaches might have been necessary with GPT-3 era models, but with the more modern SotA models it might be time to adapt alignment strategies to be a bit more nuanced and intelligent.

Google wouldn't be damned if they'd tread a middle ground right now in between do and don't.