← Back to context

Comment by michaelrpeskin

1 year ago

> I'm not saying the current situation is good or optimal. But it's not obvious what the right result should be.

Yes, it's not obvious what the first result returned should be. Maybe a safe bet is to use the current ratio of sexes/races as the probability distribution just to counter bias in the training data. I don't think all but the most radical among us would get too mad about that.

What probability distribution? It can't be that hard to use the country/region of where the query is being made? Or the country/region about which the image is being asked for? All reasonable choices.

But, if the image generated isn't what you need (say the image of senators from the 1800's example). You should be able to direct it to what you need.

So just to be PC, it generates images of all kind of diverse people. Fine, but then you say, update it to be older white men. Then it should be able to do that. It's not racist to ask for that.

I would like for it to know the right answer right away, but I can imagine the political backlash for doing that, so I can see why they'd default to "diversity". But the refusal to correct images is what's over-the-top.