Comment by bonzini

1 year ago

> If most nurses are women, I have no objection to a prompt for “nurse” showing a woman.

But if you're generating 4 images it would be good to have 3 women instead of four, just for the sake of variety. More varied results can be better, as long as they're not incorrect and as long as you don't get lectured if you ask for something specific.

From what I understand, if you train a model with 90% female nurses or white software engineers, it's likely that it will spit out 99% or more female nurses or white software engineers. So there is an actual need for an unbiasing process, it's just that it was doing a really bad job in terms of accuracy and obedience to the requests.

> So there is an actual need

You state this as a fact. Is it?

  • If a generator cannot produce a result that was in the training set due to overly biasing on the most common samples, then yes. If something was in 10% of the inputs and is produced in 1% of the outputs, there is a problem.

    I am pretty sure that it's possible to do it in a better way than by mangling prompts, but I will leave that to more capable people. Possible doesn't mean easy.