← Back to context

Comment by lolinder

1 year ago

This isn't very surprising if you've interacted much with these models. Contrary to the claims in the various lawsuits, they're not just regurgitating images they've seen before, they have a good sense of abstract concepts and can pretty easily combine ideas to make things that have never been seen before.

This type of behavior has been evident ever since DALL-E's horse-riding astronaut [0]. There's no training image that resembles it (the astronaut even has their hands in the right position... mostly), it's combining ideas about what a figure riding a horse looks like and what an astronaut looks like.

Changing Albert Einstein's skin color should be even easier.

[0] https://www.technologyreview.com/2022/04/06/1049061/dalle-op...

> Contrary to the claims in the various lawsuits, they're not just regurgitating images they've seen before,

I don't think "just" is what the lawsuits are saying. It's the fact that they can regurgitate a larger subset (all?) of the original training data verbatim. At some point, that means you are copying the input data, regardless of how convoluted the tech underneath.

  • Fair, I should have said something along the lines of "contrary to popular conception of the lawsuits". I haven't actually followed the court documents at all, so I was actually thinking of discussions in mainstream and social media.