Comment by samiwami
20 hours ago
do they have anything similar to SynthID, or are they just pretending that problem doesn't exist?
I know this is probably mega cherry-picked to look more impressive, but some of the images are terrifyingly realistic. They seem to have put a lot of effort into the lighting.
> Integrating an imperceptible, robust, and content-specific watermark
From the system card someone linked elsewhere in the discussion
Zhao et al. 2023 showed any imperceptible watermark is provably removable by generative regeneration: pass the image through an img2img or VAE, the model reconstructs it visually identical but starts from a different latent. Watermark gone. SynthID and similar schemes do hold up well against normal sharing: recompression, crops, color tweaks, Twitter's pipeline. That covers most users. But the asymmetry is stuck — normally a GPU and a bit of motivation should be enough to strip it. Right? Got a tool to share? ;-)
> do they have anything similar to SynthID, or are they just pretending that problem doesn't exist?
At least they aren't pretending that a solution exists.
I feel like asking the image generators to mark AI images is the wrong way to go about it. It's like trying to maintain a blocklist. It seems better to me to have the major camera manufacturers or cell phones cryptographically sign their images as real.
I feel like this idea comes up often and in my opinion it doesn't solve anything. Take a picture of an AI image and you've made this approach useless. Which then goes to the argument of "well you'll see it's a picture of a picture" to which I will say there are plenty of ways to make this not appear so, and the ultimate form of this argument is that you can eventually project light directly into the photosensors, or otherwise hack the input between the photosensors and the rest of whatever digital magic that turns light into a JPG on your phone.
SynthID survives basic transforms including screenshots/photos, although it can of course be defeated. Even still it helps with the laziest fakes, which there seem to be a lot of - I've seen several quite widespread misinformative images over the past couple months that failed a synthID check.
Anyways I think approaching the problem from both directions is probably good.
Maybe a stupid question, but does the SynthID still exist if you screenshot and crop your generated image? What if you screenshot, rotate _just_ a bit, and crop? Or apply some other effect to the image like adjusting the coloring a little bit, adding some blur, etc.
The paper they published last year goes over some of these transformations: https://arxiv.org/pdf/2510.09263
I think we are just going to have to accept that realistic images can be easily fabricated now.
Seeing is not believing anymore, and I don't think SynthID or anything like it can restore that trust in images.
It's going to mess up accountability.
Some politician will be recorded doing something & he'll have his people release a thousand photos/videos of him doing crimes. And they'll say, look, it's a smear campaign.
This is just one stupid example, but people will have better schemes.
Also global coordinated releases of fake content and hypertargeted possibly abusive content. Virtual kidnappings will take off, automated & scaled.
Some politician will be recorded doing something & he'll have his people release a thousand photos/videos of him doing crimes. And they'll say, look, it's a smear campaign.
And his enemies will do the same, hopefully resulting in less blind trust for everyone in the population, which can only be a good thing.
Hopefully the arms race will balance out with improved AI image detection, but I can see how that will never be guaranteed to be reliable.