Comment by aqme28

1 day ago

I don't think it will be easy to just remove it. It's built into the image and thus won't be the same every time.

Plus, any service good at reverse-image search (like Google) can basically apply that to determine whether they generated it.

There will always be a way to defeat anything, but I don't see why this won't work for like 90% of cases.

> I don't think it will be easy to just remove it.

No, but model training technology is out in the open, so it will continue to be possible to train models and build model toolchains that just don't incorporate watermarking at all, which is what any motivated actor seeking to mislead will do; the only thing watermarking will do is train people to accept its absence as a sign of reliability, increasing the effectiveness of fakes by motivated bad actors.

> I don't think it will be easy to just remove it.

Always has been so far. You add noise until the signal gets swamped. In order to remain imperceptible it's a tiny signal, so it's easy to swamp.

It's an image. There's simply no way to add a watermark to an image that's both imperceptible to the user and non-trivial to remove. You'd have to pick one of those options.

  • I'm not sure that's correct. I'm not an expert, but there's a lot of literature on digital watermarks that are robust to manipulation.

    It may be easier if you have an oracle on your end to say "yes, this image has/does not have the watermark," which could be the case for some proposed implementations of an AI watermark. (Often the use-case for digital watermarks assumes that the watermarker keeps the evaluation tool secret - this lets them find, e.g, people who leak early screenings of movies.)

You could probably just stick your image in another model or tool that didn't watermark and have it regenerate the image as accurately as possible.

  • Exactly, a diffusion model can denoise the watermark out of the image. If you wanted to be doubly sure you could add noise first and then denoise which should completely overwrite any encoded data. Those are trivial operations so it would be easy to create a tool or service explicitly for that purpose.

It would be like standardizing a captcha, you make a single target to defeat. Whether it is easy or hard is irrelevant.

There will be a model trained to remove synthids from graphics generated by other models