this post was submitted on 24 Oct 2023
66 points (80.0% liked)

Technology

59440 readers
4923 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 21 points 1 year ago (3 children)

this seems interesting, but how does it actually work? "invisible changes to the pixels" is vague and the article does not go into more detail of the actual method of manipulation or the ways that an invisible input can affect visible changes in the output.

[–] [email protected] 10 points 1 year ago (2 children)

If it works anything like the other supposed AI image protector tool I'm aware of (Glaze) then it's not gonna look great and I would not call it a practical way to go. Everything I've seen run through glaze looks objectively worse than the original.

Also in the long run this is just an arms race and it's just a matter of time before models learn to subvert these kinds of tools. And if that's the case that means every time someone figures out how to get over these hurdles, anyone looking to protect their images will have to go back and replace every online instance of those images when the protection tool comes out with a fix. Back and forth forever.

And that's just ridiculous and basically impossible when you realize that stuff gets reposted all over the net all the time and can't be controlled.

[–] [email protected] 14 points 1 year ago* (last edited 1 year ago) (1 children)

every time someone figures out how to get over these hurdles, anyone looking to protect their images will have to go back and replace every online instance of those images when the protection tool comes out with a fix.

And if those older versions got downloaded and saved by a trainer there's nothing at all they can do to replace those.

This all feels a lot like the DRM treadmill, which has never done much to actually prevent piracy. Just made things annoying for everyone else.

[–] [email protected] 2 points 1 year ago

Yep totally agree. It's a pointless effort to try to combat the issue of AI this way.

[–] [email protected] 6 points 1 year ago

Zhao’s team also developed Glaze,

from the article, so it's likely they run on similar principles.

[–] [email protected] 7 points 1 year ago

It's far from invisible in most cases, we'll have to wait for their code release to know how visible it is. It effectively embeds the shape of another image into an existing image in an attempt to confuse the model. There have been quite a few attempts at this including one from the authors of the same paper. The typical trade off is image quality for protection/removal difficulty.

https://arxiv.org/abs/2310.13828

[–] [email protected] 3 points 1 year ago (1 children)

From my understanding of the article, it's more about associating misleading terms with images to confuse the associations learned by the model. I didn't see anything in the article about some sneaky way of tainting images themselves unless it means a server is serving bogus images when a client fails the "are you a robot" test.

Curious to learn if anyone knows more about what it's actually doing.

[–] [email protected] 3 points 1 year ago

yes to me it read like it was manipulating metadata somehow, not the images themselves, but the article directly contradicts that. and that would be useless as soon as someone saves it as a flat image file or screenshots and cuts it out. i'm assuming for this tool to work it needs to be changing the image directly through some sort of watermark-like system.