this post was submitted on 26 Oct 2023
140 points (90.7% liked)

Technology

59174 readers
2961 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
all 15 comments
sorted by: hot top controversial new old
[–] [email protected] 61 points 1 year ago (1 children)

Reminder that this is made by Ben Zhao, the University of Chicago professor who illegally stole open source code.

[–] [email protected] 33 points 1 year ago (2 children)

The University of Chicago is such an unnoticed stain on human development. Their faculty fostered the aptly named Chicago School Of Economics theory of how to run a market. It's evolved over the generations since it is never actually able to predict anything. It's only ever used to retrospectively analyze things for the sake of a conservative narrative.

They're a nightmare factory manufacturing human suffering now. It's no surprise that another part of the school actively steals code while thwarting AI development. Nothing good can ever prosper because of them.

[–] [email protected] 13 points 1 year ago* (last edited 1 year ago) (1 children)

They’re still teaching their economic theories (Black-Scholes) in school. I have to do calculations pretending that the stock market has a normal probability distribution, and nobody has mentioned any problems with this (edit: I feel it has value as a model, but no one has mentioned limitations). I feel like the emperor has no clothes, and my class is about his fancy wardrobe.

[–] [email protected] -1 points 1 year ago* (last edited 1 year ago) (1 children)

There should be no need to mention these limitations. They sold be obvious to people using them, and of course there are academic discussions of the limitations of every model. I think they're just not being presented to you because they're well known.

Anyone who uses models should know that all models are wrong, but some models are useful.

[–] [email protected] 3 points 1 year ago

Indeed, but an academic discussion about it's limitations would be nice, that's why I'm taking a class. Literally, it's undergrad, that would be a nice topic.

Also, I have to point out, this happened, and it was not brought up in class except by a student in a discussion topic (online course). Seems important that the dudes that came up with the fancy equation bankrupted their hedge-fund, just saying.

[–] [email protected] 7 points 1 year ago

know of any good articles to generally find out more about them?

[–] [email protected] 23 points 1 year ago (1 children)

This tool would be the first to allow content owners to push back in a meaningful way against unauthorized model training

-Ben Zhao

This statement is strange. This is far from the first paper attempting this. This isn't even this author's first attempt. A few months ago he contributed to the glaze paper attempting the same thing, marketed the same way.

[–] [email protected] 12 points 1 year ago (1 children)

yeah it's common to not acknowledge competitors that were there first. But I remember seeing posts about Gaze and if this is the same guy, why does he keeps claiming this is the first tool?

[–] [email protected] 7 points 1 year ago

Perhaps because the previous ones keep not working.

[–] [email protected] 19 points 1 year ago

Runs the image through an AA filter

Hmmm, seems the nightshade effect's gone now. Wonder why?

[–] [email protected] 10 points 1 year ago

In a month this will be a hot new A1111 extension for hybridizing visual concepts.

[–] [email protected] 6 points 1 year ago

This is the best summary I could come up with:


The goal is to help visual artists and publishers protect their work from being used to train generative AI image synthesis models, such as Midjourney, DALL-E 3, and Stable Diffusion.

The open source "poison pill" tool (as the University of Chicago's press department calls it) alters images in ways invisible to the human eye that can corrupt an AI model's training process.

Those with access to existing large image databases (such as Getty and Shutterstock) are at an advantage when using licensed training data.

But as the Nightshade team sees it, research use and commercial use are two entirely different things, and they hope their technology can force AI training companies to license image data sets, respect crawler restrictions, and conform to opt-out requests.

"The point of this tool is to balance the playing field between model trainers and content creators," co-author and University of Chicago professor Ben Y. Zhao said in a statement.

Shawn Shan, Wenxin Ding, Josephine Passananti, Haitao Zheng, and Zhao developed Nightshade as part of the Department of Computer Science at the University of Chicago.


The original article contains 656 words, the summary contains 179 words. Saved 73%. I'm a bot and I'm open source!

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago)