this post was submitted on 14 Oct 2024
9 points (84.6% liked)

Technology

34815 readers
242 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 3 weeks ago (6 children)

I'm actually on this man's side.

The idea-stealing he talks about is not unheard of, and multiple people or groups coming up with similar ideas at the same time by looking at market trends is actually quite common.

If you also look at the fact that he has evidence for pretty much all his claims,

AND

He has gotten the domain and has evidence for the ideas and ownership of "Open AI" before Altman's "OpenAI" was formed

AND

He says a lot of his ideas never came to fruition because he couldn't get funding but the one thing he didn't need crazy funding for, investing in Bitcoin when it was $10 per coin, is something he ends up doing and leaves him well-off.

All that to me is enough evidence that this man is one hell of an unlucky individual.

And as such, I believe him.

[–] [email protected] 2 points 3 weeks ago (4 children)

i'm not. just because he's an underdog here means that you're gonna ignore all the harms of generative ai up to this day? it's like complaining that big oil stole the idea of adding tetraethyllead to gasoline from you and you got no profits from that as a result

[–] [email protected] 2 points 3 weeks ago (1 children)

Not necessarily. A lot of the harms disappear when everything goes open, which is what this person stands for, and what OpenAI was supposed to stand for.

Open LLM + Open Training Data = Open AI

Copyright and IP concerns disappear with an open dataset.

Open models are inherently more trustworthy because of an obvious reduction in vendor lock-in.

[–] [email protected] 1 points 3 weeks ago (1 children)

Copyright and IP concerns disappear with an open dataset.

i don't think i'd agree with that, doesn't matter if dataset goes open if content went there without consideration for authors

also even things like thispersondoesnotexist were used to mass-create fake identities and such

[–] [email protected] 1 points 3 weeks ago (1 children)

Yeah, but something like that would be super easy to find and fix without going through lawsuits. And I'd argue the dataset creators would be far less likely to add copyrighted material to the training data when it's all out in the open and they can be immediately made to remove and retrain the AI without that data.

[–] [email protected] 1 points 3 weeks ago

the problem with that is that training can't be done "immediately" it takes tons of compute

load more comments (2 replies)
load more comments (3 replies)