Technology
This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.
Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.
Rules:
1: All Lemmy rules apply
2: Do not post low effort posts
3: NEVER post naziped*gore stuff
4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.
5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)
6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist
7: crypto related posts, unless essential, are disallowed
view the rest of the comments
I'm actually on this man's side.
The idea-stealing he talks about is not unheard of, and multiple people or groups coming up with similar ideas at the same time by looking at market trends is actually quite common.
If you also look at the fact that he has evidence for pretty much all his claims,
AND
He has gotten the domain and has evidence for the ideas and ownership of "Open AI" before Altman's "OpenAI" was formed
AND
He says a lot of his ideas never came to fruition because he couldn't get funding but the one thing he didn't need crazy funding for, investing in Bitcoin when it was $10 per coin, is something he ends up doing and leaves him well-off.
All that to me is enough evidence that this man is one hell of an unlucky individual.
And as such, I believe him.
i'm not. just because he's an underdog here means that you're gonna ignore all the harms of generative ai up to this day? it's like complaining that big oil stole the idea of adding tetraethyllead to gasoline from you and you got no profits from that as a result
Not necessarily. A lot of the harms disappear when everything goes open, which is what this person stands for, and what OpenAI was supposed to stand for.
Open LLM + Open Training Data = Open AI
Copyright and IP concerns disappear with an open dataset.
Open models are inherently more trustworthy because of an obvious reduction in vendor lock-in.
i don't think i'd agree with that, doesn't matter if dataset goes open if content went there without consideration for authors
also even things like thispersondoesnotexist were used to mass-create fake identities and such
Yeah, but something like that would be super easy to find and fix without going through lawsuits. And I'd argue the dataset creators would be far less likely to add copyrighted material to the training data when it's all out in the open and they can be immediately made to remove and retrain the AI without that data.
the problem with that is that training can't be done "immediately" it takes tons of compute
same