this post was submitted on 24 Jan 2024
270 points (90.9% liked)

Open Source

31265 readers
217 users here now

All about open source! Feel free to ask questions, and share news, and interesting stuff!

Useful Links

Rules

Related Communities

Community icon from opensource.org, but we are not affiliated with them.

founded 5 years ago
MODERATORS
 

First, they restricted code search without logging in so I'm using sourcegraph But now, I cant even view discussions or wiki without logging in.

It was a nice run

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 9 months ago (11 children)

It's an interesting debate isn't it? Does AI transform something free into something that's not? Or does it simply study the code?

[–] [email protected] 5 points 9 months ago* (last edited 9 months ago) (7 children)

There's no debate. LLMs are plagiarism with extra steps. They take data (usually illegally) wholesale and then launder it.

A lot of people have been doing research into the ethics of these systems and that's more or less what they found. The reason why they're black boxes is precisely the reason we all suspected; they were made that way because if they weren't we'd all see them for what they are.

[–] [email protected] 7 points 9 months ago (2 children)

The reason they're black boxes is because that's how LLMs work. Nothing new here, neural networks have been basically black boxes for a long time.

[–] [email protected] 5 points 9 months ago* (last edited 9 months ago) (1 children)

Sure, but nothing is theoretically stopping them from documenting every single data source input into the training module and then crediting it later.

For some reason they didn't want to do that of course.

[–] [email protected] 1 points 9 months ago

Llama and stability AI published their sources, did they not?

load more comments (4 replies)
load more comments (7 replies)