this post was submitted on 06 Jul 2024
142 points (90.3% liked)
Technology
59608 readers
3435 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Now this sounds like a complicated debate - but it seems to me like everyone against this bill are people who would benefit monetarily from not having to deal with the safety aspect of AI, and that does sound suspicious to me.
In regards to the open source models, while it makes sense that if a developer takes the model and does a significant portion of the fine tuning, they should be liable for the result of that...
But should the main developer still be liable if a bad actor does less than 25% fine tuning and uses exploits in the base model?
One could argue that developers should be trying to examine their black-boxes for vunerabilities, rather than shrugging and saying it can't be done then demanding they not be held liable.
This kind of goes against the model that open source has operated on for a long time, as providing source doesn't represent liability. So providing a fine-tuned model shouldn't either.
I didn't mean in terms of providing. I meant that if someone provided a base model, someone took that, built upon it, then used it for a harmful purpose - of course the person modified it should be liable, not the base provider.
It's like if someone took a version of Linux, modified it, then used that modified version for an illegal act - you wouldn't go after the person who made the unmodified version.
You wouldn't necessarily punish the person that modified Linux either, you'd punish the person that uses it for a nefarious purpose.
Important distinction is the intention to deceive, not that the code/model was modified to be able to be used for nefarious purposes.