EvilMonkeySlayer

joined 1 year ago
[–] [email protected] 27 points 8 months ago (3 children)

I think it's those stupid hard coded buttons on my remote that I accidentally press every so often then have to repeatedly try and back/exit out of the stupid thing it launched that I cannot remove/uninstall from my tv.

[–] [email protected] 1 points 8 months ago

Oh, I've broken you.

[–] [email protected] 1 points 8 months ago

Pointing out your arguments inconsistency is forgetting?

Are you okay?

[–] [email protected] 1 points 8 months ago (4 children)

My man, now you're just trying to put the onus on me.

Which is it?

Is it they don't retain or they do?

You made the claim. 🤷‍♂️

[–] [email protected] 1 points 8 months ago (6 children)

So, now it's gone from "reasonable effort" to most definitely you can say without any doubt that all the trained models contain no copyrighted data at all?

Come on. Make up your mind.

[–] [email protected] 1 points 8 months ago (8 children)

So, me pointing out the flaw in your argument is trolling?

What?

If you choose to use weasel wording to try and get out of something that is your call.

[–] [email protected] 1 points 8 months ago (11 children)

You're contradicting yourself.

In one sentence you say it doesn't memorize (with "reasonable effort") then in the next you admit it does.

"Reasonable effort" is weasel wording.

Make up your mind.

[–] [email protected] 2 points 8 months ago

And as said they didn’t “train chat GPT on a piracy site” the scraping algorithm put some stuff form there in the training data. There is no person doing that.

"Your honour my program that I created to slurp up data from the internet using my paid for internet connection, into my AI trained model that I own and control happened to slurp up copyrighted data.. I um, it's not my fault it slurped up copyrighted data even though I put no checks in place for it to check what it was slurping up or from where."

That is the argument you are putting forth.

Do you think any judge/court of law would view that favourably?

[–] [email protected] 1 points 8 months ago (13 children)

It contains large parts of the data in order to create. In my link I provided it shows that the models do contain chunks of the original works.

Otherwise, how would it create the words etc.

I am amazed that we now have people on the level of crypto coin idiocy going on about ai models who don't understand this.

[–] [email protected] 3 points 8 months ago

This entire comment screams of 0 technical knowledge

Yes, your comment does.

There is literally software to extract this stuff from models now.

This "it's just math" is techbro idiocy. It's like the idiots regurgitating crypto coin bullshit.

[–] [email protected] 10 points 8 months ago (25 children)

No it doesn’t, the training data isn’t inside the LLM.

This is factually incorrect. You can extract the data. How do you think the legal cases are being brought?

For example

The model has to contain the data in order to produce works.

Wholesale commercial copyright infringement where you're profiting off of others work on a large scale is a whole different ball game.

They're training their models on large amounts of pirated content and profiting off it.

Of course the rights holders are going to say "wait a minute, why are you making money off my content without my permission? And how much of my work did you pirate to use?"

You cannot hand wave away mass piracy to train their models, and then distribute said models based on an act of mass copyright infringement.

Do you not understand the basics of the law?

its idiotic to think that its reasonable to demand such a thing.

Again, the law is the law. If they mass pirate a bunch of media which then the model contains chunks of they are breaking the law.

I can't believe this is a hard concept for someone to understand.

[–] [email protected] 12 points 8 months ago (40 children)

How so?

The trained model includes vast swathes of copyrighted material. It's the rights holders who get to decide whether someone can use it.

Just because it makes it inconvenient or harder for someone to train an AI model does not justify wholesale stealing.

A lot of models are even trained on large numbers of pirated material like books downloaded from pirate sites etc. I guarantee you OpenAI and others didn't even buy a lot of the material they use to train the AI models on.

view more: next ›