this post was submitted on 30 Jul 2023
0 points (NaN% liked)

Technology

37717 readers
409 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

Greg Rutkowski, a digital artist known for his surreal style, opposes AI art but his name and style have been frequently used by AI art generators without his consent. In response, Stable Diffusion removed his work from their dataset in version 2.0. However, the community has now created a tool to emulate Rutkowski's style against his wishes using a LoRA model. While some argue this is unethical, others justify it since Rutkowski's art has already been widely used in Stable Diffusion 1.5. The debate highlights the blurry line between innovation and infringement in the emerging field of AI art.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 1 year ago (15 children)

While some argue this is unethical, others justify it since Rutkowski's art has already been widely used in Stable Diffusion 1.5.

What kind of argument is that supposed to be? We've stolen his art before so it's fine? Dickheads. This whole AI thing is already sketchy enough, at least respect the artists that explicitly want their art to be excluded.

[–] [email protected] 0 points 1 year ago (14 children)

His art was not "stolen." That's not an accurate word to describe this process with.

It's not so much that "it was done before so it's fine now" as "it's a well-understood part of many peoples' workflows" that can be used to justify it. As well as the view that there was nothing wrong with doing it the first time, so what's wrong with doing it a second time?

[–] [email protected] 0 points 1 year ago (1 children)

Yes, it was.

One human artist can, over a life time, learn from a few artists to inform their style.

These AI setups are telling ALL the art from ALL the artists and using them as part of a for profit business.

There is no ethical stance for letting billion dollar tech firms hoover up all the art ever created to the try and remix it for profit.

[–] [email protected] 0 points 1 year ago (1 children)

No, it wasn't. Theft is a well-defined word. When you steal something you take it away from them so that they don't have it any more.

It wasn't even a case of copyright violation, because no copies of any of Rutkowski's art were made. The model does not contain a copy of any of the training data (with an asterisk for the case of overfitting, which is very rare and which trainers do their best to avoid). The art it produces in Rutkowski's style is also not a copyright violation because you can't copyright a style.

There is no ethical stance for letting billion dollar tech firms hoover up all the art ever created to the try and remix it for profit.

So how about the open-source models? Or in this specific instance, the guy who made a LoRA for mimicking Rutkowski's style, since he did it free of charge and released it for anyone to use?

[–] [email protected] 0 points 1 year ago (1 children)

Yes copies were made. The files were downloaded, one way or another (even as a hash, or whatever digital asset they claim to translate them into) then fed to their machines.

If I go into a Ford plant, take pictures of their equipment, then use those to make my own machines, it's still IP theft, even if I didn't walk out with the machine.

Make all the excuses you want, you're supporting the theft of other people's life's work then trying to claim it's ethical.

[–] [email protected] 1 points 1 year ago (1 children)

Yes copies were made. The files were downloaded, one way or another (even as a hash, or whatever digital asset they claim to translate them into) then fed to their machines.

They were put on the Internet for that very purpose. When you visit a website and view an image there a copy of it is made in your computer's memory. If that's a copyright violation then everyone's equally boned. When you click this link you're doing exactly the same thing.

[–] [email protected] 0 points 1 year ago (1 children)

By that logic I can sell anything I download from the web while also claiming credit for it, right?

Downloading to view != downloading to fuel my business.

[–] [email protected] 1 points 1 year ago (1 children)

No, and that's such a ridiculous leap of logic that I can't come up with anything else to say except no. Just no. What gave you that idea?

[–] [email protected] 0 points 1 year ago (1 children)

Because this thread was about the companies taking art feeding it into their machine a D claiming not to have stolen it.

Then you compared that to clicking a link.

[–] [email protected] 1 points 1 year ago

Yes, because it's comparable to clicking a link.

You said:

By that logic I can sell anything I download from the web while also claiming credit for it, right?

And that's the logic I can't follow. Who's downloading and selling Rutkowski's work? Who's claiming credit for it? None of that is being done in the first place, let alone being claimed to be "ok."

load more comments (12 replies)
load more comments (12 replies)