this post was submitted on 04 Dec 2023
888 points (97.9% liked)

Technology

59675 readers
3086 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 280 points 11 months ago (4 children)

It's kind of odd that they could just take random information from the internet without asking and are now treating it like a trade secret.

[–] [email protected] 121 points 11 months ago (11 children)

This is why some of us have been ringing the alarm on these companies stealing data from users without consent. They know the data is valuable yet refuse to pay for the rights to use said data.

[–] [email protected] 49 points 11 months ago

Yup. And instead, they make us pay them for it. 🤡

[–] [email protected] 4 points 11 months ago* (last edited 11 months ago) (1 children)

According to most sites TOS, when we write our posts we give them basically full access to do whatever they like including make derivative works. Here is the reddit one (not sure how Lemmy handles this):

When Your Content is created with or submitted to the Services, you grant us a worldwide, royalty-free, perpetual, irrevocable, non-exclusive, transferable, and sublicensable license to use, copy, modify, adapt, prepare derivative works of, distribute, store, perform, and display Your Content and any name, username, voice, or likeness provided in connection with Your Content in all media formats and channels now known or later developed anywhere in the world. This license includes the right for us to make Your Content available for syndication, broadcast, distribution, or publication by other companies, organizations, or individuals who partner with Reddit. You also agree that we may remove metadata associated with Your Content, and you irrevocably waive any claims and assertions of moral rights or attribution with respect to Your Content.

[–] [email protected] 2 points 11 months ago* (last edited 11 months ago)

According to most sites TOS, when we write our posts we give them basically full access to do whatever they like including make derivative works.

2 points:
1 - I'm generally talking about companies extracting data from other websites, such as OpenAI scraping posts from reddit or other such postings. Companies that use their own collection of data are a very different thing.
2 - Terms of Service and Intellectual Property are not the same thing and a ToS is not guaranteed to be a fully legally binding document (the last part is the important part.) This is why services that have dealt with user created data that are used to licensing issues (think deviant art or other art hosting services) usually require the user to specify the license that they wish to distribute their content under (cc0, for example, would be fully permissible in this context.) This also means that most fan art is fair game as licensing that content is dubious at best, but raises the question around whether said content can be used to train an AI (again, intellectual property is generally different from a ToS).

It's no different from how Github's Copilot has to respect the license of your code regardless of whether you've agreed to the terms of service or not. Granted, this is legally disputable and I'm sure this will come up at some point with how these AI companies operate -- This is a brave new world. Having said that, services like Twitter might want to give second thought of claiming ownership over every post on their site as it essentially means they are liable for the content that they host. This is something they've wanted to avoid in the past because it gives them good coverage for user submitted content that they think is harmful.

If I was a company, I wouldn't want to be hinging my entire business on my terms of service being a legally binding document -- they generally aren't and can frequently be found to be unbinding. And, again, this is different from OpenAI as much of their data is based on data they've scraped from websites which they haven't agreed to take data from (finders-keepers is generally not how ownership works and is more akin to piracy. I wouldn't want to base a multinational business off of piracy.)

load more comments (9 replies)
[–] [email protected] 38 points 11 months ago (2 children)

There was personal information included in the data. Did no one actually read the article?

[–] [email protected] 31 points 11 months ago

Tbf it's behind a soft paywall

[–] [email protected] 20 points 11 months ago* (last edited 11 months ago) (1 children)

Well firstly the article is paywalled but secondly the example that they gave in this short bit you can read looks like contact information that you put at the end of an email.

[–] [email protected] 2 points 11 months ago

That would still be personal information.

[–] [email protected] 26 points 11 months ago (2 children)

They do not have permission to pass it on. It might be an issue if they didn't stop it.

[–] [email protected] 50 points 11 months ago (31 children)

As if they had permission to take it in the first place

load more comments (31 replies)
[–] [email protected] 39 points 11 months ago

In a lot of cases, they don't have permission to not pass it along. Some of that training data was copyleft!

[–] [email protected] 6 points 11 months ago

You don't want to let people manipulate your tools outside your expectations. It could be abused to produce content that is damaging to your brand, and in the case of GPT, damaging in general. I imagine OpenAI really doesn't want people figuring out how to weaponize the model for propaganda and/or deceit, or worse (I dunno, bomb instructions?)