this post was submitted on 07 Dec 2023
292 points (96.8% liked)

Technology

59424 readers
2893 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
all 50 comments
sorted by: hot top controversial new old
[–] [email protected] 63 points 11 months ago

Surprised no one

[–] [email protected] 50 points 11 months ago* (last edited 11 months ago) (4 children)

So I assume they added any necessary stuff to the TOS to allow this.

My question is if there's any legal mechanism to prevent this on other platforms? Pixelfed for example.

Companies will likely federate and pull images regardless, but can we go after them when they're caught? Nothing prevents them from taking the images for internal R&D, but at least we can stop them from selling products with that training data

[–] [email protected] 10 points 11 months ago (3 children)

My question is if there’s any legal mechanism to prevent this on other platforms? Pixelfed for example.

Good question!

I’ve been saying for a while that the fediverse is blind to this issue as everything here is completely scrapable through either the public web or by running federated servers. On top of that, being culturally inclined toward more “serious” conversation and providing content warnings and alt-text for images, we’re probably generating relatively valuable training data.

And yet everything is public as though it’s still 2012.

There are alternatives. BlueSky for instance is basically private to members only. They recently announced that content would be made public to the web and a number of users were upset.

Group chats and Discord servers are probably similar, and from what I can tell “new” popular places for social activity online.

A major issue the fediverse has, IMO, is that it’s kinda stuck trying to fight Twitter and Facebook circa 2012, when that battle was lost and we’re on to new battle fronts now.

[–] [email protected] 4 points 11 months ago

Yea that's something that's been on my mind as well

There are benefits from that openness and verifiability in public spaces (ex. Lemmy communities), since now it's easier to determine if there's vote manipulation or astroturfing. But I think the fediverse needs a lot of work around privacy, and also education about what is/isn't private on these platforms.

There should also be more of a focus on setting up a legal requirement on what can be done with the information, but I'm not sure if that's a thing just yet. We developed GPLv3 to make sure FOSS products can't be incorporated for profit, but I'm not sure how it would work for data.

ex. It should be easy to save, record, and share posts on the fediverse, such as with embeds/screenshots/news stories

But also we want to prevent abuse, misuse, and AI training

[–] [email protected] 1 points 11 months ago (1 children)

Bluesky being only accessible by members doesn't completely prevent the content from being scraped by bots, though. Bots can be given user access in Bluesky too, and bots can read posts, create own posts and scrape posts and user profiles.

[–] [email protected] 1 points 11 months ago

My main point with BlueSky was that many of the users there had gotten quite comfortable with what appeared to be their closed/private space, which, despite examples like yours, was relatively true compared to the norms of Twitter and Mastodon.

The point was that many over there seemed to like it, and, if a BlueSky competitor opened up today promising all the same stuff but closed/private with the ability to opt out and make something public, many would probably jump ship or demand the same from BlueSky.

[–] [email protected] 1 points 11 months ago* (last edited 11 months ago)

afaik activitypub/fediverse doesn’t have to be fully open… there’s private messages and followers only profiles on mastodon… sure, any server admins of your followed would be able to see anything you post (and thus in this case for threads for example, if you accept any follower from threads then meta can see your stuff) but this also doesn’t grant them a license to use the content

also, bluesky will eventually be the same: it only doesn’t have those issues now because they haven’t opened up their software… it’ll have federation in the future, which means it has to be somewhat programmatically open to others

[–] [email protected] 3 points 11 months ago

I think in order to fight against these composite using our data for AI training we souks have to do something like watermarking our images explicitly stating that they are not for AI training. Or we create some type of counter measure that messes up the training.

[–] [email protected] -1 points 11 months ago* (last edited 11 months ago) (1 children)

You're never going to get rights over the training data your pictures that are freely available for anything to scan creates. By being on the internet your pictures basically have the right to be viewed by anyone or anything even an AI. You have never gotten to control who looks at your content after you post it.

You're trying to make the same argument the "don't copy my nft" bros tried to make.

Imagine going into court and saying you should get paid for all the stuff u gave away for free on the Internet willingly.

[–] [email protected] 7 points 11 months ago* (last edited 11 months ago) (1 children)

Well there's a difference between "don't look at my work without paying me, even if it's posted publicly" and "don't sell my work without paying me, even if it's posted publicly"

Like I said, there's nothing we can do about companies using all the data they can get their hands on for private R&D. It IS possible to protect against the second case, where companies can't sell an LLM product with copyrighted training data.

My question was about how that second case could be extended to stuff posted on the Fediverse, such as if an instance had a blanket "all rights belong to the user posting the content".

These laws exist, if companies can use them then so can we

[–] [email protected] 40 points 11 months ago (1 children)

That ought to satisfy all those who wanted "consent" for training data.

[–] [email protected] 18 points 11 months ago (16 children)

I wonder how they worked around user violations of copyright... Imagine all the content uploaded to Instagram/Facebook that the poster didn't create but simply uploaded their download/screenshot.

load more comments (16 replies)
[–] [email protected] 30 points 11 months ago (5 children)

Is there an example of AI generated images that aren't hyper realistic or have perfect bokeh? I'm taking about an out of focus shot or the subject looks like a regular slob like you and I?

[–] [email protected] 24 points 11 months ago (1 children)

It's mostly bias in the training data. Most people aren't posting mediocre images of themselves online so models rarely see that. Most are also finetuned to specifically avoid outputting that kind of stuff because people don't want it.

Out of focus is easy for most base models but getting an average looking person is harder.

[–] [email protected] 7 points 11 months ago

I would usually try to add things to the prompt you'd expect to find in a more casual scenario, like "smartphone" with half weight or something, or "video", or maybe like "Facebook". Just meta information you think attaches to more casual photos. Maybe even add "photo".

[–] [email protected] 10 points 11 months ago (1 children)

you can do that with stable diffusion and loras, yes

[–] [email protected] 3 points 11 months ago (1 children)

Loras are amazing. You can do anything or create anyone.

[–] [email protected] 2 points 11 months ago
[–] [email protected] 5 points 11 months ago

On the focus part, I've seen some impressive results from people who input a specific camera, lens, and focal distance.

[–] [email protected] 4 points 11 months ago

The models are deliberately engineered to create "good" images, just like cameras get autofocus, anti-shake and stuff. There are many tools that will auto-prettify people, not so many for the reverse.

There are enough imperfect images around for the model to know what that looks like.

[–] [email protected] 4 points 11 months ago* (last edited 11 months ago)

I assumed this was because it's making an average. Human attraction is highly sensitive to symmetry so this creates that symmerty by the way it works.

[–] [email protected] 22 points 11 months ago (1 children)

Considering most of the people on Instagram don’t even look like the photos of “themselves” that they post on Instagram, this might be an uncanny valley image generator.

[–] [email protected] 1 points 11 months ago

Thing is they have the original data aswell to train on, so the machine knows what the average of someone looks like and the average to which they change it so they could in theory have a good grasp of the uncanny valley or at least nó the gap the scale back to am original look.

[–] [email protected] 21 points 11 months ago* (last edited 11 months ago) (2 children)

All this AI photo generation is leading me to think that all imagery is going to be essentially meaningless. Is it real? Is it fake? Did a bot make it, or a human? As this tech continues to grow, i will be studying every image i come across while i ask myself those questions subconsciously.

I mean on one hand, you can "see" almost anything you can type out descriptively enough. Pretty neat! But now virtually anything can be "seen" which includes things that shouldn't be this easy to show. I'm thinking propaganda, deepfakes, blatantly making up fake news with imagery and video to back it all up. I guess we were always headed in this direction one way or another.

[–] [email protected] 7 points 11 months ago (1 children)

you can 'see' almost anything you can type out descriptively enough

A significant fraction of the population can't! https://en.m.wikipedia.org/wiki/Aphantasia

[–] [email protected] 2 points 11 months ago* (last edited 11 months ago)

Even those people who have difficulties with imagining something visually can use AI image generators somehow. As long as they can write and understand what a sentence means, they can use any sentence as a prompt to get a calculated image. You don't need any artistic talent or phantasy to get started with creating basic artificial images. That's exactly why artists around the world feel their skills are now being devalued by AI generators.

[–] [email protected] 6 points 11 months ago* (last edited 11 months ago)

The upside of generative AI better be no less than an end to toil because in the short run, they’re going to ruin the internet and prompt several genocides.

[–] [email protected] 11 points 11 months ago

Deviantart has already done this since the AI image hype train first started. Every picture by default is selected as material for AI training; pictures have to be manually deselected by the user to be excluded. And of course it's a nightmare for those with tons of art submissions.

Facebook/Instagram may end up having to something like that in the future but I doubt it until someone higher up does something about it.

[–] [email protected] 4 points 11 months ago

This is the best summary I could come up with:


Previously, Meta's version of this technology—using the same data—was only available in messaging and social networking apps such as Instagram.

Images include a small "Imagined with AI" watermark logo in the lower left-hand corner.

We put Meta's new AI image generator through a battery of low-stakes informal tests using our "Barbarian with a CRT" and "Cat with a beer" image synthesis protocol and found aesthetically novel results, as you can see above.

(As an aside, when generating images of people with Emu, we noticed many looked like typical Instagram fashion posts.)

The generator appears to filter out most violence, curse words, sexual topics, and the names of celebrities and historical figures (no Abraham Lincoln, sadly), but it allows commercial characters like Elmo (yes, even "with a knife") and Mickey Mouse (though not with a machine gun).

It doesn't seem to do text rendering well at all, and it handles different media outputs like watercolors, embroidery, and pen-and-ink with mixed results.


The original article contains 513 words, the summary contains 160 words. Saved 69%. I'm a bot and I'm open source!

[–] [email protected] 3 points 11 months ago

I wonder if they'd release the weights and training/inferencing code. They did it for LLaMA.

There's been a lot of open source alternatives to Stable Diffusion lately and it's great.