this post was submitted on 17 Mar 2025
496 points (99.2% liked)

Technology

66892 readers
4960 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

The Chinese Communist Party’s (CCP’s) national internet censor just announced that all AI-generated content will be required to have labels that are explicitly seen or heard by its audience and embedded in metadata. The Cyberspace Administration of China (CAC) just released the transcript for the media questions and answers (akin to an FAQ) on its Measures for the Identification of Artificial Intelligence Generated and Synthetic Content [machine translated]. We saw the first signs of this policy move last September when the CAC’s draft plans emerged.

This regulation takes effect on September 1, 2025, and will compel all service providers (i.e., AI LLMs) to “add explicit labels to generated and synthesized content.” The directive includes all types of data: text, images, videos, audio, and even virtual scenes. Aside from that, it also orders app stores to verify whether the apps they host follow the regulations.

Users will still be able to ask for unlabeled AI-generated content for “social concerns and industrial needs.” However, the generating app must reiterate this requirement to the user and also log the information to make it easier to trace. The responsibility of adding the AI-generated label and metadata falls on the shoulders of this end-user person or entity.

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 30 points 22 hours ago* (last edited 22 hours ago) (6 children)
[–] [email protected] 9 points 14 hours ago

Except when the government is immune to this and passes things off as real since it isn't marked as AI.

[–] [email protected] 19 points 20 hours ago

yeah, if im going to live in a hypercapitalist shit hole where the internet is rabidly censored and there are no environmental protections, I'd rather have watermarks on the AI slop than nazis everywhere.

[–] [email protected] 2 points 14 hours ago

not rare. it's a tool that will be used to silence critics.

[–] [email protected] 6 points 20 hours ago

Not really rare these days when you compare them to America

[–] [email protected] 2 points 21 hours ago (1 children)

For all the humans rights abuses, one has to admit that China is at least ruthlessly efficient.

[–] [email protected] 3 points 21 hours ago (1 children)

Not really. Sure, China is able to make unpopular decisions better then democracies, but that makes them inefficient in different directions. E.g. high speed rail in areas where it is not needed but greatly lacking freight trains. Or their housing bubble.

[–] [email protected] 8 points 21 hours ago

China no doubt has it's problems. It's just crazy to think how fast the country has progressed in the last 50 years.

[–] [email protected] 0 points 14 hours ago

you don't think this will be exploited to hurt people?

[–] [email protected] 13 points 20 hours ago
[–] [email protected] 17 points 21 hours ago* (last edited 21 hours ago) (4 children)

Anyone's praising this doesn't understand that this request is basically impossible and is merely posturing.

I'm a developer and I work a lot with LLM data and the only way to detect LLM text is through watermarks where some words or expressions are statistically preferred over others. This means it's only effective on large bodies of text that are not modified further.

If you take LLM content and remix it using traditional natural language processing then it's done - the content is indistinguishable and untraceable and it takes like 50 lines of python code and a few milliseconds of computing.

[–] [email protected] 7 points 14 hours ago (1 children)

I think you're looking past the real reason why this is happening.

it's not to improve the quality of information.

it's to silence anyone with dissenting opinions of the Chinese Government.

it's easy to label something. some Chinese citizen posts something about the government, tagged as AI, they're fined, jailed, etc.

[–] [email protected] 3 points 5 hours ago

100% thats exactly what's happening and I can't believe people are so blinded by AI generato hate to praise stuff like that.

[–] [email protected] 15 points 20 hours ago (1 children)

A government policy isn’t just posturing because the state now has a rule to cite if they’re gonna issue you a fine or whatever the punishment is supposed to be. So you will either comply, or go underground or abroad. That’s a real consequence.

[–] [email protected] 2 points 15 hours ago (2 children)

So it's just another way for authoritarians to exert power over people then?

The only way to address AI is through low level laws we already have like anti-discrimination, defamation, online bullying etc. But those give people more rights and protections and you can't have that.

[–] [email protected] 2 points 13 hours ago

it’s just another way for authoritarians to exert power over people then?

exert power over people to stop them from doing what?

[–] [email protected] 1 points 12 hours ago

I mean you can just say that about any and every law or policy. No need to be so knee-jerk about it. The point I'm making is it isn't just posturing. It's not like a company pretending to promise to watermark their AI outputs; it's a government saying you must comply with new rule.

[–] [email protected] 1 points 20 hours ago

I would love to delve into this a little more.

[–] [email protected] 0 points 20 hours ago (1 children)

It's impossible because the text these LLM-based models produce would be obtuse to watermark.

Huh?

What about photos and video and audio?! Why are you asking?

[–] [email protected] 1 points 15 hours ago (2 children)

What about photos, videos and audio? You should see what the second L means in the LLM before you go at it

[–] [email protected] 3 points 13 hours ago

The directive includes all types of data: text, images, videos, audio, and even virtual scenes.

LLMs are only one aspect of this, but yeah, probably the most difficult to discern, at least at the moment.

[–] [email protected] 1 points 13 hours ago

You should read some of the content you're commenting on before posting a critique.

[–] [email protected] 7 points 23 hours ago

I suggest the shit emjoi being used as the indicator.

[–] [email protected] 3 points 21 hours ago

Too late now, pal. You've all poisoned the well and now you have to drink it too.

[–] [email protected] 34 points 1 day ago

Meanwhile best we can do in America is hide tracking dots in every color printer.

[–] [email protected] 83 points 2 days ago (4 children)

...I'm...

In full agreement with this*

*with the provision that there are ways to ensure this isn't weaponized so that dissident or oppositional speech/photos/art isn't flagged as AI so that it can be filtered out.

[–] [email protected] 4 points 14 hours ago (1 children)

There are no ways of ensuring that. Wanting this is suicide for anyone but authoritarians.

[–] [email protected] 1 points 14 hours ago

Then it should not be done. I laid out my conditionals for it not being terrible.

[–] [email protected] 8 points 1 day ago (1 children)

Or propaganda that doesn’t have it is taken as legitimate.

[–] [email protected] 4 points 1 day ago

That doesn't change anything though.

[–] [email protected] 14 points 2 days ago* (last edited 1 day ago) (1 children)

It doesn't matter whether this is used against dissidents or not. Their speech is censured either way. It shouldn't affect the much larger positive effect this will have on the majority of people.

[–] [email protected] 5 points 1 day ago

This does provide another tool for them to claim it isn't censored but label it as AI to hurt the credibility of dissidents though. I don't think it doesn't matter.

[–] [email protected] 8 points 2 days ago (1 children)

So in short you disagree which is reasonable given the circumstances.

Besides, wouldn‘t it make much more sense to verify and mark genuine content rather than the slob which is becoming the majority of content?

[–] [email protected] 5 points 1 day ago

I like that approach better. Just like I'd rather know what doesn't cause cancer in the state of California at this point.

[–] [email protected] 43 points 2 days ago (1 children)

When the dirty commies do the reforms we all know we need in our countries...

We're so fucked. ⚰️

[–] [email protected] 8 points 1 day ago

As a dirty commie: you’ll get over it someday.

[–] [email protected] 22 points 2 days ago* (last edited 2 days ago) (2 children)

Not a bad law if applied to companies and public figures. Complete wishful thinking if applied to individuals.

For companies it's actually enforceable but for individuals it's basically impossible and even if you do catch someone uploading AI-generated stuff: Who cares. It's the intent that matters when it comes to individuals.

Were they trying to besmirch someone's reputation by uploading false images of that person in compromising situations? That's clear bad intent.

Were they trying to incite a riot or intentionally spreading disinformation? Again, clear bad intent.

Were they showing off something cool they made with AI generation? It is of no consequence and should be treated as such.

[–] [email protected] 9 points 1 day ago (4 children)

I agree that it's difficult to enforce such a requirement on individuals. That said, I don't agree that nobody cares for the content they post. If they have "something cool they made with AI generation" - then it's not a big deal to have to mark it as AI-generated.

[–] [email protected] 5 points 1 day ago

Notice: Those are not my girlfriend's boobs. I used Photoshop with an AI plug in to make them look fuller.

No thanks mate. Government and anyone selling anything should be held to those standards. If you are an influencer pushing a product for profit that applies to you too.

load more comments (3 replies)
[–] [email protected] 1 points 1 day ago (1 children)

Would applying a watermark to all the training images force the AI to add a watermark?

[–] [email protected] 1 points 23 hours ago

Nope. In fact, if you generate a lot of images with AI you'll sometimes notice something resembling a watermark in the output. Demonstrating that the images used to train the model did indeed have watermarks.

Removing such imaginary watermarks is trivial in image2image tools though (it's just a quick extra step after generation).

[–] [email protected] 9 points 1 day ago (1 children)

They don't want to pollute their training data.

[–] [email protected] 20 points 1 day ago

Honestly?

Good. I assume this is more about controlling narratives but it's something that should be happening no matter what side of the AI debate you're on.

[–] [email protected] 6 points 1 day ago* (last edited 1 day ago) (1 children)

My favorite genre of comment section is when every other post is talking about how someone/thing the poster doesn't like does something they think is good but they gotta reassure everyone that it'll still be bad.

Yeah, he saved the kitten from the tree, But at what cost? 😔

load more comments
view more: next ›