this post was submitted on 21 Mar 2024
97 points (89.4% liked)
Technology
59398 readers
2735 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I've been calling this for awhile now.
I've been calling it the Ouroboros effect.
There's even bigger parts at play the paper didn't even dig into, and that's selective bias dye to human intervention.
See at first let's say an AI has 100 unique outputs for a given prompt.
However, humans will favor let's say half of em. Humans will naturally regenerate a couple times and pick their preferred "cream of the crop" result.
This will then ouroboros for an iteration.
Now the next iteration only has say 50 unique responses, as half of them have been ouroboros'd away by humans picking the one they like more.
Repeat, each time "half-lifing" the originality.
Over time, everything will get more abd more sameish. Models will degrade on originality as everything muddles into corporate speak.
You know how every corporate website uses the same useless "doesn't mean anything" jargon string of words, to say a lot without actually saying anything?
That's how AI is going to local minima to, as it keeps getting selectively "bred" to speak in an appealing and nonspecific way for the majority of online content.
I mean that's kind of how chatgpt is now, I've been slowly getting into llama, can't quite get it as good as gpt yet but I'm learning