FaceDeer

joined 6 months ago
[–] [email protected] 1 points 58 minutes ago

Things change. There was a period before this information was easily available; this repository only goes back to 2013. Now there's a period after this information, too. Things start and eventually they end.

Here's hoping that some neat new things start up in its place.

[–] [email protected] 1 points 2 hours ago

They're not both true, though. It's actually perfectly fine for a new dataset to contain AI generated content. Especially when it's mixed in with non-AI-generated content. It can even be better in some circumstances, that's what "synthetic data" is all about.

The various experiments demonstrating model collapse have to go out of their way to make it happen, by deliberately recycling model outputs over and over without using any of the methods that real-world AI trainers use to ensure that it doesn't happen. As I said, real-world AI trainers are actually quite knowledgeable about this stuff, model collapse isn't some surprising new development that they're helpless in the face of. It's just another factor to include in the criteria for curating training data sets. It's already a "solved" problem.

The reason these articles keep coming around is that there are a lot of people that don't want it to be a solved problem, and love clicking on headlines that say it isn't. I guess if it makes them feel better they can go ahead and keep doing that, but supposedly this is a technology community and I would expect there to be some interest in the underlying truth of the matter.

[–] [email protected] 11 points 7 hours ago

No, researchers in the field knew about this potential problem ages ago. It's easy enough to work around and prevent.

People who are just on the lookout for the latest "aha, AI bad!" Headline, on the other hand, discover this every couple of months.

[–] [email protected] 4 points 7 hours ago (2 children)

AI already long ago stopped being trained on any old random stuff that came along off the web. Training data is carefully curated and processed these days. Much of it is synthetic, in fact.

These breathless articles about model collapse dooming AI are like discovering that the sun sets at night and declaring solar power to be doomed. The people working on this stuff know about it already and long ago worked around it.

[–] [email protected] 13 points 1 day ago

Sometimes headshots develop spontaneously. It's a rare condition, but convenient. Some claim John F. Kennedy suffered from this condition.

[–] [email protected] 12 points 1 day ago (2 children)

Last I heard they hadn't found the knife yet.

[–] [email protected] 13 points 1 day ago (1 children)

I recall seeing a list of the most dangerous jobs in America and "President of the United States" topped it due to the high percentage of people with that job who've been shot.

[–] [email protected] 6 points 3 days ago

But at least that crappy bug-riddled code has soul!

[–] [email protected] 2 points 3 days ago (1 children)

But yeah I mean there probably would be some survivors.

This is literally the whole point I'm making. I really don't get the downvotes, it seems perfectly straightforward.

[–] [email protected] 0 points 3 days ago

I'm not Malthusian. What does Malthusianism have to do with this?

[–] [email protected] -5 points 3 days ago (3 children)

It's very straightforward math based on the article you posted. It's not saying that a nuclear war wouldn't be bad, or shouldn't be avoided. Of course that should be avoided.

My issue is with the people who insist that humanity as a species is at risk from nuclear war. That's the part that's wrong.

view more: next ›