this post was submitted on 05 Dec 2023
74 points (98.7% liked)

Technology

34865 readers
41 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
all 14 comments
sorted by: hot top controversial new old
[–] [email protected] 20 points 11 months ago

If you design your testing set well enough and then don't care about the accuracy of the output, then it's not hard to get that kind of accuracy even without a brain scanner.

[–] [email protected] 11 points 11 months ago

Idk what metric they're using for "accuracy" but those images look sort of vaguely like the original at best

[–] [email protected] 11 points 11 months ago (1 children)

Well, that's scary. Now we can have mind reading robots.

[–] [email protected] 6 points 11 months ago

this season on Black Mirror, we re-write the first episode of Torchwood (the one where they briefly reanimate dead people to ask them how they died)

[–] [email protected] 4 points 11 months ago

"You are detained, you have the right to remain silent, since everything you think can be used against you"

[–] [email protected] 3 points 11 months ago* (last edited 11 months ago) (2 children)

It's not that impressive, they're just generating ChatGPT prompts then feeding those into Stable Diffusion. It's like peak AI grifting.

[–] [email protected] 3 points 11 months ago* (last edited 11 months ago)

They're hiding it behind academic language, but when you actually know how the underlying tech works, that's what they're doing.

[–] [email protected] 1 points 11 months ago (1 children)

Could you explain why this is a grift? The produced images in the example look scarily close to the original, whatever method they use to go from brainwaves -> image.

[–] [email protected] 1 points 11 months ago (1 children)

Because predictive texting and stable diffusion are both EXTREMELY easy to fudge data for. If you've ever used either you'll realize it's extremely hard to get it to do what you want and extremely easy to get it to do what it wants. All those high quality art renders you see are always texas sharpshoots where they try a bunch of random crap to see what looks good then say "Look at what it can do!" when it does.

[–] [email protected] 2 points 11 months ago (1 children)

But where exactly is the "fudging" happening?

the subjects were shown an image different from the 1,200, and their brain activity was measured under the fMRI 30 minutes to an hour later while asked to imagine what kind of image they had seen. Inputting the records, the neural signal translator then created score charts. The charts were input into another generative AI program in order to reconstruct the image, undergoing a 500-step revision process.

This sounds pretty straightforward. Even if the methods involve "fudging" and "throwing random crap at the wall", what matters in the end is the accuracy of the results, as long as there's no human-in-the-middle tweaking anything during each prediction.

[–] [email protected] 2 points 11 months ago

Today's not a good day for me so i'm not going to argue this any more, sorry.