Technology
This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.
Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.
Rules:
1: All Lemmy rules apply
2: Do not post low effort posts
3: NEVER post naziped*gore stuff
4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.
5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)
6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist
7: crypto related posts, unless essential, are disallowed
view the rest of the comments
If the source is literally a piracy website that serves up applications on how to remove DRM from ebooks, it's absolutely piracy. You can't just deny the source and be like "it's not piracy!" The way the data came into your hands was illicitly, not legally. Especially if DRM has been circumvented and removed before it came into your hands.
They didn't go out and buy copies of thousands of books.
I don't, I was making a point about how absurdly large the language models have to be, which is to say, if they have to have that much data on top of thousands of pirated books, it means they fundamentally cannot make the models work without also scraping the internet for data, which is surveillance.
And if they went to a library and scanned all the books?
I mean, it's just not surveillance, by definition. There's no observation, just data ingestion. You're deliberately trying to conflate the words to associate a negative behavior with LLM training to make your argument.
I really don't get why LLMs get everybody all riled up. People have been running Web crawlers since the dawn of the Web.
The AI literally observes the training data
Insofar as my computer observes the data on my hard disk. But I suspect you know what I meant.