this post was submitted on 22 Dec 2023
861 points (96.4% liked)

Technology

59594 readers
2911 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

More than 200 Substack authors asked the platform to explain why it’s “platforming and monetizing Nazis,” and now they have an answer straight from co-founder Hamish McKenzie:

I just want to make it clear that we don’t like Nazis either—we wish no-one held those views. But some people do hold those and other extreme views. Given that, we don’t think that censorship (including through demonetizing publications) makes the problem go away—in fact, it makes it worse.

While McKenzie offers no evidence to back these ideas, this tracks with the company’s previous stance on taking a hands-off approach to moderation. In April, Substack CEO Chris Best appeared on the Decoder podcast and refused to answer moderation questions. “We’re not going to get into specific ‘would you or won’t you’ content moderation questions” over the issue of overt racism being published on the platform, Best said. McKenzie followed up later with a similar statement to the one today, saying “we don’t like or condone bigotry in any form.”

(page 3) 50 comments
sorted by: hot top controversial new old
[–] [email protected] 14 points 11 months ago

This is the first time I heard of the platform and I intend to keep it that way

[–] [email protected] 14 points 11 months ago (1 children)

I want them to explain how it makes things worse.

[–] [email protected] 13 points 11 months ago

"We would make less money, and that's worse than more money."

[–] [email protected] 8 points 11 months ago* (last edited 11 months ago) (1 children)

Submitted for good faith discussion: Substack shouldn’t decide what we read. The reason it caught my attention is that it's co-signed by Edward Snowden and Richard Dawkins, who evidently both have blogs there I never knew about.

I'm not sure how many of the people who decide to comment on these stories actually read up about them first, but I did, such as by actually reading the Atlantic article linked. I would personally feel very uncomfortable about voluntarily sharing a space with someone who unironically writes a post called "Vaccines Are Jew Witchcraftery". However, the Atlantic article also notes:

Experts on extremist communication, such Whitney Phillips, the University of Oregon journalism professor, caution that simply banning hate groups from a platform—even if sometimes necessary from a business standpoint—can end up redounding to the extremists’ benefit by making them seem like victims of an overweening censorship regime. “It feeds into this narrative of liberal censorship of conservatives,” Phillips told me, “even if the views in question are really extreme.”

Structurally this is where a comment would usually have a conclusion to reinforce a position, but I don't personally know what I support doing here.

[–] [email protected] 14 points 11 months ago* (last edited 11 months ago)

IDGAF if it feeds into the narrative. It also shuts down a recruitment pipeline. It reduces their reach. It makes the next generation less likely to continue the ideology. De-platforming is a powerful tool that should be reserved for only the most crucial fights, but the fight against Nazi is one of those fights.

The Nazis were already full-blown conspiracy theorists. EVERYTHING is spun to feed into their narrative. That ship has sailed.

A platform operator needs to AT MINIMUM demonetize the content and censure it, and is likely only being responsible if they ban it outright. If you aren't prepared to wade into the fraught, complex world of content moderation, don't run a content platform.

[–] [email protected] 8 points 11 months ago

Now I'm curious if anybody adds something about LGBTQ on there, if that's promptly removed.

load more comments
view more: ‹ prev next ›