site banner

ACX: Moderation is Different from Censorship

astralcodexten.substack.com

A brief argument that “moderation” is distinct from censorship mainly when it’s optional.

I read this as a corollary to Scott’s Archipelago and Atomic Communitarianism. It certainly raises similar issues—especially the existence of exit rights. Currently, even heavily free-speech platforms maintain the option of deleting content. This can be legal or practical. But doing so is incompatible with an “exit” right to opt back in to the deleted material.

Scott also suggests that if moderation becomes “too cheap to meter,” it’s likely to prevent the conflation with censorship. I’m not sure I see it. Assuming he means something like free, accurate AI tagging/filtering, how does that remove the incentive to call [objectionable thing X] worthy of proper censorship? I suppose it reduces the excuse of “X might offend people,” requiring more legible harms.

As a side note, I’m curious if anyone else browses the moderation log periodically. Perhaps I’m engaging with outrage fuel. But it also seems like an example of unchecking (some of) the moderation filters to keep calibrated.

15
Jump in the discussion.

No email address required.

First, on a personal note, this is exactly what I stoner-hot-take predicted Musk would do with Twitter in a prior motte thread. This freaks me out. Not that it's all that creative a take, but it's something I've noticed before when I was spending too much time in narrow epistemic corners (team fan blogs, fashion blogs) where I'd start to think the same thoughts that showed up on the blogs a week later. The same trade ideas for fan blogs, or I'd pick something up at a thrift store that caught my eye and a week later it would get anointed a trend. It's sort of a weird hivemind thing, we're all thinking about the same issues based on the same influences at the same time if we're all consuming the same set of blogs and news sites. So you can look there for my joking-not-joking predictions as to how this will go, it's a good plan but it won't survive contact with the enemy (users).

Second, to address a specific point SA makes:

A minimum viable product for moderation without censorship is for a platform to do exactly the same thing they’re doing now - remove all the same posts, ban all the same accounts - but have an opt-in setting, “see banned posts”.

Twitter would be completely unusable without any bans or filters, filled with bots and scams and obvious harassment. That's not a viable product. The minimum viable product has to filter out enough to make the product usable. I don't want to see "banned" posts, I want to see posts banned for political incorrectness. Maybe. Shout out to the mods of themotte, would themotte be usable in your judgment without that kind of basic filtering?

Any Chinese person could get accurate information on Xinjiang, Tiananmen Square, the Shanghai lockdowns, or the top fifty criticisms of Xi Jinping - just by clicking a button on their Weibo profile.

"Accurate" is a point of contention here. It's not unusual to have certain topics be overwhelmingly dominated by a particularly numerous or energetic viewpoint on the topic, even just something simple like Toronto sports can get weird with sportswriters admitting they softpedaled coverage of the Raptors and Blue Jays because the articles got a ton of clicks from Canadian fans, along with a ton of comments yelling at the writer if they insulted Toronto's honor. If you're any country other than the USA, you risk being flooded by American content and American viewpoints. While I'm not defending Chinese censorship per se, I do think that saying opening up to all sources of information increases accuracy can be disputed.

First, on a personal note, this is exactly what I stoner-hot-take predicted Musk would do with Twitter in a prior motte thread. This freaks me out. Not that it's all that creative a take, but it's something I've noticed before when I was spending too much time in narrow epistemic corners (team fan blogs, fashion blogs) where I'd start to think the same thoughts that showed up on the blogs a week later.

Funnily enough, ~same for me. Tho I suggested this nearly half a year ago, so maybe that's a little different... Link

Optional moderation. Say, someone thinks /u/ZorbaTHut is great at removing content which they don't want to see. So they subscribe to his moderation services - and they don't see this content anymore. Some other people think he's doing it wrong - and they see this content.

Twitter could have default list of mods, and it would function exactly as it does now - except interested people could selectively unsubscribe from that.

Everyone wins! Except for the people who want to block people from consensually communicating with each other.

I mean, it wouldn't even be that much of a change on Reddit. Just assume Administrators do not moderate, there's no Anti-Evil Ops, subreddit mods become "default mods" - each of which could be unsubscribed for - and people could subscribe to the nonstandard mods too.

And so there would never be deleted comments/posts - only hidden from the people who don't want to see them.

.........

Shout out to the mods of themotte, would themotte be usable in your judgment without that kind of basic filtering?

Zorba responded, explaining why it might not work; not quoting it here because of the length. So probably not(?).

I don't remember it, but I wonder if I read your post and forgot about it. Cool that you already asked and answered that question.