site banner

Culture War Roundup for the week of January 30, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

13
Jump in the discussion.

No email address required.

New ACX post: Mostly Skeptical Thoughts On The Chatbot Propaganda Apocalypse, discuss!

I'll begin: this quote from philosophybear (what's his account here btw?):

The capacity of the wealthy to command vast armies of bots (GPU’s to run machine learning are expensive) will further erode what “democracy” there is on the internet. If fee structures are bought in like I described to keep bots out, that will make the internet less democratic too. My advice is to log off and start forming connections and organizations in real life now.

... made my inner Curtis Yarvin giggle hysterically. Observe the uncanny uniformity of ideological positions on pretty much all social issues in most large newspapers, most top universities, most large corporations. The only deviations are of the "we need fifty Stalins" kind (until they become the norm eventually). Can you imagine a Harvard professor, a New York Times editor, and the Raytheon PR department having a substantial disagreement on whether we need more trans drone pilots? I can't.

And it took people the possibility of bots faking a pale shadow of such consensus to start worrying that maybe democracy is susceptible to being secretly not the rule of the people but the rule of whoever tells the people what to believe, and the unbelievable synopticity of what we are told to believe means that this rule is being actively exercised?

Later Scott sort of touches on this a bit quoting a comment that said:

On the 'disinformation vs. establishment bot' question, check out bots interacting with climate change: 83.1% of bot tweets support activism, 16.9% skepticism according to https://www.sciencedirect.com/science/article/pii/S1674927821001490 .

The abstract ends with:

Based on the above findings, we suggest cultivating individuals’ media literacy in terms of distinguishing malicious social bots as a potential solution to deal with social bot skeptics disguised as humans, as well as making use of benign social bots for science popularization.

My skepticism to the "propaganda chatocalypse" is that is that it operates on a flawed model of how humans operate. That are humans view of the world is hopelessly tied to cultural consumption, media and so on. So our inputs from culture and media needs to be controlled at all costs so we don't end up making the world worse by being exposed to dangerous ideas cultural and/or media content. This is the quintessential post-modernistic thought process that if they stuff the "the message" in everything it will create a better world. But we already did this experiment with the totalitarian states of the 20th century and people still thought differently even when they were force fed with their culture and media with the right way of thinking. It didn't work and it is not going to work this time either even if we have more powerful tools.

One of the interesting and kind of worrying aspects of this though is that our media is overrun by these postmodernists that are hell bent on anthropomorphising various artifacts also. Based on that the simulacra of the human form(whether it is in characters in text, images/film, robots, simulation in games or chat AI) is somehow interchangeable with humans and somehow bestows them some rights because that real humans have them. The whole debacle around the chatbot lamda being sentient is based on this. But what about the campaign against sex robots that casually excludes the male form of robots from the campaign. The whole thing around postmodern critique of videogames, movies and books, where they try to give fictional identities inalienable rights. All of these receiving mainstream media time and not listening to anyone going like "hang on a minute, we are not talking about real humans here!". This is at the fringes and I seem to be the only who notices this...

But we already did this experiment with the totalitarian states of the 20th century and people still thought differently even when they were force fed with their culture and media with the right way of thinking. It didn't work and it is not going to work this time either even if we have more powerful tools.

It might not have worked totally, but it worked well enough to keep the totalitarian system in place for quite a while no?

Yes the systems worked for the majority but individual minds weren't captured enough not to escape or attempt to escape. Or even discussing within small groups that weren't conforming to the totalitarian governments views. The fall of the european totalitarian governments might have happened overnight but grip on the culture and minds was lost way before, despite the total control of media, culture and discourse.