site banner

Culture War Roundup for the week of January 30, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

13
Jump in the discussion.

No email address required.

New ACX post: Mostly Skeptical Thoughts On The Chatbot Propaganda Apocalypse, discuss!

I'll begin: this quote from philosophybear (what's his account here btw?):

The capacity of the wealthy to command vast armies of bots (GPU’s to run machine learning are expensive) will further erode what “democracy” there is on the internet. If fee structures are bought in like I described to keep bots out, that will make the internet less democratic too. My advice is to log off and start forming connections and organizations in real life now.

... made my inner Curtis Yarvin giggle hysterically. Observe the uncanny uniformity of ideological positions on pretty much all social issues in most large newspapers, most top universities, most large corporations. The only deviations are of the "we need fifty Stalins" kind (until they become the norm eventually). Can you imagine a Harvard professor, a New York Times editor, and the Raytheon PR department having a substantial disagreement on whether we need more trans drone pilots? I can't.

And it took people the possibility of bots faking a pale shadow of such consensus to start worrying that maybe democracy is susceptible to being secretly not the rule of the people but the rule of whoever tells the people what to believe, and the unbelievable synopticity of what we are told to believe means that this rule is being actively exercised?

Later Scott sort of touches on this a bit quoting a comment that said:

On the 'disinformation vs. establishment bot' question, check out bots interacting with climate change: 83.1% of bot tweets support activism, 16.9% skepticism according to https://www.sciencedirect.com/science/article/pii/S1674927821001490 .

The abstract ends with:

Based on the above findings, we suggest cultivating individuals’ media literacy in terms of distinguishing malicious social bots as a potential solution to deal with social bot skeptics disguised as humans, as well as making use of benign social bots for science popularization.

Based on the above findings, we suggest cultivating individuals’ media literacy in terms of distinguishing malicious social bots as a potential solution to deal with social bot skeptics disguised as humans, as well as making use of benign social bots for science popularization.

Clear proponents of using bots for information and cultural warfare.

As an aside, the psy-op groups within military and intelligence organisations must already be exploiting the advanced chatbots. How much of the Ukraine war propaganda is already using this praxis?

One thing that the Ukraine war has demonstrated is that Russian Bots are a paper tiger, probably.

I was led to believe that Russia had the capacity to use cyberattacks to cripple large swaths of other nations' infrastructure for minimal cost and effort.

Maybe they still do, but somehow it hasn't been brought to bear in any way that has been noticeable outside the country.

There's a wide gap between being able to disrupt infrastructure and to cripple it. Most cyberattacks a country can pull off belong to the former capacity: log into a server controlling a water treatment plant and switch everything off. Crippling something requires a much more elaborate exploit.

So even if Russia can cyberattack Ukraine, this will not have a greater effect than just dropping missiles onto its power plants. Crippling its military communications would've been really useful when the war broke out, but it seems it was beyond Russia's capabilities.

I dunno, I'd count a ransomware attack that locks operators out of control systems and access to data as 'crippling' for most pursuits, even if the physical infrastructure is unharmed.

When I worked at a power plant all of the computers that actually controlled anything were offline and had their USB ports disabled. They had to be able to update them for maintenance and stuff so I'm sure there was some way in, but if you need physical access then it's going to have to be a more targeted attack and not just uploading a virus that knocks out a whole country or region.