site banner

Culture War Roundup for the week of May 22, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

10
Jump in the discussion.

No email address required.

What are the limits of the weak man?

Note: Although this post cites specific real-life examples, the intent of the discussion is intended to be entirely at the meta level.

Scott Alexander's definition is apt to cite:

The straw man is a terrible argument nobody really holds, which was only invented so your side had something easy to defeat. The weak man is a terrible argument that only a few unrepresentative people hold, which was only brought to prominence so your side had something easy to defeat.

Also instructive is Bryan Caplan's gradation:

OK, what about "collective straw manning" -- questionably accusing a group for its painfully foolish positions?  Now we have:

3. Criticizing a viewpoint for a painfully foolish position no adherent holds.

4. Criticizing a viewpoint for a painfully foolish position some adherents hold.

5. Criticizing a viewpoint for a painfully foolish position many adherents hold.

6. Criticizing a viewpoint for a painfully foolish position most adherents hold.

What Caplan is describing as "collective straw manning" seems to be a good scale for weakmanning's range. And lastly, consider also Julian Sanchez's disclaimer:

With a "weak man," you don't actually fabricate a position, but rather pick the weakest of the arguments actually offered up by people on the other side and treat it as the best or only one they have. As Steve notes, this is hardly illegitimate all the time, because sometimes the weaker argument is actually the prevalent one. Maybe the best arguments for Christianity are offered up by Thomas Aquinas or St. Augustine, but I doubt there are very many people who are believers because they read On Christian Doctrine. Probably this will be the case with some frequency, if only because the less complex or sophisticated an argument is, the easier it is for lots of people to be familiar with it. On any topic of interest, a three-sentence argument is unlikely to be very good, but it's a lot more likely to spread.

At least in theory, I think weakmanning should be avoided, but I struggle with how to draw the line exactly. If your goal is to avoid weakmanning, there's at least two axes that you must consider:

  1. All the possible arguments for position X, ranked on a spectrum from least to most defensible.

  2. All the possible arguments for position X, ranked on a spectrum from least to most *representative *of believers in X.

Weakmanning is not much of an issue if you're arguing against a single individual, because they either endorse the particular arguments or not. You can't showcase the error of one's ways by refuting arguments they never held.

But generally we tend to argue over positions endorsed by many different people, where each person may differ with regard to which argument they either advance or prioritize, so what should count as "representative"?

For example, many people believe in the theory of evolution, but some believers do so under the erroneous belief that evolutionary change occurs within an individual organism's lifespan. [I know some smartass in the comments will pipe up about some endangered tropical beetle or whatever does demonstrate "change-within-lifespan" evolutionary changes. Just remember that this is not an object-level discussion.] If you use a crude heuristic and only poll relevant experts (e.g. biology professors) you're not likely to encounter many adherents of the "change-within-lifespan" argument, so this could be a decent filter to narrow your focus on what should count as "representative" for a given position. This is generally an effective tactic, since it helps you avoid prematurely declaring victory at Wrestlemania just because you trounced some toddlers at the playground.

But sometimes you get a crazy position believed by crazy people based on crazy arguments, with a relatively tiny minority within/adjacent to the community of believers aware of the problems and doing the Lord's work coming up with better arguments. InverseFlorida coined the term "sanewashing" to describe how the meaning of "defund the police" (DTP) shifted [TracingWoodgrains described the same dynamic with the gentrification of /r/antiwork. Credit also to him for most of the arborist-themed metaphor in this post.] to something much more neutered and, correspondingly, much more defensible:

So, now say you're someone who exists in a left-adjacent social space, who's taken up specific positions that have arrived to you through an "SJW" space, and now has to defend them to people who don't exist in any of your usual social spaces. These are ideas that you don't understand completely, because you absorbed them through social dynamics and not by detailed convincing arguments, but they're ones you're confident are right because you were assured, in essence, that there's a mass consensus behind them. When people are correctly pointing out that the arguments behind the position people around your space are advancing fail, but you're not going to give up the position because you're certain it's right, what are you going to do? I'm arguing you're going to sanewash it. And by that I mean, what you do is go "Well, obviously the arguments that people are obviously making are insane, and not what people actually believe or mean. What you can think of it as is [more reasonable argument or position than people are actually making]".

Keep in mind that this is not an object-level discussion on the merits of DTP. Assume arguendo that the "sanewashed" arguments are much more defensible than the "crazy" ones they replaced. If someone were to take a position against DTP by arguing against the now obsolete arguments, one of the sanewashers would be technically correct accusing you of weakmanning for daring to bring up that old story again. This fits the literal definition of weakmanning after all.

As Sanchez noted above, for most people for most positions, intuition predates rationality. They stumble around in the dark looking for any sort of foothold, then work backwards to fill in any necessary arguments. Both the sanewashers and the crazies are reliant on the other. Without the sanitization from the hygiene-minded sanewashers, the position would lack the fortification required to avoid erosion; and without the crazy masses delivering the bodies and zeal, the position would fade into irrelevance. The specific ratio may vary, but this dynamic is present in some amount on any given position. You very likely have already experienced the embarrassment that comes from a compatriot, purportedly on your side, making an ass of both of youse with their nonsensical arguments.

If your ultimate goal is truth-seeking, weakmanning will distract you into hacking away at worthless twigs rather than striking at the core. But sometimes the goal isn't seeking truth on the specific position (either because it's irrelevant or otherwise already beyond reasonable dispute) and instead the relevant topic is the collective epistemological dynamics [I dare you to use this phrase at a dinner party without getting kicked out.]. InverseFlorida's insightful analysis would not have been possible without shining a spotlight on the putative crazies — the very definition of weakmanning in other words.

Here's the point, at last. Normally someone holding a belief for the wrong reasons is not enough to negate that belief. But wherever a sanewasher faction appears to be spending considerable efforts cleaning up the mess their crazy neighbors keep leaving behind, it should instigate some suspicion about the belief, at least as a heuristic. Any honest and rational believer needs to grapple for an explanation for how the crazies managed to all be accidentally right despite outfitted — by definition — with erroneous arguments. Such a scenario is so implausible that it commands a curious inquiry about its origin.

It's possible that this inquiry unearths just another fun episode in the collective epistemological dynamics saga; it's also possible the probe ends up exposing a structural flaw with the belief itself. In either circumstances, a weakmanning objection is made in bad faith and intended to obfuscate. Its only purpose is to get you to ignore the inconvenient, the annoying. You should pay no heed to this protest and continue deploying the magnifying glass; don't be afraid to focus the sun's infernal rays into a burning pyre of illumination. Can you think of any reasons not to?

I feel about this post the way I feel about articles that say wine or chocolate in moderation has mild health benefits. Maybe there are some people who would benefit from adding a small amount of dark chocolate to their diet and this is valuable information to them, but most people are going to use that information to justify excessive consumption. Maybe there are some people who are so devoted to steel manning that they're missing out on important insights because they accept too many bad-faith weakman objections, but most people need to be pushed to focus on their opponent's best arguments. There are many, many sites on the internet that can be described as a magnifying glass focused on the outgroup's crazies, and most of them produce circle-jerks and dunk contests rather than a burning pyre of illumination. There's no alpha left in trying to detect structural flaws in your opponent's position based on their dumbest arguments.

There might be a tiny tiny bit of alpha in trying to explain the outgroup's collective epistemological dynamics, but "my opponents say they believe this because of x, but they really are motivated by y" is not exactly an untapped field of inquiry online either.

Any honest and rational believer needs to grapple for an explanation for how the crazies managed to all be accidentally right despite outfitted — by definition — with erroneous arguments. Such a scenario is so implausible that it commands a curious inquiry about its origin.

I don't think this is as unlikely as you say. Many political issues are directional in the near term (e.g. should taxes/welfare/prison sentence length go slightly up or down relative to status quo). Many crazies who you disagree with about the optimal tax level are going to end up on your side of the "should taxes go up or down' debate. Your opponents and engagement-driven social media have strong incentives to emphasize the crazies on your side, and you have a strong incentive to downplay their extremity by sane-washing them.

I feel about this post the way I feel about articles that say wine or chocolate in moderation has mild health benefits. Maybe there are some people who would benefit from adding a small amount of dark chocolate to their diet and this is valuable information to them, but most people are going to use that information to justify excessive consumption.

https://slatestarcodex.com/2014/03/24/should-you-reverse-any-advice-you-hear/

Qua Scott, there are many pathologically selfish and selfless people in the world. Selfish people need to hear the message that other people's preferences matter and you can't always put yourself first. Selfless people need to hear the message that it is sometimes okay to put yourself first. But group A might hear the message/content intended for group B and use it to justify their destructive behaviour, and vice versa.

I feel this way about a lot of modern "self-care" content, which seems like it was (at least initially) intended to give some measure of succour to genuine victims of child abuse, bullying or similar, but was quickly co-opted by selfish narcissists (invariably claiming to suffer from some nebulously defined "trauma") to rationalize their self-absorption and inconsiderate behaviour. Or indeed a lot of journalistic content about the dangers of excessive exercise, which may be useful advice to the handful of legitimate fitness freaks out there, but likely ends up being consumed by the obese and sedentary as a reason not to go for a walk.