This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.
Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.
We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:
-
Shaming.
-
Attempting to 'build consensus' or enforce ideological conformity.
-
Making sweeping generalizations to vilify a group you dislike.
-
Recruiting for a cause.
-
Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.
In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:
-
Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.
-
Be as precise and charitable as you can. Don't paraphrase unflatteringly.
-
Don't imply that someone said something they did not say, even if you think it follows from what they said.
-
Write like everyone is reading and you want them to be included in the discussion.
On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.
Jump in the discussion.
No email address required.
Notes -
What are the limits of the weak man?
Note: Although this post cites specific real-life examples, the intent of the discussion is intended to be entirely at the meta level.
Scott Alexander's definition is apt to cite:
Also instructive is Bryan Caplan's gradation:
What Caplan is describing as "collective straw manning" seems to be a good scale for weakmanning's range. And lastly, consider also Julian Sanchez's disclaimer:
At least in theory, I think weakmanning should be avoided, but I struggle with how to draw the line exactly. If your goal is to avoid weakmanning, there's at least two axes that you must consider:
All the possible arguments for position X, ranked on a spectrum from least to most defensible.
All the possible arguments for position X, ranked on a spectrum from least to most *representative *of believers in X.
Weakmanning is not much of an issue if you're arguing against a single individual, because they either endorse the particular arguments or not. You can't showcase the error of one's ways by refuting arguments they never held.
But generally we tend to argue over positions endorsed by many different people, where each person may differ with regard to which argument they either advance or prioritize, so what should count as "representative"?
For example, many people believe in the theory of evolution, but some believers do so under the erroneous belief that evolutionary change occurs within an individual organism's lifespan. [I know some smartass in the comments will pipe up about some endangered tropical beetle or whatever does demonstrate "change-within-lifespan" evolutionary changes. Just remember that this is not an object-level discussion.] If you use a crude heuristic and only poll relevant experts (e.g. biology professors) you're not likely to encounter many adherents of the "change-within-lifespan" argument, so this could be a decent filter to narrow your focus on what should count as "representative" for a given position. This is generally an effective tactic, since it helps you avoid prematurely declaring victory at Wrestlemania just because you trounced some toddlers at the playground.
But sometimes you get a crazy position believed by crazy people based on crazy arguments, with a relatively tiny minority within/adjacent to the community of believers aware of the problems and doing the Lord's work coming up with better arguments. InverseFlorida coined the term "sanewashing" to describe how the meaning of "defund the police" (DTP) shifted [TracingWoodgrains described the same dynamic with the gentrification of /r/antiwork. Credit also to him for most of the arborist-themed metaphor in this post.] to something much more neutered and, correspondingly, much more defensible:
Keep in mind that this is not an object-level discussion on the merits of DTP. Assume arguendo that the "sanewashed" arguments are much more defensible than the "crazy" ones they replaced. If someone were to take a position against DTP by arguing against the now obsolete arguments, one of the sanewashers would be technically correct accusing you of weakmanning for daring to bring up that old story again. This fits the literal definition of weakmanning after all.
As Sanchez noted above, for most people for most positions, intuition predates rationality. They stumble around in the dark looking for any sort of foothold, then work backwards to fill in any necessary arguments. Both the sanewashers and the crazies are reliant on the other. Without the sanitization from the hygiene-minded sanewashers, the position would lack the fortification required to avoid erosion; and without the crazy masses delivering the bodies and zeal, the position would fade into irrelevance. The specific ratio may vary, but this dynamic is present in some amount on any given position. You very likely have already experienced the embarrassment that comes from a compatriot, purportedly on your side, making an ass of both of youse with their nonsensical arguments.
If your ultimate goal is truth-seeking, weakmanning will distract you into hacking away at worthless twigs rather than striking at the core. But sometimes the goal isn't seeking truth on the specific position (either because it's irrelevant or otherwise already beyond reasonable dispute) and instead the relevant topic is the collective epistemological dynamics [I dare you to use this phrase at a dinner party without getting kicked out.]. InverseFlorida's insightful analysis would not have been possible without shining a spotlight on the putative crazies — the very definition of weakmanning in other words.
Here's the point, at last. Normally someone holding a belief for the wrong reasons is not enough to negate that belief. But wherever a sanewasher faction appears to be spending considerable efforts cleaning up the mess their crazy neighbors keep leaving behind, it should instigate some suspicion about the belief, at least as a heuristic. Any honest and rational believer needs to grapple for an explanation for how the crazies managed to all be accidentally right despite outfitted — by definition — with erroneous arguments. Such a scenario is so implausible that it commands a curious inquiry about its origin.
It's possible that this inquiry unearths just another fun episode in the collective epistemological dynamics saga; it's also possible the probe ends up exposing a structural flaw with the belief itself. In either circumstances, a weakmanning objection is made in bad faith and intended to obfuscate. Its only purpose is to get you to ignore the inconvenient, the annoying. You should pay no heed to this protest and continue deploying the magnifying glass; don't be afraid to focus the sun's infernal rays into a burning pyre of illumination. Can you think of any reasons not to?
I feel about this post the way I feel about articles that say wine or chocolate in moderation has mild health benefits. Maybe there are some people who would benefit from adding a small amount of dark chocolate to their diet and this is valuable information to them, but most people are going to use that information to justify excessive consumption. Maybe there are some people who are so devoted to steel manning that they're missing out on important insights because they accept too many bad-faith weakman objections, but most people need to be pushed to focus on their opponent's best arguments. There are many, many sites on the internet that can be described as a magnifying glass focused on the outgroup's crazies, and most of them produce circle-jerks and dunk contests rather than a burning pyre of illumination. There's no alpha left in trying to detect structural flaws in your opponent's position based on their dumbest arguments.
There might be a tiny tiny bit of alpha in trying to explain the outgroup's collective epistemological dynamics, but "my opponents say they believe this because of x, but they really are motivated by y" is not exactly an untapped field of inquiry online either.
I don't think this is as unlikely as you say. Many political issues are directional in the near term (e.g. should taxes/welfare/prison sentence length go slightly up or down relative to status quo). Many crazies who you disagree with about the optimal tax level are going to end up on your side of the "should taxes go up or down' debate. Your opponents and engagement-driven social media have strong incentives to emphasize the crazies on your side, and you have a strong incentive to downplay their extremity by sane-washing them.
https://slatestarcodex.com/2014/03/24/should-you-reverse-any-advice-you-hear/
Qua Scott, there are many pathologically selfish and selfless people in the world. Selfish people need to hear the message that other people's preferences matter and you can't always put yourself first. Selfless people need to hear the message that it is sometimes okay to put yourself first. But group A might hear the message/content intended for group B and use it to justify their destructive behaviour, and vice versa.
I feel this way about a lot of modern "self-care" content, which seems like it was (at least initially) intended to give some measure of succour to genuine victims of child abuse, bullying or similar, but was quickly co-opted by selfish narcissists (invariably claiming to suffer from some nebulously defined "trauma") to rationalize their self-absorption and inconsiderate behaviour. Or indeed a lot of journalistic content about the dangers of excessive exercise, which may be useful advice to the handful of legitimate fitness freaks out there, but likely ends up being consumed by the obese and sedentary as a reason not to go for a walk.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link