site banner

Culture War Roundup for the week of January 30, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

13
Jump in the discussion.

No email address required.

The way I think about this is by considering that a strong enough AI program would be theoretically capable of using a clothed picture of someone to actually reconstruct the way the exact way they look naked, which would be quite literally equivalent to the aforementioned situation/looking at them with x-ray glasses, etc. which again (I think) we and most people agree would be wrong. And so, less-powerful AI capable of doing something similar seem to be at least on that gradient of wrong, if not exactly as bad.

Barring some development in physics that enable something like time travel, I don't see how an AI program like that could even theoretically be possible. The information just isn't there, and it's trivial to come up with extremely plausible examples where no AI could reconstruct the exact way someone looks naked - e.g. if someone has a tattoo on their stomach or something.

As such, I don't see this as a gradient. When we're talking about AI or deepfakes, we're necessarily talking about using computers to create brand new illustrations that didn't exist before. In the future, AI could (and likely will) get so much better that we could use clothed pictures of someone to easily and quickly create illustrations of that person that look, to a very high degree of accuracy, as if it were just a nude photo of that person. However, given that that accuracy can't get to 100% or even all that close to it, it's categorically different from, say, snapping a nude photograph of that person.

To me, the strongest argument against deepfakes is the potential distress caused to someone who might accidentally (or intentionally, for that matter) run into such things of themselves. I place this in the same category of suffering as the suffering of some Fundamentalist Christian or Muslim at observing gay people kissing or a cartoon of Mohammed; given how much damage some gay couple or a bunch of pixels can do to these individuals (i.e. none) it's suffering caused entirely by the individual's decision to deem those things as causing themselves suffering. A bunch of pixels that are arranged to look like an accurate photograph of someone nude simply lacks the ability to do anything to affect that person beyond what that person chooses. At best, perhaps a bunch of people jacking it to pictures of someone would lead to those people treating that person differently, but a. I don't particularly see why this would be true and b. even if it were, the issue would be the people choosing to treat that person differently, not in them having access to such pictures.

To me, the strongest argument against deepfakes is the potential distress caused to someone who might accidentally (or intentionally, for that matter) run into such things of themselves. I place this in the same category of suffering as the suffering of some Fundamentalist Christian or Muslim at observing gay people kissing or a cartoon of Mohammed; given how much damage some gay couple or a bunch of pixels can do to these individuals (i.e. none) it's suffering caused entirely by the individual's decision to deem those things as causing themselves suffering.

I've been thinking a lot of this question given certain internet drama I'm plugged into and this analogy was really helpful to me to firm up my own internal position on this so thanks.