site banner

Culture War Roundup for the week of January 1, 2024

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

6
Jump in the discussion.

No email address required.

I think classical utilitarianism as a whole suffers from a lack of respect for duty to the near in ways that this sort of misconduct highlights.

I take it a bit further and argue that utilitarianism in general, and the more classical and universalist varieties of utilitarianism in particular are incompatible with "duty" as a concept.

One of the core reasons that I continue to maintain that "AI Alignment Problem" is misnamed and has far more to with the inherent flaws of utilitarianism as espoused by men like Singer, Benatar, Yudkowsky, Et Al than it does intelligence artificial or otherwise is that a utilitarian cannot credibly commit to cooperate in a prisoners' dilemma or be held to an existing agreement. Because, contra Scott Alexander, one of the core tenets of utilitarianism is that one has a moral duty to renege on said agreements should it become convenient to do so if doing so can be plausibly framed as "increasing net utility."

ETA: As I have argued previously, the difference between virtue signaling and real virtue is that virtue entails sacrifices. It requires suffering and foregoing things you want, things like defrauding your investors (looking at you SBF) or in this case banging your students/coworkers and cheating on your wife.

Yudkowsky

cannot credibly commit to cooperate in a prisoners' dilemma or be held to an existing agreement.

Come on man. Yudkowsky has made more progress on this general problem than any other living philosopher.

First, I'm not convinced that Yudkowsky Et Al actually demonstrated what they seem to think they've demonstrated (convient assumptions and seemingly arbitrary weights abound). Second, it's debatable whether adding the equivalent of epicycles to geo-centerism constitutes "progress".

If sufficiently advanced utilitarianism approaches virtue ethics in its' outputs, it raises the question; "why bother with utilitarianism in the first place?"

You have to distinguish between the ridiculously-abstract question of what should lie at the base of an ethical system, and the ridiculously-empirical question of what kinds of ethical injunctions people will successfully understand and consistently obey.

There’s no reason that we should have to choose the same answer to both of these questions. Arguing that we should feels to me like arguing that we shouldn’t, say, take the axiom of choice in our system of mathematical logic because the average dude on the street is likely to mis-apply it.

I will further clarify that in both cases, the abstract question of “what should we take at the base of our system?”, divorced from its actual consequences, seems wildly pointless to me. Only what happens when you adopt a rule should matter. Which is, despite everything, what you’re saying above, right?

Because you can't program virtue ethics into an AI. You need a utility function.

All of Yudkowsky's philosophical work is grounded on the framework of AI development.

Because you can't program virtue ethics into an AI. You need a utility function.

I am not even sure Yudkowsky would argue this. In any case this is not defensible unless you think that virtue ethics is in principle not computable.

You need a utility function.

Debatable, there are alternative architectures available and Yudkowsky is pretty open about the fact that he sees AI development as a means not an end. I think it would be more accurate to say that it's the other way. You go back and read his early writings on LessWrong it's all about using advanced computational methods to finally "solve" the question of morality once and for all and immanentize the eschaton be it through Fully-Automated Luxury Gay Space Communism or by paving the universe with computronium.

Credit where credit is due though, while he may have started out in the "hyper intelligence for hyper intelligence's sake" camp I think at some point he realized that a true hyper intelligence's that shared his philosophy and preferences would view him with the same indifference/lack-of-moral-regard with which he views "the less intelligent" (IE normies) and that this triggered something of an existential crisis which is the source of his current focus.