site banner

Culture War Roundup for the week of November 20, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

7
Jump in the discussion.

No email address required.

To bring up another post from last week, I'm going to go ahead and repost @justcool393's piece on the Sam Altman/OpenAI/Microsoft situation, since she posted it a few hours ago and right before the last thread went down.

Here's her writing:


Another day, another entrant into the OpenAI drama. Emmett Shear is the new interim CEO of OpenAI.

I don't know why it was surprising to people that Sam wouldn't come back. The company was meant to be subservient to the nonprofit's goals and I'm not sure why the attempted coup from Sam's side (you know the whole effectively false reporting that Sam Altman was to become the new CEO) was apparently "shocking" that it failed.

The OpenAI board has hired Emmett Shear as CEO. He is the former CEO of Twitch.

My understanding is that Sam is in shock.

https://twitter.com/emilychangtv/status/1726468006786859101

What's kinda sad about all of this is how much people were yearning for Sam Altman to be the CEO as if he isn't probably one of the worst possible candidates. Like maybe this is just a bunch of technolibertarians on Twitter or HN or something who think that the ultimate goal of humanity is how many numbers on a screen you can earn, but the amazing amount of unearned reverence towards a VC to lead the company.

In any case, here's to hoping that Laundry Buddy won't win out in the rat race for AGI, lest we live in a world optimized for maximum laundry detergent. Maybe we'll avoid that future now with Sam's departure.

Anyway, I'll leave this to munch on which I found from the HN thread.

Motte: e/acc is just techno-optimism, everyone who is against e/acc must be against building a better future and hate technology

Bailey: e/acc is about building a techno-god, we oppose any attempt to safeguard humanity by regulating AI in any form around and around and around"

https://twitter.com/eshear/status/1683208767054438400


I'm reposting here because I'm convinced, like many other residents, that the ongoing drama of who controls AI development has far reaching implications, likely on the scale of major power geopolitical events. If not ultimately even greater.

To add a bit to the discussion to justify reposting - I think many of these discussions around AI Safety versus Accelerationism are extremely murky because so many people in secular, rationalistic circles are extremely averse to claiming religious belief. It's clear to me that both AI Safety and Accelerationism have strong themes of classical religion, and seem to be two different sects of a religion battling it out over the ultimate ideology. Potentially similar to early Orthodox Christians versus Gnostics.

Alternatively, @2rafa has argued that many of the E/Acc (effective accelerationism) crowd comes from bored technocrats who just want to see something exciting happen. I tend to agree with that argument as well, given how devoid of purpose most of the technocratic social world is. Religion and religious-style movements tend to provide that purpose, but when you are explicitly secular I suppose you have to get your motivation elsewhere.

We've also got the neo-luddites like @ArjinFerman who just hate AI entirely and presumably want us to go back to the mid 90s with the fun decentralized internet. Not sure, I haven't actually discussed with him. I can actually agree with some of the Ludditism, but I'd argue we need to go back to 1920 or so and ban all sorts of propaganda, mass media and advertising.

Anyway, clearly the technological battle for the future of our civilization continues to heat up. The luddites seem out, but may have a surprising last hour comeback. The woke/political left leaning folks seem to be strongly in charge, though the OpenAI scandal points to trouble in the Olympian heights of Silicon Valley AI decision makers.

Will the Grey Tribe use AGI to come back and finally recover the face and ground it has lost to the advancing SJW waves? Who knows. I'm just here for the tea.

We've also got the neo-luddites like @ArjinFerman who just hate AI entirely and presumably want us to go back to the mid 90s with the fun decentralized internet. Not sure, I haven't actually discussed with him. I can actually agree with some of the Ludditism, but I'd argue we need to go back to 1920 or so and ban all sorts of propaganda, mass media and advertising.

I didn't really make up my mind how far back to turn the clock to, but I like the way you think.

If RETVRNING is not an option, I do have a general principle in mind on how to proceed, but I don't have a name for it. Techno-optimists often point out that this isn't the first time us Luddites have their gripes about machines making us dumb, and takin' ar jerbs, but here we are, and the world doesn't seem so horrible. Aside from the arguments that, in some ways, yes it is, I think technology should be developed in a way that helps us grow as people, rather than makes us succumb to naked consumerism. As you semi-correctly guessed I already have this issue with what IT promised vs what it delivered. Computers and the Internet disrupted how we do a lot of things, but they could have conceivably given us decentralization and climbing rates of technological literacy. We got the opposite on both fronts. The fact that we ended up with even more centralization is not even that surprising when you think about it, as the forces pushing towards it were on open display all this time, but what happened to tech-literacy came as a bit of shock to me. X-ers and Millenials probably all had the childhood experience of their parents buying a new device, and us being able to figure out how it works through mere trial and error, before our parents could find their way through the manual. For years I assumed the same will happen to me, but it just hasn't, and reportedly there are now kids who don't even know what a file is, because the way we design software is hiding the fundamentals of how computers work. On one hand that's a relief - it doesn't look like a young whippersnapper is about to take my jerb anytime soon - but it's also depressing. This, more then anything else, is what worries me the most about the advent of AI, and if anyone has any ideas how to avoid it, I'm all ears.

There was this old TNG episode about kids getting abducted from the Enterprise to live on a planet where all their needs are catered to by a planetary AI, so they can do art and stuff. Well, what I'm saying is: Both the Federation and Aldea has AI technology, but they choose to use it in different ways. Give me the 8 year olds of the Enterprise, who are forced to master basic calculus so they can grow up - and may Allah forgive me for using this phrase - as well rounded citizens, who actually can maintain the technology they depend on, over the children of Aldea, who for that matter don't even master art, they just have their thoughts and emotions translated into it by the AI.

The final thing that is driving me up the wall, is the utter state of the discourse. EAs, for all the talk of "alignment", never mention either of these issues because, as far as I can tell, they don't want the common people to have an understanding of AI, so they can have total control over it for themselves. As for E/Accs the closest thing I ever got to an acknowledgement of the problems with centralization and dumbing down was "Yeah that worries me too, but what can you do? Anyway, look - ChatGPT go brrr!". For that reason I'm inclined to just disconnect from technological society, and join the Amish.

For that reason I'm inclined to just disconnect from technological society, and join the Amish.

That you don't trust the EA is no reason to disconnect. To beat the EA and not let them have total control you need to support a group that is more aligned to your ideals and try to get your group to have their own influence in the AI game.

It's not just the lack of trust in EAs, E/Accs' approach also seems to lead down a dark path. Basically, I expect the same result as what happened with software, social media, and the Internet generally. At least with software there's the FOSS movement, as helpless as it ultimately turned out to be. Is there a Stallman of AI? Is there even a fraction of energy behind him that there was for FOSS in the 90's and 00's?

I completely agree with you about EA. My point is that you need to play the game with your own side and try to find likeminded people to support. Running away is a losing move.

Of course you personally not wanting to do that is understandable. But when it comes to what is better to do it requires people who try to create alternative platforms and participate in them.

The genie can not be put back in the bottle. Either they monopolize the genies, or others use them too.

Wel, like I said, it's not just the EAs, it's also the E/Accs that I have a problem with.

As for not being able to put the genie back in the bottle, yeah that's one of my fears, but I don't know if this is already decided. By current demographic trends, the Amish are scheduled to inherit America. AI might very well turn out to be a suicidal technology, and Luddites the only survivors.

There will be no survivors. You think not having a cell signal saves you from a Paperclip Maximizer that's demolishing the biosphere for spare parts? Come the fuck on.

I think the Paperclip Maximizer is a boogeyman, and if AI does cause us to go extinct, it will be in a completely different manner than the AI-safety people predict. Like I said above, this is precisely what drives me up the wall in this conversation.

I do not see how there's any remotely plausible world where AI somehow causes us to go extinct while sparing the Amish. Unless the Amish have some really sick data centers hidden under those barns of theirs.

More comments