site banner

Culture War Roundup for the week of November 20, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

7
Jump in the discussion.

No email address required.

NYT: Before Altman’s Ouster, OpenAI’s Board Was Divided and Feuding

The NYT scooped everybody. We finally know why Sam Altman was fired:

A few weeks before Mr. Altman’s ouster, he met with [OpenAI board member Helen Toner] to discuss a paper she had recently co-written for Georgetown University’s Center for Security and Emerging Technology.

Mr. Altman complained that the research paper seemed to criticize OpenAI’s efforts to keep its A.I. technologies safe while praising the approach taken by Anthropic, according to an email that Mr. Altman wrote to colleagues and that was viewed by The New York Times.

In the email, Mr. Altman said that he had reprimanded Ms. Toner for the paper and that it was dangerous to the company, particularly at a time, he added, when the Federal Trade Commission was investigating OpenAI over the data used to build its technology.

Ms. Toner defended it as an academic paper that analyzed the challenges that the public faces when trying to understand the intentions of the countries and companies developing A.I. But Mr. Altman disagreed.

“I did not feel we’re on the same page on the damage of all this,” he wrote in the email. “Any amount of criticism from a board member carries a lot of weight.”

Senior OpenAI leaders, including Mr. Sutskever, who is deeply concerned that A.I. could one day destroy humanity, later discussed whether Ms. Toner should be removed, a person involved in the conversations said.

There are a few other minor issues mentioned in the article, but this sounds like the big one. Rationalist/EA types take being told that they can't criticize "allies" in public very negatively, a position I am quite sympathetic to. Helen Toner works at an Open Philanthropy-funded think tank, so she's as blue blood an effective altruist as they get. My guess is that this was the moment that she decided that Sam had to be eliminated before he took control of the board and jeopardized OpenAI's mission.

What gets me is how disingenuous this makes the original firing announcement: "Mr. Altman’s departure follows a deliberative review process by the board, which concluded that he was not consistently candid in his communications with the board, hindering its ability to exercise its responsibilities." It sounds like he was perfectly candid. They just didn't like what he was about.

In completely unrelated news, ChatGPT has been down for the last three hours.

Senior OpenAI leaders, including Mr. Sutskever, who is deeply concerned that A.I. could one day destroy humanity

These people are so weird. They are obsessed with making something they think will destroy the world. Just build AI or don't. If AI is so dangerous, nothing you can do will really stop it in the long run. Eventually, over hundreds or even thousands of years, the inevitable will happen. I feel like this is almost a religion for them and they are the absolute worst people to be in charge or this kind of technology.

  1. They think AGI is inevitable, and if it's not made by people with safety and alignment in mind, the odds of it being misaligned skyrocket.

  2. Control of an aligned Superintelligent AGI is equivalent to having the keys to the lightcone, if you make it through the gauntlet of it not killing you and it listens to what you tell it, then you have the means to dominate everyone else, including others who make misaligned AGI, if yours is capable of squashing them at birth, or at the very least capable of panopticon surveillance to prevent anyone from building one in the first place.

  3. Even prior to that, being on the cutting edge of AI research gives you a voice, people can dismiss Eliezer Yudkowsky as a crank, far harder to do that to Geoffrey Hinton or Ilya Sutskever. You have far more power to get governments to regulate things, or develop industry best standards that reduce the risk of more laissez-faire competitors YOLOing the whole thing.

AI is dangerous, as is anything much smarter than you that potentially doesn't share your goals, but if you see it as inevitable, then your best bet is making sure it comes out with goals you share or control.

My own p(doom) from AI has dropped to about 30% from a high of 70%, when RLHF and other techniques showed that it was possible to ~mostly align the best AI today, in the form of LLMs, which are the frontrunners for the best AIs of tomorrow. Just because I currently do not think AI will probably kill us in a decade doesn't mean I don't think it's a serious risk, and there are few things on the planet more worth being passionate about.

It is not remotely as simple as build AI or don't, even if everyone associated with the AI X-risk community died overnight, the genie is out of the bottle, and others will pursue it with billions of dollars. OAI had a value, before this debacle, of about $80 billion, with operating expenses on the order of 500 million p/a.

If you can't undo that, or enforce a "pause", then you grudgingly find a way to stay ahead of the competition while doing your best not to be the reason why it all went to hell.

This is an excellent answer. One small quibble:

Control of an aligned Superintelligent AGI is equivalent to having the keys to the lightcone, if you make it through the gauntlet of it not killing you and it listens to what you tell it, then you have the means to dominate everyone else, including others who make misaligned AGI, if yours is capable of squashing them at birth, or at the very least capable of panopticon surveillance to prevent anyone from building one in the first place.

For the record I think Yudkowsky and friends are wrong about this one. Control of the only superintelligent AGI, if that AGI is a single coherent entity, might be the keys to the lightcone, but so far it looks to me like AGI scales horizontally much better than it scales vertically.

This, if anything, makes things more dangerous rather than less, because it means there is no permanent win condition, only the deferral of the failure condition for a bit longer.

Thanks!

but so far it looks to me like AGI scales horizontally much better than it scales vertically.

This particular concern hinges on recursive self-improvement, and I agree that we haven't seen much evidence of that, yet, but it's still the early days.

I think that the intelligence of LLMs needs to at least reach that of the average ML researcher capable of producing novel research and breakthroughs before we can call it one way or another, and we're not there yet, at least in terms of released models, not that I expect Gemini or GPT-5 to be that smart yet. The closest I can think of is training LLMs on synthetic data curated by other models, or something like Nvidia using ML models to optimize their hardware, but that's still weaksauce.

If it turns out to be feasible, it still remains to be seen whether we have a hard take-off with a Singleton or a slow (yet fast on human timescales, just months or years) takeoff which might allow for multipolarity. I remain agnostic yet gravely concerned myself.

This particular concern hinges on recursive self-improvement

And most of the talk on that issue assumes that the point where said self-improvement hits steep diminishing returns must necessarily be somewhere far above human intelligence — again, apparently based on nothing beyond it being more conducive to one's preferred outcomes than the alternative.

Diminishing returns != no or negative returns. Intelligence is the closest thing we have to an unalloyed good, and the difference in capabilities between people with just 20 or 30 IQ points is staggering enough.

Nothing at all suggests that the range of IQ/intelligence seen in unmodified humans constrained by a 1.4 kg brain in a small cranium applies at all to an entity that spans data-centers, especially those that can self-modify and fork themselves on demand. You don't need a bazillion IQ points to be immensely dangerous, human scientists with maybe 160 or 170 invented nukes.

We have AI that already matches human intelligence on many or even most cognitive tasks, the scaling laws still hold, and companies and nations can easily afford to throw several OOMs more money at the problem.

Humanity itself has seen exponential or even super-exponential advancement, and we've barely gained a handful of IQ points from the Flynn effect, most of it was merely technological compounding.

Since the theoretical or practical upper limits on the size and speed of an AGI are massive, I wish to see what reason anyone has to claim they'll bottom out within spitting distance of the smartest humans. That is ludicrous prima facie, even if we don't know how fast further progression will be.

You don't need a bazillion IQ points to be immensely dangerous, human scientists with maybe 160 or 170 invented nukes.

Yes, but you're assuming there's a lot more even more dangerous things "out there" for a smarter entity to discover.

What is intelligence for? That is, what is its use? Primarily,

  1. modeling reality
  2. modeling other agents.

Our first day of Physics lab classes at Caltech, the instructor told us that it doesn't matter how many digits of pi we'd all memorized (quite a bunch), just use 3.14, or a scientific calculator's pi key, whichever was faster, because any rounding error would be swamped out by the measurement error in our instruments.

When it comes to modeling the physical world, sure, going from knowing, say, Planck's constant to two decimal places to knowing it to three decimal places will probably net you a bunch of improvements. But then going from, say, ten decimal places to eleven, or even ten decimal places to twenty, almost certainly won't net the same level of improvement.

When modeling other minds, particularly modeling other minds modeling you modeling… — the whole "I know that you know that I know…" thing — well, that sort of recursion provides great returns on added depth… in certain games, like chess. But AIUI, in most other situations, that kind of thing quickly converges to one or another game-theoretic equilibrium, and thus the further recursion allowed by greater intelligence provides little additional benefit.

I'm not saying we can't produce an intelligence "that spans data-centers" much smarter than us, and I'm not saying it's impossible that there are dangerous and powerful things such an intelligence might figure out, I'm just saying it can't just be assumed, or treated as highly likely by default. That it's unsupported extrapolation to reason 'smart=nukes, therefore super-smart=super-nukes and mega-smart=mega-nukes.' I'm not saying that machine intelligence will "bottom out within spitting distance of the smartest humans," I'm saying that it's possible that the practical benefits of such intelligence, no matter how much vaster than our own, may "bottom out" well below the dreams of techno-optimists like yourself, and you can't just rule that out a priori on an unsubstantiated faith that there's vast undiscovered realities beyond our limited comprehension just waiting for a smarter being to uncover.

I want you to at least consider, just for a moment, the idea that maybe we humans, with our "1.4 kg brain[s] in a small cranium," may have a good enough understanding of reality, and of each other, that a being with "900 more IQ points" won't find much room to improve on it.

I'm not saying "a machine can never be smarter than a man!" I'm saying "what if a machine a thousand times smarter than us says, 'yeah, you already had it mostly figured out, the rest is piddly details, no big revelations here'?"

Yes, but you're assuming there's a lot more even more dangerous things "out there" for a smarter entity to discover.

I repeat that, while I think this is true, it's still not necessary for a genius AI to be an existential risk. I've already explained why multiple times.

Nukes? They exist.

Pandemics? They exist. Can they be made more dangerous? Yes. Are humans already making them more dangerous for no good reason? Yes.

Automation? Well underway.

Our first day of Physics lab classes at Caltech, the instructor told us that it doesn't matter how many digits of pi we'd all memorized (quite a bunch), just use 3.14, or a scientific calculator's pi key, whichever was faster, because any rounding error would be swamped out by the measurement error in our instruments.

When it comes to modeling the physical world, sure, going from knowing, say, Planck's constant to two decimal places to knowing it to three decimal places will probably net you a bunch of improvements. But then going from, say, ten decimal places to eleven, or even ten decimal places to twenty, almost certainly won't net the same level of improvement.

I do not think that the benefits of additional intelligence as seen even in human physicists is well addressed by this analogy. The relevant one would be comparing Newtonian physics to GR, and then QM. In the domains where such nuance becomes relevant, the benefits are grossly superior.

For starters, while the Standard Model is great, it still isn't capable of conclusively explaining most of the mass or energy in the universe. Not to mention that even if we have equations for the fundamental processes, there are bazillions of higher-order concerns that are intractable to simulate from first-principles.

AlphaFold didn't massively outpace SOTA on protein folding by using QM on a molecule by molecule basis. It found smarter heuristics, that's also something intelligence is indispensable for. I see no reason why a human can't be perfectly modeled using QM, it is simply a computationally intractable problem even for a single cell within us.

In other words, knowing the underling rules of a complex system != knowing all the potential implications or applications. You can't just memorize the rules of Chess and then declare it's a solved problem.

That it's unsupported extrapolation to reason 'smart=nukes, therefore super-smart=super-nukes and mega-smart=mega-nukes.'

I'm sure there people who might make such a claim. I'm not one of them, and like I said, it's not load bearing. Just nukes is sufficient really. Certainly in combination with automation so the absence of those pesky humans running the machines isn't a problem.

I want you to at least consider, just for a moment, the idea that maybe we humans, with our "1.4 kg brain[s] in a small cranium," may have a good enough understanding of reality, and of each other, that a being with "900 more IQ points" won't find much room to improve on it.

I have considered it, at least to my satisfaction, and I consider it to be exceedingly unlikely. Increases in intelligence, even within the minuscule absolute variation seen within humans, is enormously powerful. There seems to be little to nothing in the way of further scaling in the context of inhuman entities that are not constrained by the same biological limitations in size, volume, speed or energy. They already match or exceed the average human in most cognitive tasks, and even if returns from further increases in intelligence diminish grossly or become asymptotic, I am the furthest from convinced that stage will be reached within spitting distance of the best of humanity, or that such an entity won't be enormously powerful and capable of exterminating us if it wishes to do so.

More comments