site banner

Culture War Roundup for the week of December 11, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

6
Jump in the discussion.

No email address required.

Three months ago, LessWrong admin Ben Pace wrote a long thread on the EA forums: Sharing Info About Nonlinear, in which he shared the stories of two former employees in an EA startup who had bad experiences and left determined to warn others about the company. The startup is an "AI x-risk incubator," which in practice seems to look like a few people traveling around exotic locations, connecting with other effective altruists, and brainstorming new ways to save the world from AI. Very EA. The post contains wide-ranging allegations of misconduct mostly centering around their treatment of two employees they hired who started traveling with them, ultimately concluding that "if Nonlinear does more hiring in the EA ecosystem it is more-likely-than-not to chew up and spit out other bright-eyed young EAs who want to do good in the world."

He, and it seems to some extent fellow admin Oliver Habryka, mentioned they spent hundreds of hours interviewing dozens of people over the course of six months to pull the article together, ultimately paying the two main sources $5000 each for their trouble. It made huge waves in the EA community, torching Nonlinear's reputation.

A few days ago, Nonlinear responded with a wide-ranging tome of a post, 15000 words in the main post with a 134-page appendix. I had never heard of either Lightcone (the organization behind the callout post) or Nonlinear before a few days ago, since I don't pay incredibly close attention to the EA sphere, but the response bubbled up into my sphere of awareness.

The response provides concrete evidence in the form of contemporary screenshots against some of the most damning-sounding claims in the original article:

  • accusations that when one employee, "Alice", was sick with COVID in a foreign country and nobody would get her vegan food so she barely ate for two days turned into "There was vegan food in the house and they picked food up for her, but on one of the days they wanted to go to a Mexican place instead of getting a vegan burger from Burger King."

  • accusations that they promised another, "Chloe", compensation around $75,000 and stiffed her on it in various ways turned into "She had a written contract to be paid $1000/monthly with all expenses covered, which we estimated would add up to around $70,000."

  • accusations that they asked Alice to "bring a variety of illegal drugs across the border" turned into "They asked Alice, who regularly traveled with LSD and marijuana of her own accord, to pick up ADHD medicine and antibiotics at a pharmacy. When she told them the meds still required a prescription in Mexico, they said not to worry about it."

The narrative the Nonlinear team presents is of one employee with mental health issues and a long history of making accusations against the people around her came on board, lost trust in them due to a series of broadly imagined slights, and ultimately left and spread provable lies against them, while another who was hired to be an assistant was never quite satisfied with being an assistant and left frustrated as a result.

As amusing a collective picture as these events paint about what daily life at the startup actually looked like, they also made it pretty clear that the original article had multiple demonstrable falsehoods in it, in and around unrebutted claims. More, they emphasized that they'd been given only a few days to respond to claims before publication, and when they asked for a week to compile hard evidence against falsehoods, the writers told them it would come out on schedule no matter what. Spencer Greenberg, the day before publication, warned them of a number of misrepresentations in the article and sent them screenshots correcting the vegan portion; they corrected some misrepresentations but by the time he sent the screenshots said it was too late to change anything.

That's the part that caught my interest: how did the rationalist community, with its obsession with establishing better epistemics than those around it, wind up writing, embracing, and spreading a callout article with shoddy fact-checking?

From a long conversation with Habryka, my impression is that a lot of EA community members were left scarred and paranoid after the FTX implosion, correcting towards "We must identify and share any early warning signs possible to prevent another FTX." More directly, he told me that he wasn't too concerned with whether they shared falsehoods originally so long as they were airing out the claims of their sources and making their level of epistemic confidence clear. In particular, the organization threatened a libel suit shortly before publication, which they took as a threat of retaliation that meant they should and must hold to their original release schedule.

My own impression is that this is a case of rationalist first-principles thinking gone awry and applied to a domain where it can do real damage. Journalism doesn't have the greatest reputation these days and for good reason, but his approach contrasts starkly with its aspiration to heavily prioritize accuracy and verify information before releasing it. I mention this not to claim that they do so successfully, but because his approach is a conscious deviation from that, an assertion that if something is important enough it's worth airing allegations without closely examining contrary information other sources are asking you to pause and examine.

I'd like to write more about the situation at some point, because I have a lot to say about it even beyond the flood of comments I left on the LessWrong and EA mirrors of the article and think it presses at some important tension points. It's a bit discouraging to watch communities who try so hard to be good from first principles speedrun so many of the pitfalls broader society built guardrails around.

Concrete note on this:

accusations that they promised another, "Chloe", compensation around $75,000 and stiffed her on it in various ways turned into "She had a written contract to be paid $1000/monthly with all expenses covered, which we estimated would add up to around $70,000."

The "all expenses" they're talking about are work-related travel expenses. I, too, would be extremely mad if an employer promised me $75k / year in compensation, $10k of which would be cash-based, and then tried to say that costs incurred by me doing my job were considered to be my "compensation".

Honestly most of what I take away from this is that nobody involved seems to have much of an idea of how things are done in professional settings, and also there seems to be an attitude of "the precautions that normal businesses take are costly and unnecessary since we are all smart people who want to help the world". Which, if that's the way they want to swing, then fine, but I think it is worth setting those expectations upfront. And also I'd strongly recommend that anyone fresh out of college who has never had a normal job should avoid working for an EA organization like nonlinear until they've seen how things work in purely transactional jobs.

Also it seems to me based on how much interest there was in that infighting that effective altruists are starved for drama.

They never promised $75k/year in compensation, $10k of which would be cash-based. This was the compensation package listed in their written, mutually agreed upon employment contract:

As compensation for the services provided, the Employee shall be paid $1,000 per month as well as accommodation, food, and travel expenses, subject to Employer's discretion.

They included another text in evidence where they restated part of it:

stipend and salary mean the same thing. in this instance, it's just $1000 a month in addition to covering travel, food, housing etc

The only apparent mention of $70000 as a number happened during a recorded interview (edited for clarity, meaning retained):

We're trying to think about what makes sense for compensation, because you're gonna be living with us, you're gonna be eating with us. How do you take into account the room and the board and stuff and the travel that's already covered? What we're thinking is a package where it's about the equivalent of being paid $70k a year in terms of the housing and the food, and you'll eat out every day and travel and do random fun stuff. And then on top of that, for the stuff that's not covered by room and board and travel is $1000 a month for basically anything else.

I would not personally take a job offering this compensation structure, but they were fully upfront about what the comp package was and it came pre-agreed as part of the deal. I see no grounds for complaints about dishonesty around it.

I read the same doc you did, and like. I get that "Chloe" did in fact sign that contract, and that the written contract is what matters in the end. My point is not that Nonlinear did something illegal, but... did we both read the same transcript? Because that transcript reads to me like "come on, you should totally draw art for my product, I can only pay 20% of market rates but I can get you lots of exposure, and you can come to my house parties and meet all the cool people, this will be great for your career".

I don't know how much of it is that Kat's writing style pattern matches really strongly to a particular shitty and manipulative boss I very briefly worked for right after college. E.g. stuff like

As best as I can tell, she got into this cognitive loop of thinking we didn’t value her. Her mind kept looking for evidence that we thought she was “low value”, which you can always find if you’re looking for it. Her depressed mind did classic filtering of all positive information and focused on all of the negative things. She ignored all of my gratitude for her work. In fact, she interpreted it as me only appreciating her for her assistant work, confirming that I thought she was a “low value assistant”. (I did also thank her all the time for her ops work too, by the way. I’m just an extremely appreciative boss/person.)

just does not fill me with warm fuzzy feelings about someone's ability to entertain the hypothesis that their own behavior could possibly be a problem. Again, I am probably not terribly impartial here - I have no horse in this particular race, but I once had one in a similar race.

While I don't endorse "come on, you should totally draw art for my product"–type behavior, I do think the position would have been appealing and appropriate for a certain type of person I am not far from. My monthly salary on top of room and board was significantly larger as a military enlistee, but I also wasn't traveling the world. I think they were realistically underpaying for what they wanted but also think "don't take the job" is an adequate remedy to that.

I take your point about the writing style, but for me it's secondary to the core impression that the investigation was very badly mishandled in a way that makes examining things now feel unfair. The initial report should not have been released as-is and it reflects poorly on the whole EA/LW-rationalist community that it was. Given the poor choices around its release, I don't feel inclined to focus too much on what really looks like mundane and predictable workplace/roommate drama.

I agree that it was badly mishandled. I think it's valuable to tell EAs that the "people will try to get you to take a job where they say you'll be paid in experience/exposure, be mindful of that dynamic" but singling out a single organization to that degree makes it sound like it's a problem specific to that organization (which it is not, even within the EA space I personally know of another org with similar dynamics, and I'm not even very involved with the space).

I personally still wouldn't work for nonlinear but then I also would have noped out in the initial hash-out-the-contract phase.

Yeah. And honestly, there are worse things than being paid in exposure. I'd describe that as the primary compensation for my podcast job (my bosses pay me a perfectly fair hourly wage, but I'm certainly not doing it for the money). It's just worth being clear-eyed about precisely what that entails and when it's appropriate.