site banner

Culture War Roundup for the week of September 19, 2022

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

33
Jump in the discussion.

No email address required.

I wrote a post about de-biasing efforts in machine learning, which got a bit long, so I decided to turn it into an article instead. It's about how corporate anti-bias solutions are mostly only designed to cover their asses, and does nothing to solve the larger (actually important) issue.

(As an aside: does it still count as a "bare link" if I point to my own content, just hosted elsewhere?)

I pointed this out years ago: there are basically two natural philosophies when it comes to avoiding bias: either it's sufficient to (e.g.) mask race as an explicit input, or it's necessary to ensure the predictions treat all races identically. If you're against both forms but keep talking about how a system isn't "fair", you need to take a long look at what, exactly, you mean by "fairness".

Also this goes well beyond machine learning -- policing, insurance, etc. A lot of political issues aren't even at the point where productive discussion is possible, because nobody is willing to state unequivocally what they mean by "fair".

mask race as an explicit input

"Unfortunately" for the machine learning case, they lack the complex internal self-censorship that humans can do to be able to pull that off. Even if you mask out inconvenient inputs like race and gender the model will likely immediately notice clusters of correlated traits that stem from that, and reconstruct the race and gender from scratch.

(A fun idea for a dystopian story element: people conspicuously purchasing items and visits places associated with a "safe" demographic like elderly Asian women, to keep the eye-of-sauron AI off their backs.)

I'm not sure I really believe "self censoring" is a coherent concept. For example, if you're predicting the probability somebody will default on a loan and white people have lower credit scores than asians people, is any model that notices assigns asians a lower probability (on average) not "self censoring"? If 99% of people who eat sushi default on their loans, is it "unfair" to penalize them?

To be frank, I've never heard a definition of "ignores race" that isn't implicitly asking for just nakedly pretending all races have equal odds of defaulting on loans, regardless of whether that's accurate or not. I think people are actually asking for this "post-hoc fairness" should be explicit about that. Instead when I say "so you're saying we should just retroactively make the outputs of the result less accurate and pretend men and women are equally likely to commit murder", the response I get is "you're just straw-manning me, I just want a model that's fair".

It's frustrating that the complaints about bias seem contradictory and/or unsatisfiable, and the people making them are unwilling or unable to elaborate. If somebody is going to criticize my model, they should give me a well-defined notion of fairness that it's actually possible for my model to meet. If their definition of fairness means "deliberately cripple your model and force banks to give out loans that are unprofitable" they need to actually own that instead of hiding behind ambiguity.

Edit: The "post hoc" solution (which I think is the only solution that meaningfully satisfies progressive demands) is:

  1. train a model that uses every variable

  2. train a model that only uses variables you want to ignore (race, sex, etc.)

  3. your predictions are model1(x) - model2(x)

Apparently actual progressives disagree, but I've never heard anyone give an alternative.