site banner

Culture War Roundup for the week of May 22, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

10
Jump in the discussion.

No email address required.

Every worry is unlike every other. In terms of qualitative danger, AI is more like other worries than commonly believed, because its unique property of reactiveness, which it shares with us, loads on both ends of the scale. Sometimes it's about optimizing for the exact same criterion, just with outputs going to different people. Cases in point:

and so on. Importantly, this isn't the case where the defense has to crush every single attack to be successful whereas attackers need to only triumph once, like doomers often say. Successful attacks will not be existentially threatening (unless the attacker does have a tremendous advantage in technical capability, but that's trivial and a good reason to commoditize the technology, if anything). Attacks still have cost, their perpetrators still work with limited resources, leave a footprint and are vulnerable to discovery, and while it is not a given that attackers learn effectively from each other, the next iteration of defense is better-informed; until attacks run into fundamental constraints.

All that information will be banned, just as facial recognition software has been banned for police

US cities are already reversing facial recognition bans, New Orleans did just a few months ago iirc. If it works, it will happen.

I know this is a hobby horse, but once AI is trained on gait recognition and body language of labelled examples of millions of hours of countless criminals’ movements recorded by CCTV, tiny little telltale patterns might well allow for effective pre-crime in the case of almost all premeditated criminal activity. People show their nerves, everyone has a tell, etc.

I have trouble believing there's enough information content present in CCTV streams to uniquely identify individuals confidently. I see how it maybe could work, but it's not something I'd focus on directly. Are human gaits really that different as to be identifiable from distant security cameras? Are they even consistent for a single person day-to-day?

The longer I think about it, I've also started thinking that AI likely scales sub-linearly (logarithmic?) with the size of the training dataset. "But the AI can viably consider a larger dataset than human experts" may be true, but may not generate hugely better results.

Interesting! Any good papers or summary articles you'd recommend?

Sure, there's a decades-long history of forensic gait analysis (long predating AI of course) in criminology. A nice overview is here. It's actively employed in China integrated with AI, although not widely in the West. In the West, gait analysis by experts has been a feature of trials for a long time - even before CCTV, it was used (and still is) on footprints left at crime scenes to identify suspects.

The challenge, of course, is that for now the applications of current forensic gait analysis are highly limited. The lack of comprehensive gait libraries for the wider population means that, unlike DNA (at least in recent memory) it's generally only used to support attempts to prove a suspect on trial was or was not someone in video footage. The real benefit is in scanning a library of millions or billions of hours of video taken from a network of surveillance cameras (which have ideally already been used to build up a 'library' of the entire population) to find possible 'matches' (the search space can be narrowed by geography and other quantitative or qualitative information recorded by police) in the general public, just like police DNA databases and Ancestry.com data are today.

In the West, research has been slow for a while. It's generally focused on identifying diseases like Parkinsons, the racing industry uses it for analysing horses etc, so a lot of Western research uses Lidar and multiple cameras; these achieve extremely high accuracy (often over 90%), but obviously aren't hugely helpful when the footage is actually blurry black-and-white CCTV at night.