site banner

Culture War Roundup for the week of February 2, 2026

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

Jump in the discussion.

No email address required.

Disclaimer: I am not a programmer, though I keep myself broadly aware of trends. I've only used LLMs for coding for toy problems or personal tooling (AutoHotkey macros, Rimworld mods, a mortar calculator, automating discharge paperwork at my dad's hospital)*. I've noted that they've been excellent at explaining production code to even a dilettante like me, which is by itself immensely useful. And for everything else, I'm so used to the utility they provide me personally that I can't imagine going back.

They being said, I am not in a position to judge the economic utility a professional programmer derives from using it for their day job, though it's abundantly clear that the demand for tokens is enormous, and that the capability of SOTA LLMs is a moving target, getting better every day on both benchmarks and actual projects. And look, I understand there's a position where you say "sure, but these things still aren't actually good" - but if you're claiming they haven't gotten better, then I'm going to gently suggest you might want to check yourself for early-onset dementia. The jump from GPT-3 barely coding a working React toy-example to current models is the kind of improvement curve that should at minimum make you sit up and notice.

In other words, even if you think they're not good enough today, you should at the very least notice that a large and ever-increasing fraction of US GDP is being invested in making them better, with consistent improvements.

However, here's a tweet from Andrej Karpathy which I will reproduce in full:

@karpathy A few random notes from claude coding quite a bit last few weeks.

Coding workflow. Given the latest lift in LLM coding capability, like many others I rapidly went from about 80% manual+autocomplete coding and 20% agents in November to 80% agent coding and 20% edits+touchups in December. i.e. I really am mostly programming in English now, a bit sheepishly telling the LLM what code to write... in words. It hurts the ego a bit but the power to operate over software in large "code actions" is just too net useful, especially once you adapt to it, configure it, learn to use it, and wrap your head around what it can and cannot do. This is easily the biggest change to my basic coding workflow in ~2 decades of programming and it happened over the course of a few weeks. I'd expect something similar to be happening to well into double digit percent of engineers out there, while the awareness of it in the general population feels well into low single digit percent.

IDEs/agent swarms/fallability. Both the "no need for IDE anymore" hype and the "agent swarm" hype is imo too much for right now. The models definitely still make mistakes and if you have any code you actually care about I would watch them like a hawk, in a nice large IDE on the side. The mistakes have changed a lot - they are not simple syntax errors anymore, they are subtle conceptual errors that a slightly sloppy, hasty junior dev might do. The most common category is that the models make wrong assumptions on your behalf and just run along with them without checking. They also don't manage their confusion, they don't seek clarifications, they don't surface inconsistencies, they don't present tradeoffs, they don't push back when they should, and they are still a little too sycophantic. Things get better in plan mode, but there is some need for a lightweight inline plan mode. They also really like to overcomplicate code and APIs, they bloat abstractions, they don't clean up dead code after themselves, etc. They will implement an inefficient, bloated, brittle construction over 1000 lines of code and it's up to you to be like "umm couldn't you just do this instead?" and they will be like "of course!" and immediately cut it down to 100 lines. They still sometimes change/remove comments and code they don't like or don't sufficiently understand as side effects, even if it is orthogonal to the task at hand. All of this happens despite a few simple attempts to fix it via instructions in CLAUDE . md. Despite all these issues, it is still a net huge improvement and it's very difficult to imagine going back to manual coding. TLDR everyone has their developing flow, my current is a small few CC sessions on the left in ghostty windows/tabs and an IDE on the right for viewing the code + manual edits.

Tenacity. It's so interesting to watch an agent relentlessly work at something. They never get tired, they never get demoralized, they just keep going and trying things where a person would have given up long ago to fight another day. It's a "feel the AGI" moment to watch it struggle with something for a long time just to come out victorious 30 minutes later. You realize that stamina is a core bottleneck to work and that with LLMs in hand it has been dramatically increased.

Speedups. It's not clear how to measure the "speedup" of LLM assistance. Certainly I feel net way faster at what I was going to do, but the main effect is that I do a lot more than I was going to do because 1) I can code up all kinds of things that just wouldn't have been worth coding before and 2) I can approach code that I couldn't work on before because of knowledge/skill issue. So certainly it's speedup, but it's possibly a lot more an expansion.

Leverage. LLMs are exceptionally good at looping until they meet specific goals and this is where most of the "feel the AGI" magic is to be found. Don't tell it what to do, give it success criteria and watch it go. Get it to write tests first and then pass them. Put it in the loop with a browser MCP. Write the naive algorithm that is very likely correct first, then ask it to optimize it while preserving correctness. Change your approach from imperative to declarative to get the agents looping longer and gain leverage.

Fun. I didn't anticipate that with agents programming feels more fun because a lot of the fill in the blanks drudgery is removed and what remains is the creative part. I also feel less blocked/stuck (which is not fun) and I experience a lot more courage because there's almost always a way to work hand in hand with it to make some positive progress. I have seen the opposite sentiment from other people too; LLM coding will split up engineers based on those who primarily liked coding and those who primarily liked building.

Atrophy. I've already noticed that I am slowly starting to atrophy my ability to write code manually. Generation (writing code) and discrimination (reading code) are different capabilities in the brain. Largely due to all the little mostly syntactic details involved in programming, you can review code just fine even if you struggle to write it.

Slopacolypse. I am bracing for 2026 as the year of the slopacolypse across all of github, substack, arxiv, X/instagram, and generally all digital media. We're also going to see a lot more AI hype productivity theater (is that even possible?), on the side of actual, real improvements.

Questions. A few of the questions on my mind:

  • What happens to the "10X engineer" - the ratio of productivity between the mean and the max engineer? It's quite possible that this grows a lot.
  • Armed with LLMs, do generalists increasingly outperform specialists? LLMs are a lot better at fill in the blanks (the micro) than grand strategy (the macro).
  • What does LLM coding feel like in the future? Is it like playing StarCraft? Playing Factorio? Playing music?
  • How much of society is bottlenecked by digital knowledge work?

TLDR Where does this leave us? LLM agent capabilities (Claude & Codex especially) have crossed some kind of threshold of coherence around December 2025 and caused a phase shift in software engineering and closely related. The intelligence part suddenly feels quite a bit ahead of all the rest of it - integrations (tools, knowledge), the necessity for new organizational workflows, processes, diffusion more generally. 2026 is going to be a high energy year as the industry metabolizes the new capability.

*Sadly they can't make the Rimworld mods I want. This is a combination of a skill-issue on my part (people have successfully made rather large and performant mods with AI), and because I wanted something niche as hell, in the form of compatibility with a very large overhaul mod called Combat Extended. Hey, at least Nano Banana Pro made the art assets with minimal human input, if you think my coding skills are trash, wait till you see my art.

I can approach code that I couldn't work on before because of knowledge/skill issue.

Slopacolypse. I am bracing for 2026 as the year of the slopacolypse across all of github, substack, arxiv, X/instagram, and generally all digital media.

I find it peculiar that Karpathy doesn't see a relationship between those two things. I've noticed a trend where the most glowing reviews of AI capabilities seem to be for people who are using it in areas where they, themselves, do not have enough skill to confidently perform the task. At its worst, it's a sort of tool-assisted Dunning Kruger effect that's actually breathtaking if you can decouple enough to look at it in the abstract.

"I clearly couldn't do this thing but I can clearly tell that Grok/Claude/ChatGPT/Gemini did it right" is a hell of a thing. It's already causing real stress on public software security databases. There's a continuous trend that looks something like this:

"I ran $LIBRARY through Claude and it says there was a potential denial of service attack. I asked Claude for a mitigation and it provided this code."

"Can you explain this piece of the code?"

"I asked Claude to explain this piece of code and it said the following."

Other than feeling like a scene from Office Space, it's effectively acting as a denial of service attack on its own. The amount of time necessary to submit something like that without a deep understanding of the problem is considerably lower than the time necessary for a genuine SME to comb through it and judge it on merits.

Hell, if I were a malicious actor I'd probably try to exploit that by shit-flooding the system to buy myself more time.

I find it peculiar that Karpathy doesn't see a relationship between those two things.

Hmm? That's not my takeaway from the tweet (xeet?). He's not denying a connection between AI capabilities and code quality decline, he's making a more subtle point about skill distribution.

The basic model goes like this: AI tools multiply your output at every skill level. Give a novice programmer access to ChatGPT, Claude or Copilot (maybe not Copilot, lol) , and they go from "can't write working code" to "can produce something that technically runs." Give an expert like Karpathy the same tools, and he goes from "excellent" to "somewhat more excellent." The multiplicative factor might even be similar! But that's the rub, there are way more novices than experts.

So you get a flood of new code. Most of it is mediocre-to-bad, because most people are mediocre-to-bad at programming, and multiplying "bad" by even a generous factor still gives you "not great." The experts are also producing more, and their output is better, but nobody writes news articles about the twentieth high-quality library that quietly does its job. We only notice when things break.

This maps onto basically every domain. Take medicine as a test case (yay, the one domain where I'm a quasi-expert) Any random person can feed their lab results into ChatGPT and get something interpretable back. This is genuinely useful! Going from "incomprehensible numbers" to "your kidneys are probably fine but your cholesterol needs work" is a huge upgrade for the average patient. They might miss nuances or accept hallucinated explanations, but they're still better off than before.

Meanwhile, as someone who actually understands medicine, I can extract even more value. I can write better prompts, catch inconsistencies, verify citations, and integrate the AI's suggestions into a broader clinical picture. The AI makes me more productive, but I was already productive, so the absolute gains are smaller relative to my baseline. And critically, I'm less likely to get fooled by confident-sounding nonsense (it's rare but happens at above negligible rates).

This is where I tentatively endorse a "skill issue" framing, where everyone's output getting multiplied, but bad times a multiplier is still usually bad, and there are simply more bad actors (in the neutral sense) than good ones. The denominator in "slop per good output" has gotten larger, but so has the numerator, and the numerator was already bigger to start with. From inside the system, if you're Andrej Karpathy, you mostly notice that you're faster. From outside, you notice that GitHub is full of garbage and the latest Windows update broke your system.

This isn't even a new pattern. Every productivity tool follows similar dynamics. When word processors became common, suddenly everyone could produce professional-looking documents. Did the average quality of written work improve? Well, the floor certainly rose (less illegible handwriting, if I continue to accurately insult my colleagues), but we also got an explosion of mediocre memos and reports that previously wouldn't have been written at all. The ceiling barely budged because good writers were already good. I get more use out of an LLM for writing advice than, say, Scott.