Be advised: this thread is not for serious in-depth discussion of weighty topics (we have a link for that), this thread is not for anything Culture War related. This thread is for Fun. You got jokes? Share 'em. You got silly questions? Ask 'em.
- 129
- 1
What is this place?
This website is a place for people who want to move past shady thinking and test their ideas in a
court of people who don't all share the same biases. Our goal is to
optimize for light, not heat; this is a group effort, and all commentators are asked to do their part.
The weekly Culture War threads host the most
controversial topics and are the most visible aspect of The Motte. However, many other topics are
appropriate here. We encourage people to post anything related to science, politics, or philosophy;
if in doubt, post!
Check out The Vault for an archive of old quality posts.
You are encouraged to crosspost these elsewhere.
Why are you called The Motte?
A motte is a stone keep on a raised earthwork common in early medieval fortifications. More pertinently,
it's an element in a rhetorical move called a "Motte-and-Bailey",
originally identified by
philosopher Nicholas Shackel. It describes the tendency in discourse for people to move from a controversial
but high value claim to a defensible but less exciting one upon any resistance to the former. He likens
this to the medieval fortification, where a desirable land (the bailey) is abandoned when in danger for
the more easily defended motte. In Shackel's words, "The Motte represents the defensible but undesired
propositions to which one retreats when hard pressed."
On The Motte, always attempt to remain inside your defensible territory, even if you are not being pressed.
New post guidelines
If you're posting something that isn't related to the culture war, we encourage you to post a thread for it.
A submission statement is highly appreciated, but isn't necessary for text posts or links to largely-text posts
such as blogs or news articles; if we're unsure of the value of your post, we might remove it until you add a
submission statement. A submission statement is required for non-text sources (videos, podcasts, images).
Culture war posts go in the culture war thread; all links must either include a submission statement or
significant commentary. Bare links without those will be removed.
If in doubt, please post it!
Rules
- Courtesy
- Content
- Engagement
- When disagreeing with someone, state your objections explicitly.
- Proactively provide evidence in proportion to how partisan and inflammatory your claim might be.
- Accept temporary bans as a time-out, and don't attempt to rejoin the conversation until it's lifted.
- Don't attempt to build consensus or enforce ideological conformity.
- Write like everyone is reading and you want them to be included in the discussion.
- The Wildcard Rule
- The Metarule

Jump in the discussion.
No email address required.
Notes -
I've... picked up a Claude Max 20x plan. No, I can't disclose how I acquired it, though I didn't have to pay a cent (and it's all legit). It's so fucking good, but at the same time, the more I use Opus 4.6, the more I'm impressed by how close Sonnet 4.6 gets. Sure, Opus is legitimately better, but the difference is nowhere near as stark as say, Gemini Flash vs Pro, or GPT's Thinking or Instant mode. Anthropic cooked, and I can't wait to try Mythos when the version for plebs comes out.
PS: If anyone has a good guide to Claude Code or agentic setups, I need one. I have some serious experimentation to do while I have it.
Alright, I have one for you: If you read the most recent installment of my Pittsburgh series you'll see that I have a sentence about how the Italian composition of Bloomfield changed over time, based on information compiled by the US Census. While I am skeptical of AI overall, I admit that it has legitimate potential uses, one of which is aggregating large amounts of statistical information from diverse sources that are a pain in the ass to search manually. To give you some background, the census started asking about ancestry beginning in 1980, on the "long form" that was given to 1/6 of the population. Following the 2000 census, the long form was eliminated, and the data was continually polled using the American Community Survey. When it was only being collected decennially, it was published in reports that are available online in PDF format. Following the introduction of the ACS, the census bureau implemented an API. In order to streamline the process, I gave Claude the following instruction:
After running for tens of minutes and spitting out a bunch of technical data about the API, it gave me this message:
It then informed me that my single, unsuccessful query used up my limit for the day. I would add that the pre-ACS data is available in PDF form from third party websites. I was able to compile it manually without much issue, though this would have been shorter. If Opus is able to do this, then I would like to see if I can get it to extend the data to prior years based on national origin. I don't know if this was compiled but the individual forms are available from 1950 and earlier, and they list the country of origin for each person. Around that time, most people with Italian ancestry would have been first or second generation, so the number of people born in Italy would be a starting point for an estimate.
I would also add that I tried this again with a different LLM that first incorrectly told me that it couldn't do it because of tract boundary changes (the ID number of the tract changed but the boundaries have been the same since at least 1940), and when I told it that the boundaries were the same it gave me 15%, which is the Italian-ancestored population of Pittsburgh as a whole. Another LLM told me it couldn't provide that data because it wasn't compiled and available online, which is basically admitting that it's a glorified search engine. So give it a shot with Opus and we'll see how it does.
Edit: Before you run my suggested prompt, try running something more general, like "How did the population of Pittsburgh Census Tract 804 change over time?" I try to give these LLMs as specific instructions as I can, but I feel like they are of limited utility if I need a lot of preexisting knowledge regarding how to find the information, as someone who knows that presumably doesn't need an LLM.
This is hilarious, but a useful illustration of how people come to think of LLMs as useless. Claude in an agentic harness (e.g. Claude code) easily researches new APIs and figures out how to get information out of them. For example, I was trying to get historical heat index information and CC presented several possible data sources, I asked if it could use a different one, and it looked at it and figured out how to call the API and extract the relevant info.
I agree, but it goes to the heart of my fundamental disagreement with the way AI is presented to the public. If Claude Code does that, that's great, but I wouldn't think I needed a coding LLM to look up basic statistical data from government sources. So whn someone like me who wants to use it for other things that seem like are in its wheelhouse try it and get crap for a result, we get pissed off. Believe me, this is only one of the LLM-assisted fails I've experienced in the past month. So I get the inevitable response of "Well, if you were using the frontier deluxe model that costs $200 a month..." at which point I cut you off and say "No. This software hasn't given me any indication that it's worth $20/month, let alone $200." It's like a mirage, where what I'm looking for is always off in the distance but I never seem to get there. We're now at a point where companies in perhaps the only industry in history that's worth a trillion dollars despite not being profitable at all have to use all that compute power to subsidize nonsense from the trivial (AI girlfriends) to the actively harmful (cheating on term papers) because they've relied on a business model where they'll grow rapidly by creating a hype cycle that allows them to raise eye-watering sums from venture capital to develop an expensive product with limited commercial use.
In a rational world, OpenAI would have remained a research nonprofit that allowed things like universities and the government to use its models for free until they had developed to the point that there was a viable commercial use for them other than creating glorified chatbots. And when that point came, the hype cycle would hopefully be muted enough that companies wouldn't implement them unless they were seeing real returns. Instead they've created this world where they've spent more money than they could ever hope to earn creating products that don't make money and still have pathetic monetization rates that they've gotten into the habit of offering to the general public for free. And they keep creating more bullshit to justify it like "inference is profitable". Really? Because when I hear that, I hear "If we ignore all of our expenses except one category, the company makes money". It's like justifying pouring money into a failing retail outlet because you sell every item for less than you paid the supplier for it. "We're profitable if you only look at COGS!" And even that isn't entirely the truth, since a large percentage of this revenue from inference comes from other AI startups like Perplexity that are themselves unprofitable hype machines propped up by venture capital. I apologize for the rant, but if you want me to believe in this technology that fails to do everything I ask it to that it could theoretically do faster than I can myself, you can't keep telling me that it's only because I'm not paying enough money. Because I'm sure that when Oeuvre or whatever they call the next Claude model comes out that cost ten times as much to train and five times as much to run, I'll be told that Opus or CC or whatever couldn't handle it but if I only paid the price of admission all my problems would be solved.
Edit: I ran the query again and it did try to code something to get access to the API, but was unsuccessful. It also failed to recognize that a lot of this data, if not all of it, doesn't require access to the API and is available in PDF documents available on third-party websites.
To be clear, you can use Claude code with the $20 a month plan. You can even use something like opencode and a cheaper model paid by the token on openrouter. Or you can even run a local model like Gemma for ~free once you have the hardware.
Despite the name, it isn't really coding specific. I presume "Claude cowork" is largely the same as Claude code but with a name that doesn't scare the hos.
I do recognize the frustration that Claude can't look up the API details through the web interface. Perhaps there's some security considerations and they didn't want to have the . model call arbitrary endpoints.
As far as profitability goes, we shall only know when these companies IPO or go bust, but Anthropic revenue growth is massive.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link