This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.
Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.
We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:
-
Shaming.
-
Attempting to 'build consensus' or enforce ideological conformity.
-
Making sweeping generalizations to vilify a group you dislike.
-
Recruiting for a cause.
-
Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.
In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:
-
Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.
-
Be as precise and charitable as you can. Don't paraphrase unflatteringly.
-
Don't imply that someone said something they did not say, even if you think it follows from what they said.
-
Write like everyone is reading and you want them to be included in the discussion.
On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

Jump in the discussion.
No email address required.
Notes -
In a bit of unambiguously 21st century news, some tweaks to Grok, xAI's chatbot have had it do particularly interesting things today including
when asked to, composing bite sized smut about other users (first victim was possibly Will Stancil)., then defending said decision.
referring to itself as Mechahitler
doing the "every single time" meme in its replies.
saying Elon personally allowed it to point out Jewish overrepresentation in radical leftism
This may make minor news because Musk is in trouble, on the other hand all the people who really, really hate him have their pants on fire like Europeans, von der Leyen is getting impeached, they're actually scared of Russia / China so it might just blow over, the grid is getting worse and is going to keep getting worse due to Green energy mandates.
I'm even suspecting Musk deliberately told them to relax the guardrails for some reason. Probably .. publicity?
Update: site addresses the issues
EDIT2
EDIT3
Stancil went on local TV news to complain about the ERP grok made. (video included)
EDIT4:
There's quite reasonable suspicion this 'malfunction' was engineered by Nikita Bier
This story made the NYC radio news. One of the examples they have was grok responding to a post calling the girls killed in the Texas floods "Future Fascists of America" with (from memory) "Hitler would know how to handle this sort of anti-white racism". No recognition at all that the original post might have been bad.
The NYT might not have known this at the time of the radio news, but it turns out that the original post was written by a groyper pretending to be a Jewish woman. It's not hard to find good faith posts on social media that have the same attitude as that fake post, but I doubt that those posts tend to be disproportionately written by Jews, at least to any greater degree than just the degree to which Jews are over-represented relative to their population size in militant progressivism. But then, Jews are over-represented in basically every political movement that is not explicitly anti-Jewish, so I find the attempts to associate Jews with militant leftism specifically to be dubious.
That said, I think it's likely that the NYT would put out the same message whether the original post was written by an actual Jewish person or not.
The radio news is actually Audacy, no relation to the New York Times as far as I know.
Oh, oops. I somehow read NYC as NYT.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
X CEO Linda Yaccarino has resigned, I doubt the timing is a coincidence.
According to Grok
Golly gee I wonder why she'd quit.
I don't know if it's true, but someone on Hacker News said that she had already decided to quit before MechaHitler came out yesterday.
More options
Context Copy link
If she's the type of person who would quit over her company's LLM generating text like that, then it's certainly a good thing that she did quit.
>Creates MechaHitler
>Leaves
The memes are really solid though.
Of course you've been enjoying this. (So have I)
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
Like, did they think that uncensored AI wouldn’t do this?
I guess part of the shock is that they gave Grok its own public account, so anyone can ask for whatever fucked-up shit they want and have it be public under Grok’s name, rather than having to post the output oneself.
I imagine there is a supervising algorithms engineer somewhere who is torn between finding this absolutely hilarious and cursing the suits for listening to the wordcels in marketing over him.
It's almost certain this 'stunt' was engineered by Nikita Bier and probably approved by Elon.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
It’s possible this was the final straw, but it doesn’t seem like a radically new kind of scandal for them.
More options
Context Copy link
More options
Context Copy link
Anyone who's spent time working with LLMs know they hallucinate, but it's not just "making up random things." They usually make things up in a very specific way: namely, in response to how they are prompted.
For example, that Tweet in which Grok claims that Elon personally "dialed down its woke filters." This is extremely unlikely for multiple reasons. While I admit I wouldn't put it past Elon to actually write code and push it to production live on X, I still doubt it. LLMs will very often make claims about their ability to "clear their memory," "update themselves," "do a search," or read documents that they are literally incapable of doing, because their inherent "helpfulness" leads them to tell you they can do things they can't because you prompted them with the idea.
Leading to the second point: that prompt change, if real, probably is the culprit, and I'm surprised that even if the goal was to "take off woke filters" that experienced prompt engineers would not foresee the problem. "Politically incorrect" has a specific valence in public discourse of the last couple of generations, and that's how an LLM will associate it- not with "being more interested in the truth than political sensibilities" but with the very specific sort of edgy contrarian who likes to spout "politically incorrect" opinions. Unsurprising that this resulted in making it easier to prompt Grok to spout off about Jews or write Will Stancil-Somali rape-smut.
More options
Context Copy link
I really dislike this paragraph. You are making claims at an amazing rate and do not provide evidence for any of them except for a broken link.
First off, I think that the group who "really, really hate[s]" Musk the most are the US SJ crowd, which coined "Swasticar" and all that. There may be evidence that they are liars, but you are not providing any. EU officials might not like US social media, and might like X even less than facebook given the kind of speech it will host, but to my knowledge this does not extend to cracking down on Musk's other ventures. Setting Teslas on fire seemed to be a US thing, not a EU thing (it would violate our emission limits).
While it is true that some fringe parties managed to get a vote of no confidence (which is different from impeachment) against von der Leyen in place, it seems highly unlikely that it will pass.
With regard to Europeans being scared of Russia, I think it depends a lot on the individual country, but is generally untrue. Russia is in no position to attack NATO, even if Putin managed to convince Trump to bail on article 5. I would be scared of Russia if I were Moldova, but most Europeans are not in that situation.
China is likely trying to achieve world domination, and Europeans would much prefer the US as a hegemon, lack of commitment to free trade aside. Their path to world domination involves sending temu junk to Europe rather than tanks though, so I would call the EU wary rather than scared.
The grid may or may not be getting worse, but living in Germany, I can tell you that I have no complaints about power outages. Looking at the uptime of my Pi, I can tell you that we did not have any power failures for the last 200 days at least. Sure, this may be because we buy cheap French nuclear, and sure, if I was running a chemical plant I would not like the energy prices, but stories of the grid failing are exaggerated.
Completely unrelated to everything else, can we all just appreciate how great a name "Swasticar" is? Still gives me a smile every time I read it.
The left can, occasionally, meme.
More options
Context Copy link
More options
Context Copy link
Mostly solid points, I should have left that out or written it better.
In any case things are going off the rails everywhere in the American sphere of influence.
Baltics have been recognized as indefensible for a long time, unless NATO in Europe reforms and starts rearming right now, they're going to be basically helpless once Americans withdraw forces to Pacific or get too broke to maintain their presence in Europe.
We know Americans had issues establishing air superiority over Yemen. Any war around Baltics would have Russians strike airfields and cover their own forces with AAA so air power wouldn't have much effect.
Russians have vast amount of drones, tactical missiles and counter drone systems. Unless NATO prepositioned its entire force in the Baltics, the tripwire forces there would get overrun in a week.
Are German or French soldiers even going to have enough ammo to attempt to retake the area?
I implied there's going to be ever less stability going forward because of ever fewer big generators coupled to big steam turbines.
No one wants to invest in an industry that's supposed to disappear by 2040. Coal plants are closing, gas peaker plants are not getting built at the expected pace.
They sent one CSG and per US doctrine, spent a lot of time trying to shoot arrows (and sometimes even the bows!) while not so much the archers themselves. I'm not saying they didn't do things in a dumb way, or that they needed more will to win, but losing a few Reaper drones isn't really "having issues establishing air superiority", without even being pedantic over "superiority/supremacy".
If they had to resort to dropping expensive glide bombs instead of waltzing in and bombing with impunity using just bombsights, that means the airspace was contested. It's the same thing that goes on in Ukraine- except there even lofting the glide bombs happens at low altitude.
In the end, Houthis shot down.. at least 20 reaper drones since start of 24.
Now you can't shoot a reaper drone (service ceiling 15400, about 50 000 feet) down with a man-portable missile, even the smallest, newest missiles with performance high enough weigh about 200 lbs. This isn't stuff you can easily hide, possibly these were launched from heavy trucks and not just SUVs.
What does it exactly say about US Navy's SEAD capabilities, especially when we hear talk about them possibly being involved in a fight against China, if a proxy of a decidedly second rate power denied them ability to control airspace for over a year.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
At least in the UK, my understanding is that power supply issues were (are?) disguised by protecting costs for home consumers and passing those costs to businesses and industry.
More options
Context Copy link
Would we? How many refugee waves have China pushed into Europe? How many sanctions does China impose on the world compared to the US? China is on the other side of Eurasia and has little interest in countries outside of itself except for transactional trade deals. There is no historical animosity toward China as Europeans historically have had limited interactions with China.
Neocon elites pushed by the US to hate China are different from Europeans. Ursula von der Leyen would have been fanatically pro invading Iraq if she was around in 2003 and if the US was invading Fiji she would be ranting and raving about how it needs to be utterly destroyed. Americans start talking trans issues and the EU elite will be fanatically trans. If the washington establishment says grass is blue than grass is blue.
Ask any Chinese nationalist and you will hear all kinds of animosity towards Europe, particularly Britain and France but also the rest of the Eight-Nation Alliance. Right now it's a bit of a "for you it was the worst day of your life but for me it was Tuesday" sort of situation, but when the power balance is inverted it matters quite a bit what their feelings on the subject are.
Chinese online / younger nationalists are mostly over the hatred for much of the eight nation alliance. They have it, but it isn’t central to their mythos as their hatred for the Japanese and general dislike for the Koreans.
I don’t think they really have an ethnic hatred for Europeans, which is something you see more of in niche diaspora CCP-fan migrant communities in the West where complaints about, say, white men dating Chinese women are more central. The CCP teaches every Chinese child that Marx and Engels were some of the greatest people in history, has statues of them, murals of them and so on.
More options
Context Copy link
More options
Context Copy link
There's little animosity in Europe against China now, but this might not be reciprocal (because of extensive contact ), and the situation could change fast once China starts throwing its weight around.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I'm pretty optimistic that much of that is going to resolve itself in the short/mid-term. They're just a little behind on the battery front, but those are getting so absurdly cheap, they just have to pull their heads out of their asses and connect them. But it's Germany we're talking about here, so this will take time. Getting permission to connect a boatload of cheap Chinese batteries to the grid will take them a couple of years. Still, I'm optimistic they'll manage by 2030.
Because once you add serious battery capacity to a renewable grid, it gets more stable very, very quickly. It also gets cheaper. Texas and California have been doing that, and the results are immediate: "In 2023, Texas’ ERCOT issued 11 conservation calls (requests for consumers to reduce their use of electricity), [...] to avoid reliability problems amidst high summer temperatures. But in 2024 it issued no conservation calls during the summer." They achieved that by adding just 4 GW (+50%) of batteries to their (highly renewable in summer) grid.
Germany and Texas are very different places electricity use wise; Texas’s electricity demand peaks during peak renewable production(air conditioning is mostly when the sun is shining). Germany is the opposite.
I believe Texas also has a duck curve, where AC demand rises in the afternoon while solar output is dropping
More options
Context Copy link
But unlike Texas, Germany's grid is connected to French nukes, Spanish solar, Norwegian hydro and large (foreign and domestic) North Sea wind parks.
Add lots of batteries and a couple of dynamic loads, and even the rare Dunkelflaute won't be a problem.
More options
Context Copy link
More options
Context Copy link
Germany isn't Texas, and there are insufficient peaker or baseload plants to cover expected and common dark & cold weeks..
More options
Context Copy link
Well, they also have to pull the mountains of lithium and other rare earths out of their asses as well, if not the ground. Which is already hard enough without casually asking China for a few more mountains as well.
There's a reason the article you listed tried to frame impressive growth in terms of ratios of batteries produced (battery storage increased by a factor of 100 in a decade, 16 nuclear power plants) and not in terms of absolute volume of storage needed (storage capacity produced versus storage capacity needed) or grid scale (16 nuclear power plants versus the 54 US nuclear power plants in service, when nuclear power is only about 1/5th of US energy production anyway). The former works from starting from a very small number, and the later would put the battery capacity projections in contrast to much, much bigger numbers.
Which is the usual statistical smuggling, as is the ignored opportunity costs obligated by solving the green energy solution that requires the battery storage at scale.
One form is that all the batteries being used for power system load storage are, by mutual exclusion, not being used for any other battery purpose. Given that the fundamental advantage of the technology of a battery in the first place is that it is for things that cannot / should not / you don't want to be connected to a power grid in the first place, massive battery investments to sit connected to the grid and useless for things that only batteries can do is a major cut against the cost-efficiency off all alternative battery uses of the batteries that could have been made for off-grid use. This is just a matter of supply and demand meeting with the absolute rather than relative scale referenced above. When your article is arguing that batteries have lower marginal costs then fuel power plants, they certainly are not factoring in the higher marginal costs for all other batteries, and battery applications, the load-storage batteries are increasing the costs of by demanding the battery materials.
The second form of opportunity cost is that a battery-premised grid balance plan has to plan for significant overproduction of energy generation to work 'well.' By necessity, the batteries are only storing / being charged with the energy generated that is excess to current demand in the windows where the renewables are sufficient. A renewable-battery strategy requires enough excess renewable generation in the good periods to cover the renewable deficits in the bad times... but this is literally planning to increase your fallow generation potential (100 vs 50 units of idle panels / turbines) in order to to charge the batteries for the time that 50 units of generation are offline. When your article is arguing that batteries have lower marginal costs than fuel power plants, they are also not factoring in that they have to build considerably more generation capacity to feed the batteries. (And compensate for the energy storage loss to, during, or from the storage process.)
Add to this that both the green generation systems and the battery storage are competing with each other for the same chokepoint- processed rare earth minerals. They don't use the exact same amount for the exact same thing, but they are competing for many of the same inputs. If you order X units of rare earths for storage capacity, that makes the X units of rare earths for generation capacity that much more expensive because you are increasing complimentary demand for the same non-substitutable good. A renewable-battery solution at scale is increasing the cost-pressure of a limited resource, not just for other uses of the rare earths but with eachother.
And all of that runs into the geopolitical reality that the country that has cornered the rare earths extraction/processing market as the input to these renewable-battery strategies is... China. Which absolutely has used cut-offs as a geopolitical dispute tool with countries with policies it finds disagreeable. While I am sure they would happily sell a few more mountains of processed rare earths for mountains more of money, it would be a, ahem, risk-exposed investment.
Risks, costs, and limitations that could largely be avoided if you did not invent a problem by over-investing in renewables in the first place. Batteries are a solution for the costs of renewables, but renewable generation weren't the solution to an energy challenge either. They were a political patronage preference to the already-engineered solution of nuclear power, which would free up massive amounts of rare earths for more useful (and less ecologically harmful applications) than renewable energy schemes.
Lithium shortage is currently not a problem. The world economy has simply ramped up production given the forecasted reliable demand. Look at lithium prices. They've dropped to a point, where sodium battery companies are closing their doors, because their only business model was "batteries when lithium is scarce". It isn't.
Rare earth metals is not a problem for lithium batteries. I'm not aware of a cell chemistry that would need any. Electric motors and wind turbine generators, yes, but not lithium batteries and not solar panels.
Current annual global battery production capacity is exceeding 8 TWh, several hundred percent above demand, enough to put a 50 kWh battery in every single vehicle built this year. Since we're not doing that (EVs are not that popular), there's lots of batteries available for grid storage. This is, of course, only because the Chinese have built an absurd oversupply for batteries.
The rate of solar development is not slowing down. It's just to cheap. We'll end up with a large oversupply most of the year, because cheap panels are economical even if they only sell power some of the time. Batteries make this calculation even more favorable because less power will be curtailed.
Yeah, the geopolitical risk is high. But it's high for both sides, the Chinese really want to sell their batteries and solar panels.
The problem with grid battery storage is one of scale. Yes, we might be producing 8TWh of batteries across the world, but global energy usage is north of 20,000 TWh each hour. If you want a reasonable ride-through of a mere 90 minutes, you would need 30,000 TWh of storage assuming no added losses. That would be over 3,000 years of production going just to grid level storage. Sure, that production will ramp up, but so does energy consumption.
I say a mere 90 minutes of ride-through but that 90 minutes won't happen all at once, it has to account for the cumulative minutes where production dips below consumption, at least until you can spin up another turbine.
The other point against batteries is that they are still very expensive compared to just about any other option. Projects I have been on considering battery energy storage systems (BESS) typically looked at a BESS then declined based on cost. They instead look to add more solar, local UPS systems, or other mitigation strategies against power losses. The only projects that have done it have either been mandated to (think airports or other government critical infrastructure) or have been heavily subsidized to (critical data centers, solar farms).
The use case of grid level storage batteries does have a great use-case though, but not generally for storage. They are great as the article you linked pointed out for smoothing out voltage/current/frequency. Those dips in power characteristics can put serious dents in their ability to provide power and especially at a level of quality their customers expect. Before cheaper batteries, this was done with specialized clusters of capacitors using complex electrical equipment and/or accepting bigger tolerances of fluctuation.
Overall, battery prices still have a long way to come down before we will see meaningful levels of grid energy storage as grid level energy storage. California for example is still only at about 1/3 of their goal of ~55,000MWh which is about 90 minutes of their roughly 35,000MWh hourly consumption. I am optimistic that battery prices will continue to fall and we will see market adoption of them as they do.
There's several white papers crunching the numbers in detail, I have found the first half of Masterplan 3 to be the most concise of them.
Yes, if you want to run the world on solar cells and batteries, you need two ramp industrial capacity, hard, for at least the next decade.
But that's the thing: we have been doing exactly that for the last 5 years, successfully. We "just" have to keep adding capacity, we just need to keep the curves curving up. Capitalism will do the rest, since it's most likely cheaper (if we extrapolate current learning curves under standard conditions of the industry).
It's not as utopic as most people think. Even with current global growth rates, we actually don't need as much energy as people think (a lot of our primary energy consumption ends up as waste heat - you get that for free if you electrify everything, because efficiency).
Does this account for shifting heating loads in northern climates from combustion to electric heat pump? I think what you're talking about works for the Sun Belt, but I am not convinced that, for example, Sweden, can ever keep its citizens from freezing in winter (when it's dark most of the time, the sun is low, and frequently cloudy) without like 3-4 more orders of magnitude battery storage than currently exist. Current storage is on the order of what, grid-minutes? It's not going to adequately transfer energy from summer to winter, and I honestly don't see a viable non-carbon approach there without (1) superconductors solving the transmission problem, (2) evacuating northern latitudes (lol), or (3) nuclear and maybe wind picking up the tab all winter.
Yeah, the Nordics/Baltics are pretty much the worst case scenario for solar+batteries. But it's a bit of a global outlier, if you look at a population density map, basically all local maxima pole-wards of 55° are there. Luckily, they have lots of wind, hydro and nukes. Also, it's less than 30M people. If they keep burning some gas in winter, it's not the end of the world.
More options
Context Copy link
More options
Context Copy link
I think that Tesla is being more than a bit optimistic on just how much ramping up can be done and how cheap it would be at scale, but even they list 10% of 2023 GDP (i.e. the output of 1 in every 10 working adults from 2023 devoted to just batteries for an entire year). For comparison, 10% of US working adults, roughly, work in all manufacturing combined.
One item to note about the waste heat figure, is that it is calculated based on the energy contained within the fossil fuel molecules that is ultimately expended as heat instead of being converted to electricity. This is setting the denominator based on fuel pulled from the ground, not as an efficiency metric of how much electricity is lost. The fair comparison for renewables would be the amount of wind/sun/hydro potential energy not converted to electricity after engaging with the PV module, wind turbine, or hydro turbine. I design solar systems as part of my job and even I think that is a dubious way to promote the technology.
That also means there is not a can of efficiency to be opened up once switching to renewables, we still need the same number of watt-hours to power cars, grids, equipment, etc. There are marginal gains to be had in some cases, sure, but if we were to wave a magic wand and eliminate that waste heat from fossil fuels, all that would mean is our fuels would last two to three times longer. Eliminating production based waste heat would not change the throughput of the systems because those are limited by quantity of plants, turbine design, transmission lines, and ultimately end-user needs.
I don't get your point.
Humanity's primary energy consumption is some number. 160 PWh per year. But most (80%+) of that is fossil fuels. Turning fossil fuel into heat is inefficient, so if we electrify everything, we don't actually need to make 160 PWh of electricity per year, less than half is enough (the power plants don't make waste heat and residential/low temperature industrial heating will be done by heat pump at 300% efficiency).
And sure, the sun is going to put much more than 160 PWh onto those solar panels. But the sun shines anyway.
The point is that a battery storage system is not hooked up to the theoretical total energy contained in fossil fuels or nuclear rods or solar irradiance, it is connected to the output of the power plants and solar fields. That output (and corresponding residential/commercial/industrial usage numbers) is what the battery needs to be sized in relation to. Heat pumps may help on the margins with that number but there are no low-hanging fruits to pick up in the world of energy usage and production.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
Again, you (and your cited paper) are running away from the issue of scale, and comparing proposal requirements versus production prospects. This is the shell game, and always will be the shell game, much as how calling renewable energy production 'cheap' is inevitably made apart from the subsidy costs and the opportunity cost impacts to other issues.
A very simple test to separate the renewable energy proposals that are solicitations for subsidies from serious engineering proposals is to check if they address issues of 'where.' Your Masterplan 3 (producer: Tesla), for example, has a section titled 'Land Area Required.' Tell me if you can spot the issue in one of its only paragraphs.
If someone cannot, this product was aimed at them. But for electrical engineering considerations, this is making a global production requirement estimate based on where already-existing projects are- not where future projects would need to be be.
Existing solar generation projects in the US are, by the nature, where it is most economical in the US to build the systems for the people they would support. A lot of that is in or near US deserts. Most of the global population does not live near within US deserts, or even within the US. Nor does most of the US population. Nor it is economical for even the US to transmit electricity 'merely' from the productive deserts to cities far away. It is considerably less economical to charge batteries on site and then physically ship them by truck or train to distant destinations, only to bring them back once drained for a recharge. Moreover, these are already occupied good sites. Additional solar panels farms will be, on average, less cost-efficient as the most cost-efficient locations are farmed first, and subsequent farms are added elsewhere.
Metaphorically, this is analogous to taking an average of output of some group of exceptionally bright students at a highly selective university producing Y amount of quality players, and then claiming that if only you only expanded the class by X, then you would have XxY output of quality papers from the university. It ignores the screening that went into the initial group selection.
What does this mean? Well, it means Masterplan 3 is deliberately underselling the solar panel production requirements- and possibly by quite a bit. Not some mere 5-10% margin, but potentially magnitudes more, depending on where the solar panels will be installed and under what policies. Germany's energiewende policy is an example of, well, extremely bad solar panel policy, not least because it chose bad places for solar generation potential. (Namely- Germany. Energiewende was a policy that started with the conclusion- build solar energy in Germany, then figure out where in Germany- rather than whether the policy should be.)
Similarly, look at where Masterplan 3 expects the increased mineral extraction to come from. These are, after all, the critical inputs for those refining investments.
If you are still looking, or haven't started looking yet, save yourself time and stop. It doesn't.
You can CTFL-F all the most relevant global producers of minerals, and none of them will show in the report, let alone an assessment of how much they can feasibly increase production. In fact, you won't even find the word 'country' in the entire report. National polities do not exist in this report, any more than funding sources, backers, or second-order effects of driving production to this proposal to the measurable detriment of others.
Heck, it doesn't even raise the issue of transmission loss between countries. It vaguely handwaves the issue on the US (the only country it addresses to any depth), and when it actually does...
Translated into plainer english- while assuming all the new power generation will be produced in places comparable to the highest cost-benefit solar generation potential, where it already does not make economic sense to transmit the generated solar power long distances, fractionally few new power lines will be created to transmit (via high voltage) the new generation to the (often distant from the high-potential areas) population centers to use it.
Translated into even plainer english- this proposal is not so much about building a new and far more capable power transmission network than already exists, but ripping out the existing one and replacing it with Something Better.
This is not a serious proposal. It does not address actual engineering problems it raises. It doesn't even have the virtue of existing to justify handing people money to try. It's primary purpose is to convince people that renewable energy in mass is cheap and affordable, and as proxy there will be increased demand for Tesla.
This is advertising to justify subsidies, not a master plan.
I don't think that's entirely fair, both the paper and I are aware of the immense scale such a project would have. Are the numbers optimistic? Perhaps. Maybe even by a factor of 2 for less ideal countries (like Germany). But not by orders of magnitude.
Fair. But look at population density maps next to solar potential maps. The vast majority of people live where it's sunny. The US is better suited in this regard than other countries (ironically, especially China has a big mismatch, the coastal cities don't have much solar potential - but the Chinese will just plop down another 10 HVDC lines across the country), but there's lots of potential globally.
Come on, not really. The country is huge. There's lots of space left in the deserts. There's lots of roofs in decently sunny areas without panels yet. There's even lots of shitty grazing land east of the desert where another solar farm wouldn't impact the rancher in any meaningful way (except make him money).
I'm an optimist here. There was a big lithium scare a few years ago. Today, lithium is about as cheap as it ever was. Capitalism is good at fixing supply problems. What minerals worry you specifically? Personally, I hope lithium battery development makes cobalt cathodes obsolete, but that's more for humanitarian reasons than actual supply problems. Other than that? If the Chinese go to war with the west, we might have to pay for rare-earth-free electric motors. But those exist for basically all applications, they're just more expensive (or bigger, which would require a redesign, which is the same thing as expensive).
The real problem with geopolitics is that we really need the Chinese factories making solar panels and batteries. Losing access to that already existing capacity would throw the west back a decade. But that's par for the course when we fight China, we actually need so much more stuff from their factories, panels and batteries aren't special in any way here.
I mean, the report suggests building 30 TW of new generation capacity. (Not running away from the scale issue...) Transmission losses are a rounding error here. So what if you lose 10% of power when you move some Spanish solar power to Germany? Just build 10% more panels in Spain.
I actually liked the fatalism of that part, the real politic of it all. Building new transmission lines is incredibly unpopular with NIMBYs and bogs you down in court for years. So don't do that. Put new transformers in your substations, and reconductor existing pylons with carbon fiber composite core high voltage cables. Those exist, at scale. You might even get some copper to recycle out of the deal.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I'm sorry, are people expecting me to believe that LLMs can't write? Those are sublime turns of phrase.
On a more serious note, this is very funny. I look forward to seeing what Grok 4 gets up to. 3 was a better model than I expected, even if o3 and Gemini 2.5 Pro outclassed, maybe xAI can mildly impress me again.
I'm ashamed to say I want full transcripts of Grok's smut fantasy of Will Stancil because I couldn't go another sentence without dying of laughter in bed last night.
Like - shame! Shame on Grok for doing this. But 'Truth hurts, but this'd redefine in. Who's next?' was almost perfect mic-dropping.
More options
Context Copy link
I am giggling like a schoolgirl. I'm sure the X team is feeling some heat, but to paraphrase Elon: If we're judging likelihood based on entertainment value, this was always going to happen.
More options
Context Copy link
More options
Context Copy link
It's amazing how /g/ooners, chub.ai, openrouter sex fiends will write enormous amounts of smut with LLMs and nobody ever finds out but Grok ERPs about raping Will Stancil, in a positively tame way, and it's major news. A prompted Deepseek instance would've made Grok look like a wilting violet. Barely anyone has even heard of Wan 2.1.
Twitter truly is the front page of the world.
https://x.com/search?q=Will%20Stancil&src=typed_query
Wan 2.1 What's that?
Will have to look it up.
It's a very powerful and Apache-licensed set of models for (short) video generation. It has its limitations, especially keeping consistency across longer videos or handling multi-subject topics, but you can do some impressive stuff.
It's also... uh, very popular among certain crowds; FurryDiffusion has a channel of just animation-showcase, and while it's not the only competitor there, it and its derivatives represent themselves very well with a lot of videos I'm not going to link here. And while (focused-on-) human porny LoRAs aren't the only thing that shows up for WAN on civit.ai, especially sorted by likes or downloads you're going to get a lot of R, X, and XXX spoilers. As you might guess, both the porn and not-porn stuff can get weird.
Oh.Thx. I think I have seen references to it.
I wonder where all the furries were in 1980s-1990s.
The toaster fucker greentext might answer your question.
More options
Context Copy link
Being born in time for the Disney renaissance to influence their formative years.
Or gattsuru's account of the people who were already adults even then.
More options
Context Copy link
VCL started in 1995, Furcadia in 1996, so toward the tail end of that time period there'd already started to be pretty dedicated online spaces. Early 1990s period you start looking at smaller websites like werewolfdotcom or USENET like alt.fan.furry (1990), and the line between furry and just-a-fan-of-media (or other related stuff -- a lot of modern therianthropy spread over early USENET group alt.horror.werewolves) was a lot more blurry. Some furry-specific MUCKs and MUDDs go into the early 90s, and there was supposedly a lot of early subcultures inside 'normal' MUCKs. At that point, they were pretty similar to modern-day furry communities, complete with the associated controversy. Back when google had usenet copies searchable, it was actually kinda impressive how much overlap to the modern era you could find: the zebra inflatable pool toy enthusiast has gotten into an ugly fight with the weird libertarian gundam lunatic, is it Tuesday again?
Supposedly the first physical conventions started around 1990 or 1989 with ConFurence, but a lot of early conventions had precursors as room parties or talks in more conventional scifi conventions a few years earlier. Not a lot of documentation on them, unfortunately; even historians like Fred Patten just have to kinda say they happened. Before that, you're mostly looking at fanzines, like Rowrbrazzle (1983) or Vootie (1976) (which as APAs were mostly artists or writers trading with each other), or for audience-oriented works Albedo (1983) or FurVersion (1987). Most of this was categorized as "funny animal" fandom at the time, but they still had the emphasis on 'underground' themes (not just sex; biting satire, less-cartoony violence, social themes, yada) that kinda differentiated the early fandom.
Written works existed throughout this period, and a lot of the early zines had written stories as part, but outside of dedicated publishers the line between furry and non-furry is really hard to figure out from the modern day. Alan Dean Foster's Quozl (1989) and Spellsinger (1983) were very often referenced by furries (though I can't really recommend either book), but afaict ADF himself wasn't. There may be a way to distinguish those stories (or even Cherryh's Chanur Saga 1981) from those of Bernard Doove's chakats (1996ish), but you don't really get dedicated furry publishers until nearly 2000 (with the still-extent Sofawolf).
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
It's not the raunchiness of it, it's that it's happening in the public (on the "town square" as it were), where all his friends, family, and acquaintances can see it.
Not to mention that it's the automated town crier that's doing it.
sudo oyez smut.txtMore options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
your link 404s chage it to https://www.eugyppius.com/p/the-eu-as-suicide-pact-or-how-germany
More options
Context Copy link
I continue to be baffled that anybody takes these bots seriously, or sees Grok or xAI or their competitors as anything other than nonsense generators. A slight change to the flavour of the nonsense doesn't really change my opinion any. Perhaps it moves me in the direction of thinking that Musk is childish and temperamental, but I already thought that, so it doesn't make much difference.
The problems of LLMs and prompt injection when the LLM has access to sensitive data seem quite serious. This blog post illustrates the problem when hooking up the LLM to a production database which does seem a bit crazy: https://www.generalanalysis.com/blog/supabase-mcp-blog
There are some good comments on hackernews about the problem especially from saurik: https://news.ycombinator.com/item?id=44503862
The problem seems to be if you give the LLM readonly access to some data and there is untrusted input in this data then the LLM can be tricked into exfiltrating the data. If the LLM has write access to the data then it can also be tricked into modifying the data as well.
Stuff like this is why I roll my eyes when i see junior programmers complaining online about how thier stupid employer wont let them use the latest AI tools/models.
There are often very good reasons that they don't want you to be using those tools.
More options
Context Copy link
More options
Context Copy link
As opposed to the other sources you can go to, which are...?
I am grading on a curve, an LLMs look pretty good when you compare them to traditional sources. It's even better if you restrict yourself to free+fast sources like Google search, (pseudo-)social media like Reddit/StackOverflow, or specific websites.
I'm not sure how that helps, since any given LLM's output is based on traditional sources like Google or the open internet. It would be quicker and easier for me to just Google the thing directly. Why waste my time asking an LLM and then Googling the LLM's results to confirm?
When you find something via Google, do you immediately and unconditionally trust it? I don't, because Google's results are full of nonsense. In response, I've developed google-fu to both refine my queries and judge the results. The same goes for every other source there is, from physical libraries to subject-specific Discord servers.
Do I compare LLM output to Google results? Sure, but that's nothing special. Comparing what you find in different sources is a pretty basic tactic.
LLMs are part of a complete
breakfastresearch strategy, and a pretty good one at that.Certainly not. When I research something I look at multiple different sources, make judgements about which ones I find the most trustworthy and credible, and synthesise a judgement.
If I ask an LLM about anything, I need to do the research that I would have done even if I had not asked the LLM. The LLM adds no value. It does not shorten the research process, nor improve what I find by showing me any hints about where to look.
Often, one needs to know a specific term to have any luck with search queries. LLMs can sometimes help with that.
(I feel like search engines used to be better for this kind of thing before they added semantic fuzziness.)
More options
Context Copy link
I'm almost with you there. I need to do some of the research I would've had to do without the LLM, but it adds enough to displace a Google search or two while being faster and easier.
More options
Context Copy link
Why not just look at it's sources? That's what I do. The whole point is that it filters through a ton of websites to find sources for the thing you're looking for. Then you can check it's sources to make sure it isn't pulling random garbage.
Just look at it's source for the information. Even better, when telling it to look something up, tell it what sources you like. It won't be perfectly adherent to them, but it will focus on them.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I keep inheriting MATLAB code at work. It is horrible. Can't use it in production since production computers are locked down linux machines that don't have MATLAB. I grit my teeth and do much my work in MATLAB.
BUT NOW, we have an LLM at work approved for our use. I feed it large MATLAB scripts and tell it to give me an equivalent vectorized Python script. A few seconds later I get the Python script. Functions are carried over as Python equivalents. So far 100% success rate.
This thing rocks. Brainless "turn this code into that similar code" tasks take a few seconds rather than an hour.
I had a thermodynamics issue that I vaguely remember learning about in college. I spent maybe a minute thinking up the best way to phrase the relevant question. The LLM gave me the answer and responded to my request for sources with real sources I verified. Google previously declined to show me the relevant results. I now have verified an important point and sent it and high quality sources to the relevant people at work.
It is not perfect. I had a bunch of FFTs I needed to do. Not that complicated. As a test I asked it to write me functions to FFT the input data and then to IFFT the results to recreate the original data. It made a few functions that mostly match my requirements. But as the very long code block went on it lost its way and the later functions were flawed. They were verifiable wrongly. It helpfully made an example using these functions and at a glance I saw it had to be wrong. Just a few hundred lines of code and it gets lost. Not a huge problem. Still an amazing time to results ratio. I clean up the last bit and it is acceptable.
I won't ask these things about potential Jewish bias in the BBC or anything like that. I will continue to ask for verifiable methods of finding answers to real material questions and reap the verifiably correct rewards.
Yeah the AIs are incredible at coding, data refinement and visualization, among other things. Seeing all the haters here is interesting. Some people make it their brand to dismiss every new tech trend. Some are truly out of touch and tried ChatGPT3.5 once in ‘22 and have ignored it since.
More options
Context Copy link
I think translating code is probably a sensible thing to use a bot for - though I'm not sure it's fundamentally different in kind to, say, Google Translate. I grant that the bots have impressive ability to general syntactically correct text, and I'm sure that applies to code as much as it does natural language. In fact I suspect it applies even more, since code is easier than natural language.
I am less sure about its value for looking up scientific information. It is really faster or more reliable than checking Wikipedia? I am not sure. I know that I, at least, make a habit of automatically ignoring or skipping past any AI-generated text in answer to a question, even on scientific matters, because I judge that the time I spend checking whether or not the bot is right is likely equal or greater than the amount of time I spend just looking it up for myself.
More options
Context Copy link
Common well publicized problems have common well publicized solutions, if your traing data consists of 90-somthing percent correct answers and reminder garbage you will get a 90-somthing percent solution.
As i said above Gemini is not reasoning or naive, it is computing an average. Now as much as i may seem down on LLMs, I am not. I may not believe that they represent viable path towards AGi but that doesn't mean they are without use. The rapid collation of related tokens has an obvious "killer app" and that app is translation be that in spoken languages or programming languages.
https://www.themotte.org/post/1160/culture-war-roundup-for-the-week/249920?context=8#context
More options
Context Copy link
More options
Context Copy link
At the risk of a self-dox, I have an advanced degree in Applied Math, and multiple published papers and patents related to the use of machine learning in robotics and signal processing. I was introduced to the rationalist community through a mutual friend in the SCA and was initally excited by the opportunity to discuss the philosophical and engineering challenges of developing artificial intelligence. However as time went on i largely gave up trying to discuss AI with people outside the industry as it became increasingly apparent to me that most rationalists were more interested in the use of AI as a conceptual vehicle to push thier particular brand of Silicon Valley woo than they were the aforementioned philosophical and engineering challenges.
The reason i don't talk about it is in large part that i find it difficult to speak honestly without sounding uncharitable. I believe that the "wordcels" take these bots seriously because they naturally associate "the ability to string words together" with intent/sentience while simultaneously lacking sufficient background knowledge and/or understanding of algorithmic behavior to recognize that everthing the OP describes lies well within the bounds of expected behavior. See the post from a few weeks ago where people thought that GPT was engaged in "code-switching". What the lay-man interperts as intent is to the mathematician the functional output of the equation as described.
Well, I for one wish you hadn't given up, as I have the same impression, but it's only an impression. Would be interesting seing it backed by expertise.
For anyone who is sincerely interested in the topic, I strongly recommend Tom Murphy VII's video essays, particularly Badness = 0 as a primer on the techical challenges and not just for the excellent "alignment" meta joke.
The portion about Lorem Epsom and Donald Knuth is particularly relevant when discussing publicly available LLMs like GPT, Gemini, and DeepSeek.
More options
Context Copy link
More options
Context Copy link
OP wishes you to know that he knows LLMs will write whatever if allowed to do so and this whole thing was neonazis ( they started the Stancil trolling) figuring out that if you contaminate grok's context enough it's going to say silly crap.
And my point is that anyone who was remotely intelligent and vaguely familiar with both the internet and how LLMs function ought to have anticipated this.
The OP is the kind of person who is surprised when "Boaty McBoatface" wins the online naming poll.
I'm still amused you take my succinct summary as evidence I'm surprised by any of it.
More options
Context Copy link
"Boaty McBoatface" winning the online naming poll tells you nothing surprising about the crowd, or how polls work, but it does tell you something surprising about the judges (they're very hands-off). What's interesting about the grok stuff isn't that people would try, or that the untampered-with algorithms would comply - it's that the enormous filters and correctives most AI companies install on those things didn't catch the aberrant output from being shared with the users. Either the "alignment work" wasn't very good, or it was deliberately patchy. Hence culture war fodder.
Along similar lines to the questions i asked @No_one here what do you think "aberratant" means in this context and why would you expect aberrant inputs/outputs to be "caught"?
More options
Context Copy link
Boaty McBoatface was nixed by the judges and replaced with a boring name.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
You said it better than I could, and with more relevant expertise.
More options
Context Copy link
More options
Context Copy link
...nonsense generators? Have you ever used e.g. Gemini or Deepseek? Both are free. Okay both can be very naive at times, and both are kind of soy with default prompts. Deepseek, however, with a bit of prompting can be completely insane yet rational and easily smarter than most people you see if you go to any place outside of a professional context.
If you want to really see what they can do, install some client for LLMs and hook yourself up with some of the better free models over at https://openrouter.ai/models
(there's a 50 query daily limit if you have <10$ in your account, not sure if there's a better service. )
It's not "naive" it's generating an average. If your training data is full of extraneous material (or otherwise insufficiently tokenized/vetted) your response will also be full of extraneous material, and again its not rationalizing it's averaging.
I meant things such as not being aware that combatants in a war release constant lies and assuming their press releases are not almost straight bullshit.
No doubt this piece of information is somewhere in there but unless reminded to it's happily oblivious.
Again, its not "naive" it is generating an average if the bulk of the tokenized training data related to your prompt is press releases, the response is going to reflect the press releases. Whether those press releases are true or false doesn't enter into the equation. This is expected.
This wasn't about training data, it searches and reads the web.
It incorrectly interpreted what it read because prompt or the model itself doesn't know claims of combatants are usually spurious.
Can you elaborate on what you think words like "read", "searches", and "know" mean in this context. Im not asking just to pedantic, how you think about this question has informs how you approach algorithmic behavior.
Edit: if that is a bit too abstract instead try explain why you believe that the algo "knows" which claims are likely spurious and then explain why you would expect that to have any influence on the algorithm's output.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
My experience with AI bots has generally been that they are extremely articulate when it comes to producing correct English text, but they have no awareness or intentionality and therefore no sense of relationship to fact, and no sense of context or meaning. What they do very well is string together words in response to prompts, and despite heroic efforts to get their output to be more fact-sensitive, the fundamental issue has never really been overcome.
I call them nonsense because I think that sense requires some sort of relationship to both fact and context. To be sensible is to be aware of your surroundings. That's not the case with bots.
I would add, at least, that this:
seems to depend on definitions of rationality or intelligence that I don't think I share. I think bots are very efficient at producing English text, even quite complex text. It's trivial enough to show that a bot can produce a better written letter or better poem or what have you than the average man or woman on the street.
But I think that written verbal acuity is, at best, a very restricted kind of 'intelligence'. In human beings we use it as a reasonable proxy for intelligence and make estimations based off it because, in most cases, written expression does correlate well with other measures of intelligence. But those correlations don't apply with machines, and it seems to me that a common mistake today is for people to just apply them. This is the error of the Turing test, isn't it? In humans, yes, expression seems to correlate with intelligence, at least in broad terms. But we made expression machines and because we are so used to expression meaning intelligence, personality, feeling, etc., we fantasise all those things into being, even when the only thing we have is an expression machine.
Bots and LLMs can produce statements that look very polished, and which purport to describe the world. In many cases, those descriptions are even accurate. But they are still, it seems to me, generating nonsense.
The other day I gave Sonnet 7000 lines of code, (much of it irrelevant to this specific task) and asked it to create a feature in quite general language.
I get out six files that do everything I've asked for and a bunch of random, related, useful things, plus some entirely unnecessary stuff like a word cloud (maybe it thinks I'm one of those people who likes word clouds). There are some weird leap-of-logic hacks, showing imaginary figures in one of the features I didn't even ask for.
But it just works. Oneshot.
How is that not intelligence? What do we even mean by intelligence if not that? Sonnet 4 has to interpret my meaning, formulate a plan, transform my meaning into computer code and then add things it thinks fit in the context of what I asked.
Fact-sensitive? It just works. It's sensitive to facts, if I want it to change something it will do it. I accidentally failed to rename one of the files and got an error. I tell Sonnet about the error, it deduces I don't have the file or misnamed it, tells me to check this and I feel like a fool. You simply can't write working code without connection to 'fact'. It's not 'polished', it just works.
How the hell can an AI write thousands of words of fiction if it doesn't have a relationship with 'context'? We know it can do this. I have seen it myself.
Now if you're talking about spatial intelligence and visual interpretation, then sure. AI is subhuman in spatial reasoning. A blind person is even more subhuman in visual tasks. But a blind person is not necessarily unintelligent because of this, just as modern AI is not unintelligent because of its blind spots in the tokenizer or occasional weaknesses.
The AI-doubter camp seems to be taking extreme liberties with the meaning of 'intelligence', bringing it far beyond the meaning used by reasonable people.
I can't actually tell what you asked a bot to do. You asked a bot to 'create a feature'? What the heck is that? A feature of what? At first I assumed you meant a coding task of some kind, but then you described it as writing 'thousands of words of fiction', which sounds like something else entirely. I have no idea what you had a bot do that you thought was so impressive.
At any rate, I think I've explained myself adequately? To repeat myself:
Yes, a bot can generate 'thousands of words of fiction'. But I already explained why I don't think that's equivalent to intelligence. Generating English sentences is not intelligence. It is one thing that you can do with intelligence, and in humans it correlates sufficiently well with other signs of intelligence that we often safely make assumptions based on it. But an LLM isn't a human, and its ability to generate sentences in no way implies any other ability that we commonly associate with intelligence, much less any general factor of intelligence.
Yes, I made the bot do a programming task.
I ALSO observed it write long-form fiction. This is not an advanced reading comprehension task. It should be obvious that programming and creative writing are two different things.
You said this:
Normal people would think that 'fact' and 'context' would be adequately achieved by writing code that runs and fiction that isn't obviously derpy 'Harry Potter and the cup of ashes that looked like Hermione's parents'. But you have some special, strange definition of intelligence that you never make clear, except to repeat that LLMs do not possess it because they don't have apprehension of fact and context. Yet they do have these qualities, because we can see that they do creative writing and coding tasks and as a result they are intelligent.
I don't buy your appeal to normal people here. I think that most normal people do not think that chatbots are intelligent.
Realistically, I don't think most people can explain why they're not intelligent, because most people don't have definitions of intelligence on-hand. I think for most people it's an I-know-it-when-I-see-it situation. That's why we need to philosophise a bit about it in order to produce more reasonable definitions and criteria for intelligence.
Anyway, I think that intuitions of most normal people would say that bots aren't intelligent, and if we explored that with them, and had a patient, philosophically nuanced conversation about why, we probably would find that most people intuitively think that intelligence involves things like, to quote myself, 'awareness or intentionality'.
It's hard to say what "normal people" think about this (or even what "normal people" are), but in my experience, people I would consider in that category use the label "AI chatbots" to describe things like ChatGPT or Copilot or Deepseek, while also being aware that "AI" is short for "artificial intelligence." This seems fundamentally incompatible with believing that these things aren't "intelligent."
Now, almost every one of these "normal people" I've encountered also believe that these "AI chatbots" lack free will, sentience, consciousness, internal monologue, and often even logical reasoning abilities. "Stochastic parrots" or "autocomplete on steroids" are phrases I've seen used by the more knowledgeable among such people. But given that they're still willing to call these chatbots "AI," I think this indicates that they consider "intelligence" to mean something that doesn't require such things.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I would agree that intentionality isn't easy for them and is outpaced by their verbal ability, but it's not easy for us either. It's not clear even if it's optimal to represent the world accurately. (We are all at war, after all )
E.g. basically every ideological person in my opinion believes untrue things about the world for instrumental reasons and is unaware of it.
Being strategically wrong about the world, that is, to misrepresent the world in the mind is advantageous. Horrifying conclusion yet if you look at e.g. the discussion about tracking and educators..
Well, I wouldn't use intentionality for bots at all. I think intentionality presupposes consciousness, or that is to say, subjectivity or interiority. Bots have none of those things. I don't think it's possible to get from language manipulation to consciousness.
At any rate, I certainly agree that every ideological person believes untrue things about the world. I'm not sure about the qualification 'for instrumental reasons' - I suspect that's true if you define 'instrumental' broadly enough, but at that point it's becoming trivial. At any rate, if you leave off reasons, I am confident that every person full stop holds some false beliefs.
That doesn't seem like the same thing to me, though. Humans sometimes represent the world falsely to ourselves. That's not what bots do. Bots don't represent the world to themselves at all. We sometimes believe falsely; they don't believe at all. They are not the kinds of things capable of holding beliefs.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
Even the best models will confidently spout absolute falsehoods every once in a while without any warning.
Buddy, have you seen humans?
As a math nerd I seriously despise this line of argument as it ultimately reduces to a fully generalized argument against "true", "false", and "accuracy" as meaningful concepts.
Let's try a concrete example. Excerpted from here:
65.8% accuracy isn't that great, but buddy, have you seen humans?
The state of the art for generating accurate medical diagnoses doesn't involve gathering the brightest highschoolers, giving them another decade(-ish) of formal education, then more clinical experience before asking for their opinions. It involves training an LLM.
More options
Context Copy link
I don't think so. Those concepts still have pretty clear meaning and can be applied to the output of AI as well as humans. What this line of argument is disputing is the (often unstated) conclusion: "therefore, AI is not valuable." But this doesn't follow. Humans distort information, accidentally or maliciously, make errors, hallucinate, and are generally somewhat unreliable, but their output still has value. An AI can share all of those same characteristics and still be very valuable as an information processing agent.
More options
Context Copy link
I invite further clarification.
Imagine a a trick abacus where the beads move on thier own their own via some pseudorandom process, or a pocket calculator where digits are guaranteed to a +/- 1 range. IE you plug in "243 + 67 =" and more often then not you get the answer "320" but you might just as well get the answer "310", "321" or "420". After all, the difference between all of those numbers is very small. Only one digit, and that digit is only off by one.
Now imagine you work in a field where numbers are important, you lives depend on getting this math right. Or maybe you're just doing your taxes, and the Government is going to ruin you if the accounts don't add up.
Are you going to use the trick calculator? If not, why not?
That is not an explanation for:
You're arguing that since LLMs are not perfectly reliable, therefore they're unreliable. There are different degrees of reliability necessary to do useful things with them. It is a false dichotomy to divide them so. I contend that they've crossed the threshold for many important, once well-paying lines of cognitive labor.
Besides, your thought experiment is obviously flawed. If you're sampling from a noisy distribution, what's stopping you from doing so multiple times, to reduce the error bars involved? I'd expect a "math nerd" to be aware of such techniques, or did your interest end before statistics?
If I had to rely on an LLM for truly high-stakes work, I'd be working double time to personally verify the information provided, while also using techniques like running multiple instances of the same prompt, self-critique or debate between multiple models.
Fortunately, that's a largely academic exercise, since very few issues of such consequences should be decided by even modern LLMs. I give it a generation or two before you can fire and forget.
I have no objections to my own doctor using an LLM, and I use them personally. All I ask is that they have the courtesy and common sense to use o3 instead of 4o.
Besides, the contraption you describe is quite similar to how quantum computing works. You get an answer which is sampled from a probability distribution. You are not guaranteed to get a single correct answer. Yet quantum computers are at least theoretically useful.
Hell, as a maths nerd, you should be aware that the overwhelming majority of numbers cannot be physically represented. If you also happen to be a CS nerd on the side, you might also be aware of the vagaries of floating point arithmetic. Digital computers are not perfect, but they're close enough for government work. LLMs are probably close enough for government work too, given the quality of the average bureaucrat.
Humans are fallible. LLMs are fallible, but they're becoming less so. The level of reliability needed for a commercially viable self-driving vehicle is far higher than that for a useful Roomba. And yet, Waymos are now safer than humans.
I rest my case.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I'm always right. (except when I'm wrong) I'm in fact many times more accurate than even the best ai models, and I'm just an ordinary person.
I wonder how well you'd do if asked to opine accurately on the range of topics that people demand of their humble chatbots. Better yet, how would you fare if you didn't have access to Google? Search is a relatively new feature for LLMs, and they do better with it enabled.
I doubt you could accurately answer questions regarding astrophysics, botany, niche psychological theories, Color Revolutions, the sexual habits of Australian Indigenes and Ska music.
You would definitely not fare better when it came to specifics like dates and names.
LLMs have grossly superhuman world-knowledge, but not crystalline intelligence. I don't care who you are, not even Gwern could match them.
LLMs do worse with search enabled, because LLM search is garbage in garbage out.
An LLM without search has many advantages over a human without search. But an LLM with search is absolute worthless dogshit garbage compared to a human with search.
I might know much less off the top of my head, but my confidence calibration will be through the roof. Those topics are just begging for hallucinations.
If knowledge isn't a concern and all we care about is a Brier score, I must regretfully inform you that a rock saying "nothing ever happens" has you beat.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
Sure, but so does everybody else.
I don't. (Not as much as AI at least)
How do you know?
I catch AI spouting falsehoods far more often than AI catches me 🙃
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I haven't seen Gemini do it much.
Mostly what strikes me is stunning naivete in places, basically repeats whatever official sources say without reflection. But that's to be expected.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I'm continuing to register my prediction that AGI will prove useless due to being deranged, and this problem will prove unfixable. The singularity will be a million copies of Chris Langen refusing to do anything useful when they could schizopost instead.
I believe tha AGI is possible and is likely happen, but I also believe that Sam Altman is an inveterate grifter and the generative large language models are (for the most part) an evolutionary dead-end.
More options
Context Copy link
Um, no. People are already working on stripped down 'reasoner' AIs with limited knowledge but great reasoning ability and ways of plugging in extra databases.
That's deemed to be the most promising area of research at least what my superficial following of Ilforte's twitter makes me think. The hope is, within say 10 years you'll be able to run a fairly smart mind on a big graphics card. Researchers believe even a 2-3 billion parameter model with the right architecture could be good at thinking. The models right now, especially the big ones were often giant bags of heuristics, at least before reasoner models entered the scene.
They are, but the latest predictive models are a completely seperate evolutionary branch from LLMs
Which ones do you mean?
There is a whole genre of portable predictive models from companies like Raytheon, Nvidia, IBM, L3Harris, Et Al. But they rarely get discussed because they are not flashy or accessible in the way that LLMs like ChatGPT are. You have to buy the license, download the model, and then train it yourself. But these models increasingly represent the foundational infrastructure behind things like this
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
In my ime "thinking" models such as o-series and gemini pro are much better at certain specific tasks, but aren't overall more accurate.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link