self_made_human
Grippy socks, grippy box
I'm a transhumanist doctor. In a better world, I wouldn't need to add that as a qualifier to plain old "doctor". It would be taken as granted for someone in the profession of saving lives.
At any rate, I intend to live forever or die trying. See you at Heat Death!
Friends:
A friend to everyone is a friend to no one.
User ID: 454
Regardless of whether or not it's unkind to say the truth, it isn't up for debate that there are massive differences on average between the kind of child OP could have (if not infertile) and the kind up for adoption.
In that case, people who would have been sysadmins are either paid to become brick layers or are forced to do it because that's the only job left.
There's a reason you rarely see Asian-Americans working low end jobs in the US, while those positions are filled back in their native countries. A society of Einsteins will have a need for janitors, until they automate the solution away. It is still better to be such a society with such a population.
COD and Insurgency, while not quite as "Number Shooter" as say, Destiny, are still not that far.
You have small maps, predictable spawns and player behavior, and a constrained set of weapons. Said weapons can be somewhat easily modeled with DPS being the only really relevant property.
I play a lot of Arma, and you're not going to be able to do that. The bigger the playing field, the wider the space of strategies, the more simulationist the modeling..
You will have a hard time min-maxing Arma for the same reason nobody has solved IRL war, despite the obvious extreme optimization pressure.
DOTA2
This strategy guide for DOTA2 players is the best around. Hope that helps!
One of the reasons new multiplayer games are a lot more fun to play than old ones is that for the first few weeks after a game is released, or while it’s in beta, the nasty people, the min-maxers, the forum theorycrafters, have yet to ruin everything by Excel spreadsheeting statistical models of damage and critical chance and elemental resistance until they derive, mechanically, the ‘most efficient’ build, after which everyone adopts the new meta, increasingly of course because even the developers now design to it (see World of Warcraft’s designers building raids with the expectation that players will play the most meta builds, with all the most advantageous mods/addons). Why bother experimenting, playing, using your own intelligence when someone else who gamed the system with the ‘meta’ will curbstomp you for 1/10th the effort.
That's an artifact of playing Number Shooters (where enemies are transparently walking sacks of hitpoints you're trying to subtract), or Number RPGs/looter shooters where you're just trying to make Number Go Up. Everything can be easily reduced to metrics like DPS, to the detriment of having a game at all.
I'm grateful that I prefer my games to operate in a manner that obfuscates the fact that it's all 1s and 0s on a storage drive, and which remain fun even if you're not playing them like you're a glorified SAT-solver.
If you are a doctor and want your children to be doctors (an ancient professional right, just as the son of a blacksmith might become one), you will probably have to work them to the bone
If I want my (hypothetical) kids to be doctors, then I'd need to quite a bit of faith that there are Amish communities running around in 2050. I really don't see how it's feasible to be entering that profession otherwise, that's just not the way things are going.
One might argue we're already over Peak Doctor, we just don't know it yet. I certainly wouldn't want to even be just a bright-eyed student entering med school in the Year of Someone's Lord 2025.
At the end of the day, I'm strongly of the opinion that there's no point in worrying about the state of education if you're someone who has only young kids or no kids at all. Formal education as we know it will very likely not exist by the time they'd be old enough for it, and if it does, it'll likely just be entirely signaling as opposed to 75% signaling.
As others have already been kind enough to point out, the mod team asks that top-level posts within the CWR thread have more to them than a bare-link. Ideally, with more substantive commentary or an explanation of why this worth time and attention. At the absolute bare minimum, quotes of the source material and your thoughts on the matter where relevant.
Just wanted a tattoo because it felt cool lol. My best friend came up with a design I really liked, and on my deltoids it went. You could also consider it a getting into med school/becoming an adult gift to myself.
Error rates have fallen drastically, and I'm someone who has regularly benefited from context windows becoming OOMs larger than the best 2023 had to offer.
I know specific questions, in programming and maths most obviously, but also in medicine, where I wouldn't trust the output of a 2023 model, but where I'd be rather confident in a 2025 one being correct.
Reasoning models are also far better at task adherence and thinking logically. Agents are still less than ideal today, but they were borderline useless in 2023.
Other very nice QOL features include image and file input and generation, artifacts, voice conversations etc. If I had to go back to a 2023 GPT-4, I'd be pissed.
-
Commoditizing their complements. This is particularly true of Meta, which wanted to use Llama to undercut competitors like OpenAI. Meta doesn't need their models to be profitable, that's not the core of their company. But OAI? Without people willing to pay for access to their models (or if they're able to clone them cheaper and run them elsewhere), they'd be utterly screwed.
-
Gaining market and mind share, and either finding other ways to monetize (consulting or fine-tuning services), going paid and closed-source, or using the hype and investor confidence to raise money.
-
Attracting talented researchers, who often want recognition and the right to publish their research instead of having it all be locked down internal IP.
-
Actual ideological commitment. Or at least lip-service to help with points 2 and 3.
Uh, I haven't specifically been keeping track of most suggestions I'm afraid. I tried to go through my chat history for specific examples, but came up short since it doesn't save conversations more than a week or two old. It did note some flaws that I personally agree with, such as a predilection towards run-on sentences or instances where I'm being unclear. Most of the time, I would have run across and then fixed the flaws myself, but this approach saves me a lot of time. Unlike most authors, I spend far less time editing than writing by default. I should probably be doing more of that, and the LLMs help.
I think I get the most utility when I ask the model to rewrite whole essays for clarity, or to see how some other author would have approached that. This occasionally produced novel (to me) insights, or particular turns of phrase I might feel tempted to steal.
Or American. It's not an ethnicity, and even Native Americans can be ambiguous.
I can't think of a single use case where Gemini 2.5 Pro isn't superior to Kimi (it says plenty about the model that I have to compare it to SOTA), including cost. Google is handing away access for free, even on the API. It's nigh impossible to hit usage limits while using Gemini CLI.
A God-tier shitpost I am memetically compelled to spread due to the worms in my brain:
I'm not Dase, alas, but I want to say that I was profoundly surprised that Diffusion as a technique even works at all for text generation, at least text that maintains long-term coherence. I'm utterly bamboozled.
Excellent work as usual Dase. I was sorely tempted to write a K2 post, but I knew you could do it better.
challenges the strongest Western models, including reasoners, on some unexpected soft metrics, such as topping EQ-bench and creative writing evals (corroborated here)
I haven't asked it to write something entirely novel, but I have my own shoddy vibes-benchmark. It usually involves taking a chapter from my novel and asking it to imagine it in a style from a different author I like. It's good, but Gemini 2.5 Pro is better at that targeted task, and I've done this dozens of times.
Its writing is terse, dense, virtually devoid of sycophancy and recognizable LLM slop.
Alas, it is fond of the ol' em-dash, but which model isn't. I agree that sycophancy is minimal, and in my opinion, the model is deeply cynical in a manner not seen in any other. I'd almost say it's Russian in outlook. I would have bet money on "this is a model Dase will like".
Meta's AI failure are past comical, and into farce. I've heard that they tried to buy-out Thinking Machines and SSI for billions, but were turned down. Murati is a questionable founder, but I suppose if any stealth startup can speed away underwater towards ASI, it's going to be one run by Ilya. Even then, I'd bet against it succeeding.
I don't know if it's intentional, but it's possible that Zuck's profligity and willingness to throw around megabucks will starve competitors of talent, but I doubt the kind of researcher and engineers at DS or Moonshot would have been a priori deemed worthy.
I don't think anyone nominated me for a UVP, so I haven't had the opportunity. I probably would nominate you if it came up.
(Maybe you're thinking about the doge contest)
Thank you. I will clarify that by RL, I don't mean bog-standard RLHF, but more recent techniques like RLVR that have been around since o1.
yes yes another post about AI, sorry about that
Feel that AGI baby!
It's obvious what the trends are. I predict that, on the midnight before ASI, the Motte's going to be 124% AI commentary. It might even be AI doing the commentary.
It's a primarily agentic non-reasoner
I have read claims that it's a pseudo-reasoner, and it was trained on COT traces and had RL done even if it doesn't use explicit reasoning tokens itself. I've also heard that it's 3x as verbose as most NRLLMs, almost on par with RLMMs, making the distinction academic. This was on Twitter, and I don't have links handy. I'm not sure how strongly to index on that.
A lot of the grognards over on HN don't think it counts, but they're the type who wouldn't accept blowjobs in heaven if the angels weren't Apache licensed.
Other than reach and better animation, I don't think this is different from the AI companions that have been available for a while. Replika, the most famous one, will already do NSFW ERP. And yeah, there are men (and women!) who have decided their Replikas are preferable to real people.
That fact that it's animated is a big deal! Men are visual creatures, and the fact that previous ERP was textual made it far less appealing to the average dude, if not woman. Of course, jerking off to anime tiddies is still not a preference of the majority, but it's easy to upgrade to photorealism. That'll get more people.
I predicted this outcome ages ago, though I'd have said it was inevitable and obvious to anyone who cared. It's priced in for me, and I agree that it likely won't be catastrophic.
I don't doubt that, but once again, that doesn't mean that the vast majority of people are receiving any actual attention from the CIA.
My apologies. I was thinking of this related thread, and it's not you I was arguing with.
(Some might even call the mistake I made a hallucination, hmm)
So, some observations. First, sorry dude, but I have major side-eye for your ability to evaluate literary quality. :p
You hit below the belt. Reverend Insanity is Peak Fiction and I'm going to go down swinging!
As you probably know, even the most powerful LLMs do not have a context window large enough to store an entire large novel in memory, let alone a series, and you can't directly upload embeddings to GPT or Claude
1 million tokens is a lot! (Gemini 2.0 had 2 million, but good luck getting it to function properly when it's that full). That is 750k words. All of Harry Potter is just over a million.
I'm going to ignore Llama here, since even if it has a max 10 million token CW, mental retardation is not improved by the fact that there's a lot more of it. And why shouldn't I? Even Zuck has chosen to forget that particular failure.
I've uploaded whole medical textbooks into them without major issue. Not tiny books either.
As long as you can keep it on track, I have found that some of the GPT and Anthropic models are... not terrible as beta readers. They point out some real flaws and in a very generic sense have an "understanding" of pacing and tone and where a scene is missing something.
I am most personally familiar with uploading chapters (often half a dozen) of my own work, which works well. If I was less lazy, I'd probably be saving summaries of the whole thing and stringing them together. (Royal Road makes it so you can't export an epub of your own fic without paying, and without that option, I'd be doing a lot of copying and pasting)
When asked for critique, some of the issues raised were cogent. Too much jargon, uneven pacing and so on.
Some of that was intentional, such as the fact that since the excerpts were lifted from a larger work, most of the jargon was previously explained at one point or the other. I also have no shame about making potential readers resort to keeping a Wikipedia tab open on the side, it's niche hard scifi and I want to flex. Other issues are well worth amending before publication.
I haven't had the good fortune of having very many professional authors or editors review and critique, and I don't doubt that they'd probably give me even more useful feedback. Yet what I get is quite good and elevates the final product!
I still think we'll need true AGI to write an actual good novel. When you show me an AI that can write a coherent series, with multi-volume character arcs, plot seeds planted in early books that clearly pay off in later ones, literary allusions and metaphors that aren't just clumsy pulled-off-the-shelf ones but deeply enmeshed in the story, and a recognizable differentiable style (in the same way that fans can read Dickens or McCarthy or Hemingway and immediately recognize the author), I will believe we're there.
That aligns well with my own stance. A large novel is an unwieldy thing, let alone a good one. We're still at the competent novella or subpar novel stage, but I must stress that's a comparison against the very few human authors who make big bucks and/or accrue critical acclaim. Most things humans or LLM novelists write are slop, the former just don't scale as hard.
You're correct in that perfect recall or retention isn't feasible when using a large number of tokens (in my experience, performance degrades noticeably over 150k). When I threw in textbooks, it was for the purpose of having it ask me questions to check my comprehension, or creating flashcards. The models have an excellent amount of existing medical knowledge, the books (or my notes) just help ground it to what's relevant to me. I never needed perfect recall!
(Needle in a haystack tests or benchmarks are pretty awful, they're not a good metric for the use cases we have in mind)
Ah.. So that's how people were making epubs with ease. Thank you for the tip!
I don't think it's got much to do with copyright, it's probably just such a rare use case that the engineers haven't gotten around to implementing it. Gemini doesn't support either doc or docx, and those would probably be much more common in a consumer product. I don't recall off the top of my head if ChatGPT or Claude supports epubs either.
More options
Context Copy link