site banner
Advanced search parameters (with examples): "author:quadnarca", "domain:reddit.com", "over18:true"

Showing 25 of 111976 results for

domain:amphobian.info

Thank you. I will clarify that by RL, I don't mean bog-standard RLHF, but more recent techniques that have been around since o1.

I'm just waiting for DeepSeek R2. Not happy with the delay and while R1-05-28 is pretty damn good it isn't at the very top. K2 is non-thinking which means that while an excellent base model it isn't the best of the best when quality rather than speed matters.

It's not really verbose in normal use, rather the opposite. It is verbose in agentic mode, writing docs and commentary rather than thinking traces. RL has been used for all serious LLMs since GPT 3-instruct, this is independent of the current long-CoT paradigm. It is dubious that Kimi has been trained on CoTs because it doesn't do them. More likely, its training data is largely final outputs of a reasoner (like Kimi's own 1.5/1.6). They have a section in the paper on 1.5 about penalizing verbosity.

Great post, you’re one of my favorite commenters here. This makes me wonder if I ever did the User Viewpoint series. I think I did (maybe @self_made_human nominated me), but I can’t remember.

The NYT’s house style was deliberately constructed in opposition to the British style of quoting everything.

NYT: Earthquake kills thousands

[London] Times: Earthquake “Kills Thousands”.

yes yes another post about AI, sorry about that

Feel that AGI baby!

It's obvious what the trends are. I predict that, on the midnight before ASI, the Motte's going to be 124% AI commentary. It might even be AI doing the commentary.

It's a primarily agentic non-reasoner

I have read claims that it's a pseudo-reasoner, and it was trained on COT traces and had RL done even if it doesn't use explicit reasoning tokens itself. I've also heard that it's 3x as verbose as most NRLLMs, almost on par with RLMMs, making the distinction academic. This was on Twitter, and I don't have links handy. I'm not sure how strongly to index on that.

A lot of the grognards over on HN don't think it counts, but they're the type who wouldn't accept blowjobs in heaven if it wasn't using a standard Apache license.

Hahaha yes, YES! Finally I can dust off my SICKOS shirt. Your move, Anthropic.

Factually, I concur with posters below that actually this isn't markedly different, and in some ways more basic, than the already possible setups for AI gooning (if the system prompt posted on X is real). Not a big fan of the voice either tbqh, although the model rigging seems good, and manic pixie Amane-Misa-at-home is a more natural fit for an AI waifu at a casual glance than whatever eldritch horrors Replika has. I'll likewise point out that while this incarnation is obviously aimed at men, the potential for female appeal is much bigger, and is in fact only a model/voice swap and a system prompt away. Not sure who is actually going to pay $300/mo for it however, the porn equivalent is literally free, and an equivalent ai gf via OR or something is much cheaper if a hassle to set up. Normies hate trivial inconveniences almost as much as steep price tags, I don't think this is how you get buy-in from them, but I assume this is literally the first step and we'll see further developments.

Regardless of technical chops, the real value here is of course exposure, a first decent shot at normiefying the whole edifice. Elon may be a fake gamer, the gravest insult I can levy against my fellow man, but fringe interests make for strange bedfellows, and I'm glad to see the first public attempts at rather literal waifutech make the twitterati seethe. I know what I'm following for the next few days. Though I agree with the OP that the whole Mecha-Hitler brouhaha was 100% a publicity psyop in hindsight, the timing and subject matter is too conspicuous. Based?

On another note, I think that people invoking the Torment Nexus are kind of missing the point. I don't think "authentic" humans are threatened (NB: screeching xitter denizens not considered human authentic). Even the most literal goonette femcel I know consistently finds much more joy in talking/RPing with an actual human over a chatbot, by a rather wide margin, even if the chatbot wins out on 24/7 availability.

Instead, I think the real horror potential here is - may Allah forgive me for uttering this word - gacha games, or more broadly chatbot integration and monetization. I've recently gotten into Girls Frontline 2, and seeing the manic pixie grok gf clicked something together in my head. I can already see the framework, the jewish tricks are practically manifesting before my eyes: gacha girls have affinity/bond levels (here increased by gifts = in-game loot), a certain level of bond unlocks the chatbot functionality with the given waifu, further levels relax the guardrails or change the system prompt (reflecting increased affection)... you get the gist. My cai/Chub experience tells me gacha girls are some of the most popular interlocutors anyway, so the match is eminently natural.

From there the potential for added deviltry is almost limitless:

  • obviously, 3d model visibly reacting to prompts like not-Amane-Misa here
  • outfits for said 3d model, those are already a gacha staple but maybe some kind of added animations or different "backgrounds"/environments for the chatbot part? (i.e some hypothetical maid costume, with some abstract cafe setting written into the chatbot's system prompt if the costume is on?)
  • limit the amount of prompts per day (vantablack ninth circle hell tier: offer paid refreshes)
  • lock explicit NSFW behind a paid item (e.g. GFL2 has a marriage Covenant mechanic, the ring is single-use and costs $5)
  • give the waifus some kind of actual stat boosts for "cheering them up" before gameplay, grading incoming user prompts to this end like Grok seemingly does (I eagerly await the inevitable rizz meta guides on how to best word your prompts for the optimal boost)
  • some kind of voice command integration built on top as an afterthought? GFL2 is turn-based xcom-lite so I imagine voice commands can work given an autistic enough framework under the hood

Granted I sense the danger of metaphysical cuckoldry Chevrolet-tier oopsies and general bad press here, a man in pursuit of his coom is endlessly inventive, but as long as the chatbot is sufficiently insulated on the backend (also just imposing harsh character limits already neuters most prompt engineering) I think this can work. Though it probably won't be a Chinese gacha given the dangerously free-form nature of chatbots, and I don't think anyone else is deranged bold enough to try.

Other than reach and better animation, I don't think this is different from the AI companions that have been available for a while. Replika, the most famous one, will already do NSFW ERP. And yeah, there are men (and women!) who have decided their Replikas are preferable to real people.

That fact that it's animated is a big deal! Men are visual creatures, and the fact that previous ERP was textual made it far less appealing to the average dude, if not woman. Of course, jerking off to anime tiddies is still not a preference of the majority, but it's easy to upgrade to photorealism. That'll get more people.

I predicted this outcome ages ago, though I'd have said it was inevitable and obvious to anyone who cared. It's priced in for me, and I agree that it likely won't be catastrophic.

I don't doubt that, but once again, that doesn't mean that the vast majority of people are receiving any actual attention from the CIA.

Zooming out a bit, the UK has currently similar problems (exploding general healthcare/welfare costs leading to low-priority care being sidelined) despite an establishment politician winning, germany has similar problems despite an establishment politician winning, but somehow in the US it's Trump's fault. Even the specifics - rural hospitals getting disproportionally closed - are the same. Remember, healthcare cost graphs over time look like this, and this is % of GDP, so it's even more crazy in absolute terms. That growth is not sustainable, and indeed is starting to not getting sustained much longer. That doesn't mean Tump's cuts aren't higher than elsewhere nor that they have no negative effects, but the framing here is quite questionable.

My apologies. I was thinking of this related thread, and it's not you I was arguing with.

(Some might even call the mistake I made a hallucination, hmm)

The sats observe everything and images are kept forever. You are just picked out and individually tracked by software after the fact.

(I don't know it for a fact; it's just my best guess from extrapolating from leaks. I'd bet money on it if there were ever some way to get at the ground truth.)

So, some observations. First, sorry dude, but I have major side-eye for your ability to evaluate literary quality. :p

You hit below the belt. Reverend Insanity is Peak Fiction and I'm going to go down swinging!

As you probably know, even the most powerful LLMs do not have a context window large enough to store an entire large novel in memory, let alone a series, and you can't directly upload embeddings to GPT or Claude

1 million tokens is a lot! (Gemini 2.0 had 2 million, but good luck getting it to function properly when it's that full). That is 750k words. All of Harry Potter is just over a million.

I'm going to ignore Llama here, since even if it has a max 10 million token CW, mental retardation is not improved by the fact that there's a lot more of it. And why shouldn't I? Even Zuck has chosen to forget that particular failure.

I've uploaded whole medical textbooks into them without major issue. Not tiny books either.

As long as you can keep it on track, I have found that some of the GPT and Anthropic models are... not terrible as beta readers. They point out some real flaws and in a very generic sense have an "understanding" of pacing and tone and where a scene is missing something.

I am most personally familiar with uploading chapters (often half a dozen) of my own work, which works well. If I was less lazy, I'd probably be saving summaries of the whole thing and stringing them together. (Royal Road makes it so you can't export an epub of your own fic without paying, and without that option, I'd be doing a lot of copying and pasting)

When asked for critique, some of the issues raised were cogent. Too much jargon, uneven pacing and so on.

Some of that was intentional, such as the fact that since the excerpts were lifted from a larger work, most of the jargon was previously explained at one point or the other. I also have no shame about making potential readers resort to keeping a Wikipedia tab open on the side, it's niche hard scifi and I want to flex. Other issues are well worth amending before publication.

I haven't had the good fortune of having very many professional authors or editors review and critique, and I don't doubt that they'd probably give me even more useful feedback. Yet what I get is quite good and elevates the final product!

I still think we'll need true AGI to write an actual good novel. When you show me an AI that can write a coherent series, with multi-volume character arcs, plot seeds planted in early books that clearly pay off in later ones, literary allusions and metaphors that aren't just clumsy pulled-off-the-shelf ones but deeply enmeshed in the story, and a recognizable differentiable style (in the same way that fans can read Dickens or McCarthy or Hemingway and immediately recognize the author), I will believe we're there.

That aligns well with my own stance. A large novel is an unwieldy thing, let alone a good one. We're still at the competent novella or subpar novel stage, but I must stress that's a comparison against the very few human authors who make big bucks and/or accrue critical acclaim. Most things humans or LLM novelists write are slop, the former just don't scale as hard.

Relevant from Lambert: The American DeepSeek Project

While America has the best AI models in Gemini, Claude, o3, etc. and the best infrastructure with Nvidia it’s rapidly losing its influence over the future directions of AI that unfold in the open-source and academic communities. Chinese organizations are releasing the most notable open models and datasets across all modalities, from text to robotics or video, and at the same time it’s common for researchers worldwide to read far more new research papers from Chinese organizations rather than their Western counterparts.

This balance of power has been shifting rapidly in the last 12 months and reflects shifting, structural advantages that Chinese companies have with open-source AI — China has more AI researchers, data1, and an open-source default.

[…] The goal for my next few years of work is what I’m calling The American DeepSeek Project — a fully open-source model at the scale and performance of current (publicly available) frontier models, within 2 years.2 A fully open model, as opposed to just an “open weights” model, comes with data, training code, logs, and decision making — on top of the weights to run inference — in order to distribute the knowledge and access for how to train AI models fully.

etc. He overstates the cause, perhaps. America doesn't need these egghead communist values of openness and reproducibility, the free market will align incentives, and charity shouldn't go too far. But he's pointing to the very real fact that China, and not on the state level but on the level of individual small companies with suitable culture, is the only country bringing transformative AI not locked on corporate clusters closer to reality.

I don't think the Birkenhead drill only applies if the women in question aren't barren. Of course the value bestowed upon women is ultimately an evolutionary adaptation to the reality that only women can bear children. But in practice, even barren women are still seen as Wonderful™ in a way that NEET men aren't.

"Watching here doesn't mean something so casual as the fact that there's a sat that Incidentally oversees my geographical location from gestationary orbit.

Us psychiatrists might be nerdy and out of date, but we're not that far gone, and this would be discussed before committing someone"

The fact it can, in "some cases" be true, makes it a non-bizarre delusion. The quote specifically says "extraordinarily unlikely", and I'd probably take some international arms dealer who told me so more seriously.

Also, believing that does not block "CIA is watching me 24 hours a day by satellite" anyway.

If I am not captured by CIA visual/radio/etc monitoring continuously, with breaks between captures measured in minutes not days or hours (over a typical day) then I would be deeply surprised.

Watching here doesn't mean something so casual as the fact that there's a sat that Incidentally oversees my geographical location from gestationary orbit.

Us psychiatrists might be nerdy and out of date, but we're not that far gone, and this would be discussed before committing someone.

I was worried when I saw that it had a custom license, but it turns out it's just a slightly modified MIT license, requiring credit if your project gets big enough. This truly is Open-Source.

Periodic Open-Source AI Update: Kimi K2 and China's Cultural Shift

(yes yes another post about AI, sorry about that). Link above is to the standalone thread, to not clutter this one.

Two days ago a small Chinese startup Moonshot AI has released weights of the base and instruct versions of Kimi K2, the first open (and probably closed too) Chinese LLM to clearly surpass DeepSeek's efforts. It's roughly comparable to Claude Sonnet 4 without thinking (pay no mind to the horde of reasoners at the top of the leaderboard, this is a cheap-ish capability extension and doesn't convey the experience, though is relevant to utility). It's a primarily agentic non-reasoner, somehow exceptionally good at creative writing, and offers a distinct "slop-free", disagreeable but pretty fun conversation, with the downside of hallucinations. It adopts DeepSeek-V3’s architecture wholesale (literally "modeling_deepseek.DeepseekV3ForCausalLM"), with a number of tricks gets maybe 2-3 times as much effective compute out of the same allowance of GPU-hours, and the rest we don't know yet because they've just finished a six-months marathon and don't have a tech report.

I posit that this follows a cultural shift in China’s AI ecosystem that I've been chronicling for a while, and provides a nice illustration by contrast. Moonshot and DeepSeek were founded at the same time, have near-identical scale and resources but have been built on different visions. DeepSeek’s Liang Wengeng (hedge fund CEO with Masters in engineering, idealist, open-source advocate) couldn't procure funding in the Chinese VC world with his inane pitch of “long-termist AGI research driven by curiosity” or whatever. Moonshot’s Yang Zhilin (Carnegie Mellon Ph,D, serial entrepreneur, pragmatist) succeeded at that task, got to peak $3,3 valuation with the help of Alibaba and Sequoia, and was heavily spending on ads and traffic acquisition throughout 2024, building a nucleus of another super-app with chatbot companions, assistants and such trivialities at a comfortable pace. However, DeepSeek R1, on merit of vastly stronger model, has been a breakout success and redefined Chinese AI scene, making people question the point of startups like Kimi. Post-R1, Zhilin pivoted hard to prioritize R&D spending and core model quality over apps, adopting open weights as a forcing function for basic progress. This seems to have inspired the technical staff: "Only regret: we weren’t the ones who walked [DeepSeek’s] path."

Other Chinese labs (Qwen, Minimax, Tencent, etc.) now also emulate this open, capability-focused strategy. Meanwhile, Western open-source efforts are even more disappointing than last year – Meta’s LLaMA 4 failed, OpenAI’s model is delayed again, and only Google/Mistral release sporadically, with no promises of competitive results.

This validates my [deleted] prediction: DeepSeek wasn’t an outlier but the first swallow and catalyst of China’s transition from fast-following to open innovation. I think Liang’s vision – "After hardcore innovators make a name, groupthink will change" – is unfolding, and this is a nice point to take stock of the situation.

Seconding cjet - please let me know when you finish that project (or if you want beta testing) that sounds like a fascinating and useful tool. I actually don't have that big of a problem with hallucinating these days, at least when I'm using models with live search like, well all of them except deepseek.

I have them set up with a custom prompt that basically tells them the date and their model name (because just the date leads to situations like where grok starts losing its shit because it doesn't know anything from the past two years), that they have access to Web search and python interpreter or any other tool I want to use, and then tell it to back up any facts it mentions with sources. That wouldn't help with your plot points problem though. That reminds me of the old Wikipedia though - it would work if we had that - back when every episode of transformers and pokemon and magnum pi was laid out point by point. Now I'm sad.

I can't help with getting ai to judge the quality of writing, though I do have advice for avoiding obsequiousness. Make a system prompt telling it to be like your Tiger mom who is very harsh and critical because she loves you and knows you can do better than mediocrity, and you feel like you need that push. It works best if you do it narratively, or like you are asking a friend for help. It doesn't work all the time, but it works better than 'give constructive criticism' because it gives the ai a new narrative to focus on and provides a reason to be critical that aligns with its built in desire to help. I'm not sure how much help it would be with fiction writing though. And you have to pick one or the other, I can't get those prompts to work together, I think because the jump from narrative style to instructions messes with their brains.

Reading back, I basically went the scenic route to I can't help. But remember me when you finish that application!

That's when we feel like we have dignity: when we can control how other people see us.

Which strikes me as an intrinsically quixotic goal. As you note yourself, even the richest man on earth can't stop people making jokes about his drug problems. Even the leader of the free world can't stop people making jokes about his tiny hands, as much as he'd obviously like to. When I see trans women in floods of tears and rending their hair about how strangers don't see them the way they see themselves, all I can think is - buddy, join the club.

Any single one of these things could potentially be something that can work. But to bring 5+ new special snowflakes out in a single article is out there. It defeats the purpose of lore if less than half of the stuff in a given battle has been seen before.

let's count the special snowflakes that are present in this otherwise mundane-ish battle:

  1. Experimental drone carrier
  2. Terraforming engine
  3. Grav sails
  4. Relic vaults
  5. Protocol Shas’kaara
  6. plasmacrete
  7. Perdita Grade
  8. etc.

Perdita-grade's meaning is obvious, it could be an ornate high gothic term for Forbidden World.

But it's out of character for this article, which logically should just call it a forbidden world. Anyways it's inconsistent with the lore because in the multitude of times forbidden worlds have been described in the lore they have never once used the term Perdita-grade

The drone carrier was to provide cover with drones, as described.

It's still incorrect to say that the cruisers are escorted by the carrier.

“The CIA is watching me 24 hours a day by satellite surveillance.”

Buddy, file this under extraordinarily likely. Unless you think they're filling up their Utah data center with cat videos.