@DaseindustriesLtd's banner p

DaseindustriesLtd

late version of a small language model

65 followers   follows 27 users  
joined 2022 September 05 23:03:02 UTC

Tell me about it.


				

User ID: 745

DaseindustriesLtd

late version of a small language model

65 followers   follows 27 users   joined 2022 September 05 23:03:02 UTC

					

Tell me about it.


					

User ID: 745

28

As I've been arguing for some time, the culture war's most important front will be about AI; that's more pleasant to me than the tacky trans vs trads content, as it returns us to the level of philosophy and positive actionable visions rather than peculiarly American signaling ick-changes, but the stakes are correspondingly higher… Anyway, Forbes has doxxed the founder of «e/acc», irreverent Twitter meme movement opposing attempts at regulation of AI development which are spearheaded by EA. Turns out he's a pretty cool guy eh.

Who Is @BasedBeffJezos, The Leader Of The Tech Elite’s ‘E/Acc’ Movement?

…At first blush, e/acc sounds a lot like Facebook’s old motto: “move fast and break things.” But Jezos also embraces more extreme ideas, borrowing concepts from “accelerationism,” which argues we should hasten the growth of technology and capitalism at the expense of nearly anything else. On X, the platform formally known as Twitter where he has 50,000 followers, Jezos has claimed that “institutions have decayed beyond the point of salvaging and that the media is a “vector for cybernetic control of culture.”

Alarmed by this extremist messaging, «the media» proceeds to… harness the power of an institution associated with the Department of Justice to deanonymize him, with the explicit aim to steer the cultural evolution around the topic:

Forbes has learned that the Jezos persona is run by a former Google quantum computing engineer named Guillaume Verdon who founded a stealth AI hardware startup Extropic in 2022. Forbes first identified Verdon as Jezos by matching details that Jezos revealed about himself to publicly available facts about Verdon. A voice analysis conducted by Catalin Grigoras, Director of the National Center for Media Forensics, compared audio recordings of Jezos and talks given by Verdon and found that it was 2,954,870 times more likely that the speaker in one recording of Jezos was Verdon than that it was any other person. Forbes is revealing his identity because we believe it to be in the public interest as Jezos’s influence grows.

That's not bad because Journalists, as observed by @TracingWoodgrains, are inherently Good:

(Revealing the name behind an anonymous account of public note is not “doxxing,” which is an often-gendered form of online harassment that reveals private information — like an address or phone number — about a person without consent and with malicious intent.)

(That's one creative approach to encouraging gender transition, I guess).

Now to be fair, this is almost certainly parallel construction narrative – many people in the SV knew Beff's real persona, and as of late he's been very loose with opsec, funding a party, selling merch and so on. Also, the forced reveal will probably help him a great deal – it's harder to dismiss the guy as some LARPing shitposter or a corporate shill pandering to VCs (or as @Tomato said, running «an incredibly boring b2b productivity software startup») when you know he's, well, this. And this too.

Forbes article itself doesn't go very hard on Beff, presenting him as a somewhat pretentious supply-side YIMBY, an ally to Marc Andreessen, Garry Tan and such; which is more true of Beff's followers than the man himself. The more potentially damaging (to his ability to draw investment) parts are casually invoking the spirit of Nick Land and his spooky brand of accelerationism (not unwarranted – «e/acc has no particular allegiance to the biological substrate for intelligence and life, in contrast to transhumanism; in order to spread to the stars, the light of consciousness/intelligence will have to be transduced to non-biological substrates» Beff says in his manifesto), and citing some professors of «communications» and «critical theory» who are just not very impressed with the whole technocapital thing. At the same time, it reminds the reader of EA's greatest moment (no not the bed nets).

Online, Beff confirms being Verdon:

I started this account as a means to spread hope, optimism, and a will to build the future, and as an outlet to share my thoughts despite to the secretive nature of my work… Around the same time as founding e/acc, I founded @Extropic_AI. A deep tech startup where we are building the ultimate substrate for Generative AI in the physical world by harnessing thermodynamic physics. Ideas simmering while inventing a this paradigm of computing definitely influenced the initial e/acc writings. I very much look forward to sharing more about our vision for the technology we are building soon. In terms of my background, as you've now learned, my main identity is @GillVerd. I used to work on special projects at the intersection of physics and AI at Alphabet, X and Google. Before this, I was a theoretical physicist working on information theory and black hole physics. Currently working on our AI Manhattan project to bring fundamentally new computing to the world with an amazing team of physics and AI geniuses, including my former TensorFlow Quantum co-founder @trevormccrt1 as CTO. Grateful every day to get to build this technology I have been dreaming of for over 8 years now with an amazing team.

And Verdon confirms the belief in Beffian doctrine:

Civilization desperately needs novel cultural and computing paradigms for us to achieve grander scope & scale and a prosperous future. I strongly believe thermodynamic physics and AI hold many of the answers we seek. As such, 18 months ago, I set out to build such cultural and computational paradigms.

I am fairly pessimistic about Extropic for reasons that should be obvious enough to people who've been monitoring the situation with DL compute startups and bottlenecks, so it may be that Beff's cultural engineering will make a greater impact than Verdon's physical one. Ironic, for one so contemptuous of wordcels.


Maturation of e/acc from a meme to a real force, if it happens (and as feared on Alignment Forum, in the wake of OpenAI coup-countercoup debacle), will be part of a larger trend, where the quasi-Masonic NGO networks of AI safetyists embed themselves in legacy institutions to procure the power of law and privileged platforms, while the broader organic culture and industry develops increasingly potent contrarian antibodies to their centralizing drive. Shortly before the doxx, two other clusters in the AI debate have been announced.

First one I'd mention is d/acc, courtesy of Vitalik Buterin; it's the closest to acceptable compromise that I've seen. It does not have many adherents yet but I expect it to become formidable because Vitalik is.

Across the board, I see far too many plans to save the world that involve giving a small group of people extreme and opaque power and hoping that they use it wisely. And so I find myself drawn to a different philosophy, one that has detailed ideas for how to deal with risks, but which seeks to create and maintain a more democratic world and tries to avoid centralization as the go-to solution to our problems. This philosophy also goes quite a bit broader than AI, and I would argue that it applies well even in worlds where AI risk concerns turn out to be largely unfounded. I will refer to this philosophy by the name of d/acc.

The "d" here can stand for many things; particularly, defensedecentralizationdemocracy and differential. First, think of it about defense, and then we can see how this ties into the other interpretations.

[…] The default path forward suggested by many of those who worry about AI essentially leads to a minimal AI world government. Near-term versions of this include a proposal for a "multinational AGI consortium" ("MAGIC"). Such a consortium, if it gets established and succeeds at its goals of creating superintelligent AI, would have a natural path to becoming a de-facto minimal world government. Longer-term, there are ideas like the "pivotal act" theory: we create an AI that performs a single one-time act which rearranges the world into a game where from that point forward humans are still in charge, but where the game board is somehow more defense-favoring and more fit for human flourishing.

The main practical issue that I see with this so far is that people don't seem to actually trust any specific governance mechanism with the power to build such a thing. This fact becomes stark when you look at the results to my recent Twitter polls, asking if people would prefer to see AI monopolized by a single entity with a decade head-start, or AI delayed by a decade for everyone… The size of each poll is small, but the polls make up for it in the uniformity of their result across a wide diversity of sources and options. In nine out of nine cases, the majority of people would rather see highly advanced AI delayed by a decade outright than be monopolized by a single group, whether it's a corporation, government or multinational body. In seven out of nine cases, delay won by at least two to one. This seems like an important fact to understand for anyone pursuing AI regulation.

[…] my experience trying to ensure "polytheism" within the Ethereum ecosystem does make me worry that this is an inherently unstable equilibrium. In Ethereum, we have intentionally tried to ensure decentralization of many parts of the stack: ensuring that there's no single codebase that controls more than half of the proof of stake network, trying to counteract the dominance of large staking pools, improving geographic decentralization, and so on. Essentially, Ethereum is actually attempting to execute on the old libertarian dream of a market-based society that uses social pressure, rather than government, as the antitrust regulator. To some extent, this has worked: the Prysm client's dominance has dropped from above 70% to under 45%. But this is not some automatic market process: it's the result of human intention and coordinated action.

[…] if we want to extrapolate this idea of human-AI cooperation further, we get to more radical conclusions**. Unless we create a world government powerful enough to detect and stop every small group of people hacking on individual GPUs with laptops, someone is going to create a superintelligent AI eventually - one that can think a thousand times faster than we can - and no combination of humans using tools with their hands is going to be able to hold its own against that. And so we need to take this idea of human-computer cooperation much deeper and further. A first natural step is brain-computer interfaces.…

etc. I mostly agree with his points. By focusing on the denial of winner-takes-all dynamics, it becomes a natural big tent proposal and it's already having effect on the similarly big tent doomer coalition, pulling anxious transhumanists away from the less efficacious luddites and discredited AI deniers.

The second one is «AI optimism» represented chiefly by Nora Belrose from Eleuther and Qiuntin Pope (whose essays contra Yud 1 and contra appeal to evolution as an intuition pump 2 I've been citing and signal-boosting for next to a year now; he's pretty good on Twitter too). Belrose is in agreement with d/acc; and in principle, I think this one is not so much a faction or a movement as the endgame to the long arc of AI doomerism initiated by Eliezer Yudkowsky, the ultimate progenitor of this community, born of the crisis of faith in Yud's and Bostrom's first-principles conjectures and entire «rationality» in light of empirical evidence. Many have tried to attack the AI doom doctrine from the outside (eg George Hotz), but only those willing to engage in the exegesis of Lesswrongian scriptures can sway educated doomers. Other actors in, or close to this group:

Optimists claim:

The last decade has shown that AI is much easier to control than many had feared. Today’s brain-inspired neural networks inherit human common sense, and their behavior can be molded to our preferences with simple, powerful algorithms. It’s no longer a question of how to control AI at all, but rather who will control it.

As optimists, we believe that AI is a tool for human empowerment, and that most people are fundamentally good. We strive for a future in which AI is distributed broadly and equitably, where each person is empowered by AIs working for them, under their own control. To this end, we support the open-source AI community, and we oppose attempts to centralize AI research in the hands of a small number of corporations in the name of “safety.” Centralization is likely to increase economic inequality and harm civil liberties, while doing little to prevent determined wrongdoers. By developing AI in the open, we’ll be able to better understand the ways in which AI can be misused and develop effective defense mechanisms.

So in terms of a political compass:

  • AI Luddites, reactionaries, job protectionists and woke ethics grifters who demand pause/stop/red tape/sinecures (bottom left)
  • plus messianic Utopian EAs who wish for a moral singleton God, and state/intelligence actors making use of them (top left)
  • vs. libertarian social-darwinist and posthumanist e/accs often aligned with American corporations and the MIC (top right?)
  • and minarchist/communalist transhumanist d/accs who try to walk the tightrope of human empowerment (bottom right?)

(Not covered: Schmidhuber, Sutton& probably Carmack as radically «misaligned» AGI successor species builders, Suleyman the statist, LeCun the Panglossian, Bengio&Hinton the naive socialists, Hassabis the vague, Legg the prophet, Tegmark the hysterical, Marcus the pooh-pooher and many others).

This compass will be more important than the default one as time goes on. Where are you on it?


As an aside: I recommend two open LLMs above all others. One is OpenHermes 2.5-7B, the other is DeepSeek-67B (33b-coder is OK too). Try them. It's not OpenAI, but it's getting closer and you don't need to depend on Altman's or Larry Summers' good graces to use them. With a laptop, you can have AI – at times approaching human level – anywhere. This is irreversible.

29

I've been wrong, again, pooh-poohing another Eurasian autocracy. Or so it seems.

On 29 August 2023, to great jubilation of Chinese netizens («the light boat has passed through a thousand mountains!», they cry), Huawei has announced Mate 60 and 60 Pro; the formal launch is scheduled for September 25th, commemorating the second anniversary of return of Meng Wanzhou, CFO and daughter of Huawei's founder, from her detainment in Canada. Those are nice phones of course but, specs-wise, unimpressive, as far as flagships in late 2023 go (on benchmarks, score like 50-60% of the latest iPhone while burning peak 13W so 200% of power). Now they're joined by Mate X5.

The point, however, is that they utilize Huawei's own SoC, Hisilicon Kirin 9000S, not only designed but produced in the Mainland; it even uses custom cores that inherit simultaneous multithreading from their server line (I recommend this excellent video review, also this benchmarking). Their provenance is not advertised, in fact it's not admitted at all, but now all reasonable people are in agreement that it's SMIC-Shanghai made, using their N+2 (7nm) process, with actual minimum metal pitch around 42 nm, energy efficiency at low frequencies close to Samsung's 4nm and far worse at high (overall capability in the Snapdragon 888 range, so 2020), transistor density on par with first-gen TSMC N7, maybe N7P (I'm not sure though, might well be 10% higher)… so on the border of what has been achieved with DUV (deep ultraviolet) and early EUV runs (EUV technology having been denied to China. As a side note, Huawei is also accused of building its own secret fabs).

It's also worse on net than Kirin 9000, their all-time peak achievement taped out across the strait in 2020, but it's… competitive. They apparently use self-aligned quad patterning, a DUV variant that's as finicky as it sounds, an absurd attempt to cheat optics and etch features many times smaller than the etching photons' wavelength (certain madmen went as high as 6x patterning; that said, even basic single-patterning EUV is insane and finicky, «physics experiment, not a production process»; companies on the level of Nikon exited the market in exasperation rather than pursue it; and it'll get worse). This trick was pioneered by Intel (which has failed at adopting EUV, afaik it's a fascinating corporate mismanagement story with as much strategic error as simple asshole behavior of individual executives) and is still responsible for their latest chips, though will be made obsolete in the next generations (the current node used to be called Intel's 10 nm Enhanced SuperFin, and was recently rebranded to Intel 7; note, however, that Kirin 9000S is a low-power part and requirements there are a bit more lax than in desktop/server processors). Long story short: it's 1.5-2 generations, 3-4 years behind the frontier of available devices, 5-6 years behind frontier production runs, 7-8 years after the first machines to make such chips at scale came onto market; but things weren't that much worse back then. We are, after all, in the domain of diminishing returns.

Here are the highlights from the first serious investigation, here are some leaks from it, here's the nice Asianometry overview (esp 3:50+), and the exhilarating, if breathlessly hawkish perspective of Dylan Patel, complete with detailed restrictions-tightening advice. Summarizing:

  1. This is possible because sanctions against China have tons of loopholes, and because ASML and other suppliers are not interested in sacrificing their business to American ambition. *
  2. Yes, it qualifies for 7nm in terms of critical dimensions. Yes, it's not Potemkin tulou, they likely have passable yields, both catastrophic and parametric (maybe upwards of 50% for this SoC, because low variance in stress-testing means they didn't feel the need to approve barely-functional chips, meaning there weren't too many defects) and so it's economically sustainable (might be better in that sense than e.g. Samsung's "5nm" or "4nm", because Samsung rots alive due to systemic management fraud) [I admit I doubt this point, and Dylan is known to be a hawk with motivated reasoning]. Based on known capex, they will soon be able to produce 30K wafers per month, which means 10s of millions of such chips soon (corroborated by shipment targets; concretely it's like 300 Kirins *29700 wafers so 8.9M/month, but the cycle is>1 month). And yes, they will scale it up further, and indeed they will keep polishing this tech tree and plausibly get to commercially viable "5nm" next - «the total process cost would only be ≈20% higher versus a 5nm that utilizes EUV» (probably 50%+ though).
  3. But more importantly: «Even with 50% yields, 30,000 WPM could support over 10 million Nvidia H100 GPU ASIC dies a year […] Remember GPT-4 was trained on ≈24,000 A100’s and Open AI will still have less than 1 million advanced GPUs even by the end of next year». Of course, Huawei already had been producing competitive DL accelerators back when they had access to EUV 7nm; even now I stumble upon ML papers that mention using those.
  4. As if all that were not enough, China simply keeps splurging billions on pretty good ML-optimized hardware, like Nvidia A/H800s, which abide with the current (toothless, as Patel argues) restrictions.
  5. But once again: on a bright (for Westerners) side, this means it's not so much Chinese ingenuity and industriousness (for example, they still haven't delivered a single ≤28nm lithography machine, though it's not clear if the one they're working on won't be rapidly upgraded for 20, 14, 10 and ultimately 7nm processes – after all, SMIC is currently procuring tools for «28nm», complying with sanctions, yet here we are), as it's the unpicked low-hanging fruit of trade restrictions. In fact, some Chinese doomers argue it's a specific allowance by the US Department of Commerce and overall a nothingburger, ie doesn't suggest willingness to produce more consequential things than gadgets for patriotic consumers. The usual suspects (Zeihan and his flock) take another view and smugly claim that China has once again shot itself in the foot while showing off, paper tiger, wolf warriors, only steals and copies etc.; and, the stated objective of the USG being «as large of a lead as possible», new crippling sanctions are inevitable (maybe from Patel's list). There exists a body of scholarship on semiconductor supply chain chokepoints which confirms these folks are not delusional – something as «simple» as high-end photoresist is currently beyond Chinese grasp, so the US can make use of a hefty stick.

All that being said, China does advance in on-shoring the supply chain: EDA, 28nm scanners, wafers etc.

* Note: Patel plays fast and loose with how many lithography machines exactly, and of what capacity, are delivered/serviced/ordered/shipping/planned/allowed, and it's the murkiest part in the whole narrative; for example he describes ASML's race-traitorous plans stretching to 2025-2030, but the Dutch and also the Japanese seem to already have began limiting sales of tools he lists as unwisely left unbanned, and so the August surge or imports may have been the last, and certainly most 2024+ sales are off the table I think.

All of this is a retreading of a discussion from over a year ago, when a less mature version of SMIC N7 process was used - also surreptitiously – for a Bitcoin mining ASIC, a simple, obscenely high-margin part 19.3mm² in size, which presumably would have been profitable to make even at pathetic yields, like 10%; the process back then was near-idential to TSMC N7 circa 2018-2019. 9000S is 107 mm² and lower-margin. Nvidia GH100, the new workhorse of cutting edge ML, made with 4nm TSMC node, is 814 mm²; as GPU chips are a strategic resource, it'd be sensible to subsidize their production (as it happens, H100 with its 98 MTr/mm² must be equally or a bit less dense than 9000S; A100, a perfectly adequate 7nm downgrade option, is at 65 MTr/mm² so we can be sure they'll be capable of making those, eg resurrecting Biren BR100 GPUs or things like Ascend 910). Citing Patel again, «Just like Apple is the guinea pig for TSMC process nodes and helps them ramp and achieve high yield, Huawei will likewise help SMIC in the same way […] In two years, SMIC will likely be able to produce large monolithic dies for AI and networking applications.» (In an aside, Patel laments the relative lack of gusto in strangling Chinese radio/sensor capabilities, which are more formidable and immediately scary than all that compute. However, this makes sense if we look at the ongoing chip trade war through the historical lens, with the reasonable objective being Chinese obsolescence a la what happened to the Soviet Union and its microelectronics, and arguably even Japan in the 80s, which is why ASML/Samsung/TSMC are on the map at all; Choyna military threat per se, except to Taiwan, being a distant second thought, if not a total pretext. This r/LessCredibleDefense discussion may be of interest).


So. I have also pooh-poohed the Chinese result back then, assuming that tiny crypto ASICs are as good as they will get within the bounds assigned to them, «swan song of Chinese industry», and won't achieve meaningful yields. Just as gwern de facto did in October 2022, predicting the slow death of Chinese industry in view of «Export Controls on Advanced Computing and Semiconductor Manufacturing Items to the PRC» (even mentioning the yellow bear meme). Just as I did again 4 months ago, saying to @RandomRanger «China will maybe have 7nm in 2030 or something». I maintain that it's plausible they won't have a fully indigenized supply chain for any 7nm process until 2030 (and/or will likewise fail with securing chains for necessary components other than processors: HBM, interposers etc), they may well fall below the capacity they have right now (reminder that not only do scanners break down and need consumables, but they can be remotely disabled), especially if restrictions keep ramping up and they'll keep making stupid errors, e.g. actually starting and failing an attempt at annexing Taiwan, or going for Cultural Revolution Round II: Zero Covid Boogaloo, or provoking an insurgency by force-feeding all primary school students gutter oil breakfasts… with absolute power, the possibilities are endless! My dissmissal was informed not by prejudice but years upon years of promises by Chinese industry and academia representatives to get to 7nm in 2 more weeks, and consistent failure and high-profile fraud (and in fact I found persuasive this dude's argument that by some non-absurd measures the gap has widened since the Mao's era; and there was all the graphene/quantum computing "leapfrogging" nonsense, and so on). Their actors haven't become appreciably better now.

But I won't pooh-pooh any more, because their chips have become better. I also have said: «AGI can be completed with already available hardware, and the US-led bloc has like 95% of it, and total control over means of production». This is still technically true but apparently not in a decisive way. History is still likely to repeat – that is, like the Qing China during the Industrial Revolution, like the Soviet Union in the transistor era, the nation playing catch-up will once again run into trade restrictions, fail at the domestic fundamental innovation and miss out on the new technological stage; but it is not set in stone. Hell, they may even get to EUV through that asinine 160m synchrotron-based electron beam thing – I mean, they are trying, though it still looks like ever more academic grift… but…

I have underestimated China and overestimated the West. Mea culpa. Alphanumericsprawl and others were making good points.


Where does this leave us?

It leaves us in the uncomfortable situation where China as a rival superpower will plausibly have to be defeated for real, rather then just sanctioned away or allowed to bog itself down in imperialist adventurism and incompetence. They'll have enough suitable chips, they have passable software, enough talent for 1-3 frontier companies, reams of data and their characteristically awkward ruthlessness applied to refining it (and as we've learned recently, high-quality data can compensate for a great disparity in compute). They are already running a few serious almost-OpenAI-level projects – Baidu's ERNIE, Alibaba's Tongyi Qianwen (maybe I've mentioned it already, but their Qwen-7B/VL are really good; seems like all groups in the race were obligated to release a small model for testing purposes), maybe also Tsinghua's ChatGLM, SenseTime etc.'s InternLM and smaller ones. They – well, those groups, not the red boomer Xi – are well aware of their weaknesses and optimize around them (and borrowing from the open academic culture helps, as can be often seen in the training methods section – thanks to MIT&Meta, Microsoft, Princeton et al). They are preparing for the era of machine labor, which for now is sold as means to take care of the aging population and so on (I particularly like the Fourier Intelligence's trajectory, a near-perfect inversion of Iron Man's plot – start with the medical exoskeleton, proceed to make a full humanoid; but there are other humanoids developed in parallel, eg Unitree H1, and they seem competitive with their American equivalents like Tesla Optimus, X1 Neo and so on); in general, they are not being maximally stupid with their chances.

And this, in turn, means that the culture of the next years will be – as I've predicted in Viewpoint Focus 3 years ago – likely dominated by the standoff, leading up to much more bitter economic decoupling and kinetic war; promoting bipartisan jingoism and leaving less space for «culture war» as understood here; on the upside, it'll diminish the salience of progressive campaigns that demoralize the more traditionally minded population.

It'll also presumably mean less focus on «regulation of AI risks» than some would hope for, denying this topic the uncontested succession to the Current Thing №1.

That's about all from me, thoughts?