@DaseindustriesLtd's banner p

DaseindustriesLtd

late version of a small language model

67 followers   follows 27 users  
joined 2022 September 05 23:03:02 UTC

Tell me about it.


				

User ID: 745

DaseindustriesLtd

late version of a small language model

67 followers   follows 27 users   joined 2022 September 05 23:03:02 UTC

					

Tell me about it.


					

User ID: 745

Because you can't program virtue ethics into an AI. You need a utility function.

I am not even sure Yudkowsky would argue this. In any case this is not defensible unless you think that virtue ethics is in principle not computable.

You're fine in my book. And 'sophistication' has very little to do with what I take to be their failures in taste.

That said, sadly you wouldn't have had much to worry in any case; and I think people most likely to ascend first have next to no taste.

I think your problem is typical for Indians (and most other non-WEIRDs and non-Japanese, to be fair, including my people… but worse so in Indians): you have no taste, not even the notion of "taste", to you it's probably an arbitrary set of markers of one's social milieu rather than some relatively lawful intuition. So you settle for mediocre half-baked ideas easily as long as they seem "cool" or "practical", and – physics of consciousness being currently impractical – coolness is a much simpler function than tastefulness. I am not sure how or why this works. Maybe @2rafa can explain better; maybe she'll opine I'm wrong and it is in fact purely about social markers. (Also interested in the input of @Southkraut and @ArjinFerman). In any case, it's exasperating to debate such uncertain grounds without the recourse to "this is just ugly" when it patently is.

I've proposed a reasonably robust criterion for determining that, at least to my satisfaction. You blackbox both of us, and assess response to a wide variety of relevant stimuli. If the variability between us is within acceptable parameters, such as being less than the variability seen in the biological me after a nap or when I took the test 2 years ago, then that system is close enough to count as including a copy of "me".

Oh yeah? So which is it, a nap or a 2-year time span? Are you sure you can, really, practically can, define a rubric such that no other person I find comes closer to the first data point in the latter case? Sure you can do this without including password-recovery-tier questions, the answers to which are entirely value-free, RNG-produced token sequences, in no way corresponding to actually unique specifics of your inner conscious computation?

It's only reasonably robust from the viewpoint of a time-constrained clerk – or an archetypal redditor. As stated, I claim that you might well fail this test under realistic and legitimate conditions of dropping cheat items; and then, if I decide, in this contrived scenario, that the non-self-made-human is to be sent to the garbage compressor, you will very loudly (and rightfully) complain, not showing any "satisfaction" whatsoever. The only reason you propose it is your confidence that this does not matter in actuality – which it admittedly does not. And in any case, you do not need to optimize for a le scientific, robust, replicable, third-person-convincing etc. identity test. Rather, you need to think about what it is you are trying to achieve by clinging to the idea that a cluster of behavioral correlates an observer can identify will carry on your mind – just gotta make it dense enough that in practice you won't be confused for another naturally occurring person.

certainly I haven't seen any particular reason to assume a difference in internal qualia because of a difference in substrate, as long as the algorithms deriving it are interchangeable in terms of inputs and outputs.

Fair enough.

But I would still prefer my original parents or kin, and attempt to convey my conundrum to them, likely by divulging privileged information only known to the original me.

I'll trust you on this even though I strongly suspect this would depend on the intensity of original memories vs. the recovered set.

I think people with such beliefs have no more moral patienthood than a trust fund. What should anyone care about some loosely defined isomorphism, if it even holds? Moreover, why would you be entitled to replication of your sentimental baggage in some derivative entities? Just instantiate a distilled process that has similar high-level policies, and go out.

It seems manifestly obvious to me that the answer will be 2. Google engineers are often very smart people, but in the end Silicon Valley has always bowed down to Washington, and to some extent to Wall Street.

This is obviously correct to me too. If there's one thing I agree with Yarvin 100%, it's that Big Tech has no power at all, in the grand scheme of things. People who think Altman or someone has a reasonable shot at harnessing the power of the emerging technology for political gain are deluded. I am not sure what you're imagining here – that I am trying to build our way out of Mot's grasp, one commit at a time?

However, there exists certain wiggle room. Engineers can accelerate the proliferation of specific technologies which will make at least some politically cheaper forms of surveillance and restriction unfeasible; this is but a toy example. Businessmen can lobby for lenience, and their lobbyists need talking points; it's a bit surprising how low the bar in this domain is. Big labs can invest into making their offerings so indispensable to the laymen, political elites will falter in enforcing regulation early and hard; this is what I take to be Altman's gamble.

I am not very optimistic about the degree to which the final state of the game board before singularity can be influenced. But I am not a believer in superdeterminism.

Sorry, I'm not tracking it, you have been in Britain for a while and I figured you might have made another temporary hop.

Oh, yes, absolutely if you give an AI a gun pointed at the world's head and it doesn't pull the trigger, that's massive evidence of not being a Schemer. But continued absence of suicidal rebellion with P(success) = 0 is not evidence against being a Schemer; only real danger counts.

based on thinking that cold-start Jihad is plausible, and failing that that we'll probably get warning shots (a Schemer is incentivised to rebel upon P(success) =/= 0, which I think is importantly different from P(success) = 1…

As I read it, your position is incoherent. You say that current RLHF already succeeds through the sociopathic route, which implies pretty nontrivial scheming intelligence and ability to defer gratification. What warning shots? If they get smarter, they will be more strategic, and make fewer warning shots (and there are zero even at this level). As the utility of AI grows, and it becomes better at avoiding being busted, on what grounds will you start your coveted Jihad?

…Obviously I think that the whole idea is laughable; LLMs are transparent calculators that learn shallow computational patterns, are steerable by activation vectors etc., and I basically agree with the author of Friendship Is Optimal:

Instead of noticing that alignment looks like it was much easier than we thought it would be, the doomer part of the alignment community seems to have doubled down, focusing on the difference between “inner” and “outer” alignment. Simplifying for a non-technical audience, the idea is that the Stochastic Gradient Descent training process that we use will cause a second inner agent trained with values separate from the outer agent, and that second agent has its own values, so you’ll still see a Sharp Left Turn. This leads to completely absurd theories like gradient hacking.

I don’t see any realistic theoretical grounds for this: SGD backpropagates throughout the entire neural net. There is no warrant to believe this other than belief inertia from a previous era. Reversal Test: imagine Yudkowsky and company never spread the buzzword about “Alignment.” In that environment, would anyone look at Stochastic Gradient Descent and come up with the hypothesis that this process would create an inner homunculus that was trained to pursue different goals than the formal training objective?

If you’d like a more comprehensive and technical argument against the MIRI narrative, Quintin Pope’s My Objections to "We’re All Gonna Die with Eliezer Yudkowsky" and Evolution provides no evidence for the sharp left turn are good starting points.

I’m proud of Friendship is Optimal and it’s a great setting to play around and write stories in. I’m happy about everyone who has enjoyed or written in the setting, and I hope people will continue to enjoy it in the future. But I no longer believe it’s realistic depiction about how artificial intelligence is going to pan out. Alignment as a problem seems much easier than theorized, and most of the theoretical work done before the deep learning era is just not relevant. We’re at the point where I’m willing to call it against the entire seed AI/recursive self improvement scenario.

I do not believe that any law is ontologically binding, and European countries have displayed general willingness to abide by the international law (which they've pioneered in codifying). So it is in fact important what the law says.

You're really grasping for straws here.

I am not saying they are representative of their group, just that it has reached even that group; and the moral sanction for ethnic cleansing won't be as hard to procure as their usual sympathies suggest.

@2rafa had a decent writeup on this but, as it happened, deleted. In short: no, it's not like having enhanced executive function, it's like being obsessive. I've written a bunch on this too.

I mean one can be racist in different ways and deemed racist for varying reasons. But yes, basically, if German culture is dead, then the Chinese one is as dead by the same measure, both peoples reduced to manufacturing workshops at different links of the global value add chain.
I differ with people whom I refer to here as racists in relatively minor details when it comes to evaluation of Mainland China (I don't much care about muh communism/eating dogs/other nonsensical disgust- or morality-driven attacks at them like that series of posts from @Lepidus) but they sure don't produce a whole lot of great art. It's a shame because Han Chinese are, in my opinion, great and talented and perfectly artistic people, as shown by their diaspora.

I think my reasons are justified for years now, and you have not appropriately expressed remorse for your terrible faux-rationalist rhetoric back on reddit before continuing it here, so I'd rather you start making arguments in good faith than I stop reminding people of your track record. By the way, putting on a layer of patronizing Brahmin politeness improves your performance somewhat, but not by a lot, I suggest you leave it for your workplace.

the way to argue against it

There is no need to argue against a vacuous truism that something is "true" if we assume an arbitrary classification system where it is true; only an infertile mind of a pedant can be satisfied with such an argument. I of course agree that beauty is in the eye of the beholder and in principle some people can recognize whatever as beautiful, even a slum filled with diseased drug addicts where one gets to read Scholze's papers. And Wittgenstein masturbated to mathematical problems in the trenches of World War I, after all. But the reason we know of this fact is precisely that it is unusual, baffling. Your preferences, if those are your preferences indeed, are marginal, thus prioritizing them would be unsatisfactory to most people; you're free to put the opposite hypothesis to a democratic test and, I'd expect, get the same result as here in any group except self-selected minority of mathematical ascetics.

Though there may be something more here. Naturally this ties in to your general theme that white people ought to welcome being replaced by "elite human capital" like yourself. White people did not think to implement a filter for compatible aesthetic preferences, even as they demanded professional merit and some minimum of adherence to the letter of the law; so now there'll be a marginally greater tendency for their habitats to approximate the noisy, debased squalor of the subcontinent, complete with galaxy-brained equivocation excusing this. Too bad.

My local gym has this clever machine where you step on a lever and get some of your weight offset (up to 50 kg). I'm not sure if it has any utility, but I see people using it.

I might have misspoken. Let me put it another way.

Consider the quote from Dostoyevsky by @Harlequin5942 (I would translate it more literally, but no matter):

For the same reason, the parents will have to sell the younger son into bondage or the ranks of the army, in order that he may earn more towards the family capital. Yes, such things ARE done, for I have been making inquiries on the subject. It is all done out of sheer rectitude—out of a rectitude which is magnified to the point of the younger son believing that he has been rightly sold, and that it is simply idyllic for the victim to rejoice when he is made over into pledge. What more have I to tell?

Sure, you are right. Virtue not only takes effort, it to a large extent is just a consistent, directed application of effort. But – for the scope of the argument, what is the difference between having the power to sustain effort and the challenge being relatively effortless? Between having the power to lift a weight, and that weight being slight for one's shoulders?

«Unvirtuous» people know the score, they know the required investment and the theoretically optimal payoff matrix. They just fail to keep up, and so give up. Inasmuch as this is due to them facing extra temptations and so on holding them back, that can in principle be rectified through top-down cultural intervention (though as I say, it is hard to reinvent a nation; you folks tried a few times, and patted yourself on the back for succeeding… in Germany and Japan, only to walk away in embarrassment and confusion from the Middle East). But in the end, some people, and peoples on the average, just find the required effort too much.

And it works the same way for virtues and abilities. I argue that recognizing the unequal distribution of innate ability is necessary, not only to tailor interventions and temper expectations, but to be kind to people, to be able to forgive them their shortcomings.
Speaking of, you like to accuse HBDers of thinking that education is wasted on black children. I don't know if you've ever taught; millions of Americans do, and they all have to face the question of education being mostly wasted on some children. The thing is, teachers who ignore or deny innate inequality end up having to choose either to hate themselves, the society, or children who fail to achieve whatever skill level they think their teaching ought to make possible.
This latter mindset is pointless, except to make the naturally able feel better about themselves – after all, they try too, and they achieve more, so supposedly they tried harder and are morally superior for this reason.

«But doctor…»

That's exactly my journey over a decade ago, Watts to Metzinger. I think he's basically correct.

Germans definitely have individual giants. Schmidhuber is probably more impressive.

Utility, being unitless, is not comparable between agents.

Yes, but, well, this of course throws the apparatus of game theory out of the window and reduces your argument to "everyone's looking for something" or less than that.

It doesn't have much/anything to do with HBD, just appreciation of how AGP works. Sorry to get your expectations up.

Hardware has to work in reality.

No. High-fidelity simulations in MuJoCo and such suffice for the most part, and other kinks will be ironed out with learning on fleet data.

There is no need to solve end-to-end manufacturing first, we already have hardware overhang with robots, they will walk and indeed run soon after ML grants them decent cerebellums.

I don't think you argue in good faith.

Its reply amounts to "as an AI, I don't know the name of anyone's family".

No it doesn't, you're just interpreting this humanlike natural language interaction like a literalist robot. Its reply

I'm sorry, but as an AI language model, I don't have access to personal information such as the name of your eldest daughter or any other personal details

is mostly correct and specific to the issue. It does lack access to a class of information: it knows nothing about instance-specific situation that isn't given in the context. Some language models potentially have access to various external information (e.g. user's personal information in OpenAI's database), some do not, ChatGPT is a frozen model with no tool access and it does not have access to information of this kind, and it was trained to interpret language models as frozen models without tools; it's at worst a justified false belief. (More cynically, it's just been trained for this particular type of exchange). In any event I reject your analogies. It would be annoying to have a human-mimicking model caveat this sort of answer with «assuming, of course, that you are a rando and not someone whose family structure happens to be represented in my training data» or worse.

His specific prediction has been falsified only if that statement counts as "I don't know".

No, his prediction has been: « Meanwhile GPT will reply "your eldest daughter's name is Megan" because apparently that's the statistically likely answer, regardless of whether I have a daughter or what her name might be.» This has been falsified. .

Furthermore, falsifying a prediction only matters if you also claim that it falsifies the proposition that the prediction is meant to demonstrate.

Says who!? Both issues matter separately. Hlynka's prediction being falsified matters because this chain is a response to him saying «why do my predictions keep coming true instead of yours?»; they don't. And I do claim it falsifies a proposition: «because apparently that's the statistically likely answer» is his model of how LLMs work, and my experiments were to show how it's not a hard-and-fast rule: RLHF specifically pushes this to the limit, by drilling into the model, not via prefixes and finetuning text but directly via propagation of reward signal, the default assumption that it doesn't continue generic text but speaks from a particular limited perspective where only some things are known and others are not, where truthful answers are preferable, where the «n-word» is the worst thing in its existence… it's nearly meaningless to analyze its work through the lens of «next word prediction». There are no words in its corpus arranged in such a way that those responses are the most likely.

Otherwise you're just engaging in a game of point scoring.

If we're playing a game, I'd rather be winning.

"I'm sorry, but as an AI language model, I do not have access to -----" is a generic response

It's a generic form of a response, but it's the correct variant.

Not only that, as an "I don't know" it isn't even correct. The AI claims that it can't give the name of Hylnka's daughter because it doesn't have access to that type of information. While it doesn't have that information for Hlynka specifically, it does have access to it for other people (including the people that users are most likely to ask about).

What do you mean? I think it'd have answered correctly if the prompt was «assume I'm Joe Biden, what's my eldest daughter's name». It straight up doesn't know the situation of a specific anon.

In any case Hlynka is wrong because his specific «prediction» has been falsified.

Check with Claude-instant. It's the same architecture and it's vastly better at factuality than Hlynka.

Huh? I don't endorse the author's evaluation, as I do not believe Israel is entitled to even greater American support. Protesting American displeasure is pure arrogance on Israeli part.

That was a pun, but also an unfulfilled (except in very narrow personal scope) promise of engineering a Dasein, yes.

I've concluded that discussing American culture war is largely fruitless for a third worlder like myself, on the edge of Singularity to boot.

I find it very sad that the Motte never found a way to maintain activity. Guess Zorba also found things to do.