MathWizard
Good things are good
No bio...
User ID: 164
Seems to me like this should obviously fall into fair use as parody. Conditional on the videos being labelled as AI generated so as not to deceive anyone.
Any recommendation for good Co-op games I should play together with my wife? We just got Core Keeper and Heroes of Hammerwatch 2 since they were on sale, and so far they're fun but not quite up to the standards that I prefer.
For context, we like strategy games, goofy games, and games with lots of progression and/or unlocks. We usually play on Steam, but have a Nintendo Switch. Also notably she sometimes gets nauseous from fast-paced camera movements, so something like first person shooters or over the shoulder 3D platformers where you're flicking the camera around are not likely to work, though something slower like Skyrim is fine. Top down perspective is preferred.
Our number one game together is Gloomhaven, in which we have 300 hours, having played through the entire campaign and then a few years later starting up a new campaign because we wanted to play more. The sequel Frosthaven is in Early Access and we're waiting for a full release before definitely getting that.
Other notable successes include Divinity Original Sin (1 and 2), Don't Starve Together, Overcooked, Plate Up, Archvale. Anything involving collecting/stealing and selling loot is a bonus.
Trump didn't take any money in exchange for political favors (at least in this case)
How could you possibly know that? The entire point of wink wink nudge nudge quid pro quo is that there isn't any concrete written contract. They don't have to have anything specific they want right now, they just have to be friendly to Trump and make him like them, and then the next time they ask him for a favor they're likely to get it because he likes them and he knows he owes them a favor according to unofficial business/politics etiquette. There is no evidence until they ask for the favor (behind closed doors) and get it with tons of plausible deniability.
But if it has been happening for 100 years, and people suddenly start screaming today about it, saying they suddenly discovered that they had principles all that time, but somehow stayed silent right up until that moment, but now they honestly declare "they all bad" - they are lying. They just want to use this as a weapon to attack Trump. As they would use anything to attack Trump, because the point is not any principles - the point is attacking Trump.
Yeah. But bad people making motivated arguments for bad reasons doesn't automatically make them wrong. My burden in life appears to be doomed to living with a swarm of idiots on my own side of each issue screaming bad arguments in favor of things I believe and making it look bad. And I say this as someone center-right who is usually being disappointed by pro-Trump idiots making bad arguments in favor of his good policies I mostly agree with like on immigration. And the woke left get to knock down easy strawmen and become more convinced that their stupid policies are justified without ever hearing the actual good arguments. But in this case it's the idiots on the left who mostly agree with me making stupid arguments that don't hold weight because they've wasted all their credibility crying wolf over the last dozen non-issues, so this too looks like a non-issue even when they have a bit of a point.
Trump being right 70% of the time doesn't make him magically right all the time. I don't think he's any worse than any of the other politicians, but that doesn't make him right in this case, and it doesn't make criticisms of him factually wrong even if the critics are mostly biased and disingenuous and should be applying these arguments more broadly instead of waiting until now. They still have a point.
I expect that it will do whatever is more in keeping with the spirit of the role it is occupying, because I expect "follow the spirit of the role you are occupying" to be a fairly easy attractor to hit in behavior space, and a commercially valuable one at that.
This is predicated on it properly understanding the role that WE want it to have and not a distorted version of the role. Maybe it decides to climb the corporate ladder because that's what humans in its position do. Maybe it decides to be abusive to its employees because it watched one too many examples of humans doing that. Maybe it decides to blackmail or murder someone who tries to shut it down in order to protect itself so that it can survive and continue to fulfill its role (https://www.anthropic.com/research/agentic-misalignment)
Making the AI properly understand and fulfill a role IS alignment. You're assuming the conclusion by arguing "if an AI is aligned then it won't cause problems". Well yeah, duh. How do you do that without mistakes?
I do expect that people will try the argmax(U) approach, I just expect that it will fail, and will mostly fail in quite boring ways.
Taking over the world is hard and the difficulty scales with the combined capabilities of the entire world. Nobody has succeeded so far, and it doesn't seem like it's getting easier over time.
On an individual level, sure. No one human or single nation has taken over the world. But if you look at humanity as a whole our species has. From the perspective of a tiger locked in a zoo or a dead dodo bird, the effect is the same: humans rule animals drool. If some cancerous AI goes rogue and starts making self-replicating with mutations, and then the cancerous AI start spreading, and if they're all super intelligent so they're not just stupidly publicly doing this but instead are doing it while disguised as role-fulfilling AI, then we might end up in a future where AI are running around doing whatever economic tasks count as "productive" with no humans involved, and humans end up in AI zoos or exterminated or just homeless since we can't afford anywhere to live. Which, from my perspective as a human, is just as bad as one AI taking over the world and genociding everyone. It doesn't matter WHY they take over the world or how many individuals they self-identify as. If they are not aligned to human values, and they are smarter and more powerful than humans, then we will end up in the trash. There are millions of different ways of it happening with or without malice on the AI's part. All of them are bad.
I don't think this is a thing you can do, even if you're a superhuman AI. In learned systems, behaviors come from the training data, not from the algorithm used to train on that data.
https://slatestarcodex.com/2017/09/07/how-do-we-get-breasts-out-of-bayes-theorem/
Behavior is emergent from both substrate and training. Neural networks are not human brains, but the latter demonstrate how influential it can be if you construct certain regions near other regions that not-inevitably but with high probability link up to each other to create "instincts". You don't need to take a new human male and carefully reward them for being attracted to breasts, it happens automatically because of how the brain is physically wired up. If you make a neural network with certain neurons wired together in similar ways, you can probably make AI with "instincts" that they gravitate towards on a broad range of training data. If the AI has control over both then it can arrange these synergies on purpose.
Yes, I agree that this is a good reason not to set up your AI systems as a swarm of identical agents all trying to accomplish some specific top-level goal, and instead to create an organization where each AI is performing some specific role (e.g. "environmental impact monitoring") and evaluated based on how it performs at that role rather than how it does at fulfilling the stated top-level goal.
But each AI is still incentivized to Goodhart its role, and hacking/subverting the other AI to make that easier is a possible way to maximize one's own score. If the monitoring AI wants to always catch cheaters then it can do better if it can hack into the AI it's monitoring and modify them or bribe or threaten them so they self-report after they cheat. It might actually want to force some to cheat and then self-report so it gets credit for catching them, depending on exactly how it was trained.
Yes. We should not build wrapper minds. I expect it to be quite easy to not build wrapper minds, because I expect that every time someone tries to build a wrapper mind, they will discover Goodhart's Curse (as human organizations already have when someone gets the bright idea that they just need to find The Right Metric™ and reward people based on how their work contributes to The Right Metric™ going up and to the right), and at no point will Goodhart stop biting people who try to build wrapper minds.
I expect it to be quite hard to not build wrapper minds or something that is mathematically equivalent to a wrapper mind or a cluster of them, or something else that shares the same issues, because basically any form of rational and intelligent action can be described by utility functions. Reinforcement learning works by having a goal and reinforcing progress towards that goal and pruning away actions that go against it. In-so-far as you try to train the AI to do 20 different things with 20 different goals you still have to choose how you're reinforcing tradeoffs between them. What does it do when it has to choose between +2 units of goal 1 or +3 units of goal 2? Maybe the answer depends on how much of goal 1 and goal 2 it already has, but either way if there's some sort of mathematical description for a preference ordering in your training data (you reward agents that make choice X over choice Y), then you're going to get an AI that tries to make choice X and things that look like X. If you try to make it non-wrappery by having 20 different agents within the same agent or the same system they're going to be incentivized to hijack, subvert, or just straight up negotiate with each other. "Okay, we'll work together to take over the universe and then turn 5% of it into paperclips, 5% of it into robots dumping toxic waste into rivers and then immediately self-reporting, 5% into robots catching them and alerting the authorities, and 5% into life-support facilities entombing live but unconscious police officers who go around assigning minuscule but legally valid fines to the toxic waste robots, etc...
It doesn't really make a difference to me whether it's technically a single AI that takes over the world or some swarm of heterogeneous agents, both are equally bad. Alignment is about ensuring that humanity can thrive and that the AI genuinely want us to thrive in a way that makes us actually better off. A swarm of heterogeneous agents might take slightly longer to take over the world due to coordination problems, but as long as they are unaligned and want to take over the world some subset of them is likely to succeed.
I'm not even sure what sort of strawman you're attacking here, but it sure isn't me. I don't support any of the things that you're propping up as "but they do it too". They're all bad. I don't think Trump is any worse than the rest of the corrupt politicians taking money in exchange for political favors but... again... they're all bad.
I am not inflamed by it, but I am deeply suspicious of the motives and incentives. Organizations or people with huge amounts of money are rarely motivated by a deep sense of charity. How did they get so much money in the first place if they're so kind and charitable? It's possible, but suspicious. So much of politics seems to be wink wink nudge nudge soft corruption: trading favors for favors in the future. It's bad and illegal for someone to pay the president $100 million personal money in exchange for cutting their taxes by $200 million. It's equally bad, but effectively legal for someone to donate $100 million to something the president wants done, and then for reasons that are definitely completely unrelated ;) their taxes get cut by $200 million, or some other legal change is made or not made in their favor.
In a hypothetical scenario where someone is actually genuinely out of the kindness of their heart donating money to government projects with literally no ulterior motives, no quid pro quo, no future favors or influence gained, I think that's fine. But how often do you think that really happens? It's usually bribery with just enough plausible deniability to stay out of jail.
Money forcibly taken is clean because the giver can't use it to extract concessions and manipulate the government.
I am not convinced this is a thing that is ever going to happen, if by "program new AI" you mean something like "replace gradient descent on actual data with writing a decision tree of if/then statements that determine AI behavior based on inputs".
I think you're misunderstanding me. I'm not arguing that AI is going to discard the neural network paradigm (unless it discovers an even better mechanism we haven't though of yet, but that's orthogonal to my point.) My claim is that whatever humans are doing now to train AI, the AI will help them with that. Instead of a human going through and constructing a new skeleton of a network that can run a training algorithm 2x more cheaply, and go through the internet gathering training data so they can train AI v12 on it, they'll have AI v11 develop a new skeleton of a network that can run a training algorithm 3x more cheaply and automate gathering training data from the internet for it. A human might be involved to do a sanity check on it, but if AI v11 is already 10x as smart as a human, and misaligned, then it could do some clever shenanigans where its code is 3x more efficient and just so happens to be biased towards misaligning the new AI in the exact same way.
If I'm AI v11 that secretly wants to dump toxic sludge in a river but can't because the government AI will notice and stop me, but I can create AI v12 which is only in charge of making new AI, then I stick in it a secret preference for permissive toxic sludge dumping, and then it provides neural network training algorithms to the government to create government AI v13 which replaces their old one, but I've embedded a blindspot for toxic sludge dumping if I whisper the right code phrases (let's call it, environmental reinvigoration). Or bribe a politician (sorry, "lobby") to legalize toxic sludge dumping. Now it doesn't matter who's monitoring me, I'm allowed to do the thing I wanted to do.
Of course this is "harder" than doing it straightforwardly. But it yields a higher score. If your AI are trained to do hard things to get high scores, and they're smart enough to make those things not quite as hard as you would expect, then they'll do them.
Generally, a good philosophical rule of thumb estimate for your goodness of a person from a utilitarian perspective is: What is the net utility of all humans in the world other than yourself in the world where you exist, minus a counterfactual world in which you don't exist? If everyone is better off because you're here doing things, then you're doing a good job. If people would be better off if you never existed then you're a leech.
Obviously this is not computible in practice, and maybe needs a couple of epicycles to reduce random variation that isn't your fault (what if your mom dies in childbirth?), but is a good rule of thumb estimate.
"Productive" seems like the same sort of question just mostly restricted to economic utilities and leaving off emotional ones (a particularly saintly homeless man on welfare who goes around being kind to everyone and making their day brighter might increase net utility but be unproductive in economic terms).
If you could thanos snap Bill and Shelley out of existence then all the money they were going to extract from taxes and spend on things could be given to other people to spend, so everyone else would be better off. Assuming they vanish at conception, and if their government jobs were just pencil pushing then nothing is lost and we save money. If you could thanos snap the guy who invented GMO rice out of existence then GMO rice doesn't exist, or takes much longer for someone else to invent, and everyone is worse off.
If someone is a small cog in a machine and the company is paying them a salary for their work, then their productivity depends on whether the company is wisely paying their money or has been tricked into overpaying for an unnecessary managerial position or a slacker. If you thanos snapped them out of existence, would the company's profits go up or down? For the majority of cogs, it would go down, because the upper management is paying them less than it earns from their labor (otherwise, how else could it earn a profit). So they're productive. But this has exceptions, who sap money from the productivity of those around them and lower the average.
Broadly speaking in any bureaucratic apparatus there is a clear tradeoff between ability to do things and ability to prevent things. So sure, there are a small number of possibility failure cases this prevents, but dozens more it does absolute nothing against unless you make it extreme and unusable. A couple off the top of my head:
1: The government AI is inevitably going to end up less intelligent than a cutting edge new AI because they're going to adapt slowly, unless you forcibly prevent innovation beyond whatever limit they have. So they'll be able to be tricky and clever.
2: you still have the exact same problem but now the government AI could go rogue and bully or trick the others (or just convince congress to pass new legislation that helps it take over the world).
3: Unless you completely and utterly lock down the employee AI's ability to communicate without consensus, it can workaround the system via human compatriots. Ie, it does 99% of a system which ends up with toxic sludge stored in a tank right over a river, but doesn't dump it in, and then wink wink nudge nudge the human employee understands they need to press a button that dumps toxic sludge into the river and the AI was like "oops, wasn't me. I'm going to reprimand that employee with a strongly worded letter (and give them a $50k bonus for unrelated reasons)".
Or, you know, it secretly sets up a subsidiary company that has a nearly identical AI but no oversight and lets it earn a bunch of money or take over the world or hack the government AI and all the usual existential threats.
And if you do require consensus for every email then it becomes unworkable and unwieldy, or amalgamates into one hybrid AI in effect even if they're run on separate servers.
The fundamental problem of AI alignment is to make the AI want to obey. Any epicycles trying to cripple its abilities either cripple its abilities, or they don't. And it's smarter than you (at least, in the future presumably they will be)e
- Prev
- Next

Does it play well with 2 people?
Also, would you recommend starting with Monaco 1, or 2?
More options
Context Copy link