Newcomb's problem splits people 50/50 in two camps, but the interesting thing is that both sides think the answer is obvious, and both sides think the other side is being silly. When I created a video criticizing Veritasium's video This Paradox Splits Smart People 50/50 I received a ton of feedback particularly from the two-box camp and I simply could not convince anyone of why they were wrong.
That lead me to believe there must be some cognitive trap at play: someone must be not seeing something clearly. After a ton of debates, reading the literature, considering similar problems, discussing with LLMs, and just thinking deeply, I believe the core of the problem is recursive thinking.
Some people are fluent in recursivity, and for them certain kind of problems are obvious, but not everyone thinks the same way.
My essay touches Newcomb's problem, but the real focus is on why some people are predisposed to a certain choice, and I contend free will, determinism, and the sense of self, all affect Newcomb's problem and recursivity fluency predisposes certain views, in particular a proper understanding of embedded agency must predispose a particular (correct) choice.
I do not see how any of this is not obvious, but that's part of the problem, because that's likely due to my prior commitments not being the same as the ones of people who pick two-boxes. But I would like to hear if any two-boxer can point out any flaw in my reasoning.

Jump in the discussion.
No email address required.
Notes -
The problem with Newcomb's problem is that it basically involves time travel, and generally underspecifies how that time travel works. Consider a similar problem:
Time 1: you discover box 1 with 1,000,000 points
Time 2: you discover a box 2 with 1,000 points
Time 3: someone shows up claiming to be a time traveler shows up and says that if you hand him box 2, he will multiply it by 1,000, and then go back in time to put it in box 1. Actually, he claims, that's where box 1 came from all along and if you don't give him the 1,000 your box 1 will disappear.
Assuming you are rational/selfish, whether you say yes or no very much depends on whether he's telling the truth. If the problem carefully specifies that he actually is a time traveler telling the truth, and time travel does work this way, then obviously you should give it to him (one box). If this happened in real life, I would not give him anything and two box, because my prior on time travel existing is less than 1/1000 and he's just a liar trying to con me. If the problem is not careful and is ambiguous about his truthfulness then people's answers are going to depend on their trustfulness, suspension of disbelief, or just general attitudes towards how willing they are to buy time travel in a hypothetical logic puzzle.
Actual Newcomb's problem is basically the same as this in that decisions you make in the future affect things in the past, and the being making the boxes has to have time travel powers in order to guarantee a 100% success rate (though not all version of the problem specify this precisely, maybe it just has a 99% success rate, or a vague but high success rate) The reason people so confidently disagree is that in any well-specified version of the problem the answer is obvious, but in any vague under-specification it's ambiguous to which well-specified version people will round it to. This is the exact same reason the Monty Hall problem is controversial as well. It's not merely there being a counter-intuitive answer, it's that the problem specifications are very volatile and people keep leaving important details ambiguous that they shouldn't.
Not time travel, just perfect prediction. If you're actually a perfect predictor then you can in essence see the future. If you had a perfect model of physics and initial conditions then you could predict a coin flip with 100% accuracy. The kind of reason a human does when presented with the boxes is no different unless you a proposing some spooky non-material stuff in the reasoning. The formulation I'm familiar with is perfect prediction in which case there are four theoretical cases.
You one box and Omega correctly predicted you would one box thus you get $1m
You one box and omega incorrectly predicted you would 2 box so you get zero. This is impossible by construction, omega cannot predict wrongly.
You two box and Omega incorrectly assumed you would one box, you get $1m + $1k. This is impossible by construction, Omega cannot be wrong.
You two box and Omega correctly guesses you'll two box. You get $1k.
There are only two actually possible options with the given constraints and you get to make a choice which of them is the case. This is not a paradox unless predicting future events is impossible.
Your whole reasoning relies on there being something intrinsically impossible about predicting your decisions, even as you lay out the reasoning for them. Is it so hard to imagine that someone could read you well enough to know which outcome you'll ultimately reach?
This is only possible if you model people as deterministic mechanisms and not as rational game theory agents. If Newcomb's problem posits that you make a decision AFTER Omega makes a decision, then Omega can be wrong. For instance, if you play a mixed strategy (one box with 50% probability, two box with 50% probability), then omega has to not only model your brain perfectly, but also your coin flip. If you used quantum decay to randomize then it would have to predict that perfectly. If omega can perfectly predict that then you've removed an important tool in Game Theory. It's like trying to argue that Rock Paper Scissors as a paradox because no matter what you do your opponent can predict you and defeat you. If you tell me that I can't do mixed strategies, or that other people can, in the past, respond to my mixed strategy outcomes then that's fundamentally incompatible with 90% of Game Theory. I don't even know what the rules of this are or what the goal is. Omega can do literally whatever it wants and I'll get literally any payoff that it chooses to give me. I suppose if it's a god that wants to punish two boxers then I guess I'll obey its commandments and one box, but that's not adjency, that's not game theory, that's just submitting to the religious edicts of a higher power with arbitrary rules.
We can model it much more simply by making an alternate version where, at time 1, you decide to one box or two box. Then at time 2 omega is informed of your decision and puts stuff in a box, then at time 3 you get the result of the decision you already made. Here you obviously 1 box. This is a very straightforward, simple, and uninteresting game theory problem. The problem with this then is not what the original Newcomb's box says happens. It says you make the decision after omega does. If you actually mean that people make the decision before omega then saying they make it after is lying.
Maybe this is still useful as a critique of attempts to map Game Theory to reality. Essentially saying "Every game is an iterated game played out over the course of your lifetime. Any decisions you make will affect your personality and reputation, so doing greedy things will hurt you in the long term even if they are the rationally correct choice to a one shot game that you see in the short term." Which, sure. This is how cooperation can exist in prisoner's dilemma-like situations, because you cannot incentivize (and it is irrational to try) to cooperate in a true, pure, one-shot prisoner's dilemma with no modifications, but none of those conditions apply in real life. But you also don't have mind-reading omegas in real life either, so I don't think that's quite what people mean by this.
Ultimately, the premises are fundamentally contradictory, so the only way to come to a solution is to suspend disbelief on half of them. Either you believe that omega can perfectly predict you and you have no agency, so hope that your were born as a one-boxer (because you don't get to decide), or you believe you are a rational agent who can make decisions when it says you can, in which case omega can't predict you so might as well two box. But these are beliefs about the premises of the problem, not about what is good epistemic or rational behavior in a given coherent scenario (which always follows logically and mathematically from the premises and math to determine the maximum payoff)
These are not in tension. In some game theory scenarios adding randomness, if such a thing is actually possible, is useful to some agent. But Newcome's problem is not such a scenario. Adding any chance of walking away without the $1 million is not worth going for the extra $1k and to the best of my knowledge the best you could do by adding randomness would be to make your expected value $500,500. Whereas your expected value if you cooperate is $1m
As for the rest of the post, yeah just seems like you're demanding the hypothetical grant you libertarian free will and say something different than it says. It's very "But I did eat breakfast this morning" fighting of the hypothetical. If you want to demand that actually you can't be predicted, even hypothetically then you're just not willing to engage with the question.
I'm not demanding that I can't be predicted, I'm asking "what does that even mean?" and also "if that's true in the hypothetical, then what's the point? there's nothing left." The problem just axiomatically erases the game away with no details.
Consider another game. Suppose you are given an opportunity to play Tic Tac Toe against Umega. If you choose to play and you lose, you get 0 points. If you win or tie, you get 1000 points. But if you forfeit prior to playing the game you can get 100 points. Umega can't predict the future, but instead is extremely tricky. Somehow, it always wins at Tic Tac Toe. No matter what you originally intend to do or plan you make, things don't go the way you intended and it tricks you and lose anyway. What do you do?
The way to maximize your points here is to forfeit. Because if you try to play the game then, axiomatically, Umega tricks you and you lose. If you present a winning strategy for Tic Tac Toe that cannot be beaten and say you'll stick to that strategy no matter what Umega does then I retort that you attempt to do that, but then Umega tricks you and you lose anyway. It's a master of psychology and deceit beyond human comprehension. Okay, fine, you have no agency, you forfeit and get the 100 points because the game mathematically reduces to "forfeit: 100 points, don't forfeit: 0 points" with no other options.
Now suppose I go around presenting this to Chess Grandmasters or something (since Tic Tac Toe masters don't exist), and present it as some deep and extremely challenging variant of Tic Tac Toe and grand strategy. Chess Grandmasters stumped by this challenging variant of Tic Tac Toe!
It's not Tic Tac Toe! The optimal strategy of the game is completely and utterly independent to strategies for winning Tic Tac Toe. No matter how much or how little someone knows about Tic Tac Toe, it has nothing to do with the optimal strategy for this game (which is just forfeit) because it's not even engaging with the rules at all. You could substitute basically any strategy game there, I merely used Tic Tac Toe to point out the apparent contradiction of axiomatically declaring that you lose despite there being a known winning strategy.
This has the same fundamental problem as Newcomb's problem. If your agency is stripped from you then what even is the point of the thought experiment? "Suppose you have no free will, and will be punished for trying to act as if you have free will. What do you do?" I guess I forfeit and hope that next time I'm offered a thought experiment it allows me to make choices.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link