On This Problem Rational People Do Worse
A supercomputer that predicts human choices with near-perfect accuracy offers you a simple choice: take both boxes on the table — one containing $1,000 and one mystery box — or take only the mystery box. If the computer predicted you'd take only the mystery box, it put $1 million inside. If it predicted you'd take both, the mystery box is empty. The catch? The computer has already made its prediction and set up the boxes before you walked in the room. This seemingly simple problem divides people almost exactly in half, with each side convinced the other is obviously wrong. What makes it stranger is that the «rational» choice might actually make you poorer.
Ключевые выводы
Two-thirds of people choose only the mystery box (one-boxers), trusting the supercomputer's track record, while one-third take both boxes (two-boxers), reasoning that the boxes are already set and their choice can't change the past.
The paradox exposes a fundamental divide between evidential decision theory (using prior evidence of accuracy) and causal decision theory (only considering what you can actually cause), both of which are logically valid but lead to opposite conclusions.
One-boxers consistently walk away wealthier, yet two-boxers argue they made the rational choice — revealing that rational acts don't always lead to rational outcomes, and a rational society might be made up of people who choose «irrational» strategies.
The only way to guarantee maximum winnings ($1,001,000) would be to convince the predictor you're a one-boxer, then two-box at the last second — but if the predictor is perfect, this becomes impossible, raising questions about whether free will exists at all.
Real-world parallels like mutually assured destruction during the Cold War show that pre-committing to a worse option (nuclear retaliation that kills everyone) can produce the best outcome (deterring attack entirely), suggesting rationality is about the rules you live by, not individual choices.
Вкратце
Newcomb's paradox reveals that sometimes being a rational person requires pre-committing to seemingly irrational acts — and that the question isn't what to choose in the moment, but what kind of person you've already decided to be.
The Setup: A Choice Between Certainty and Prediction
A supercomputer offers two boxes with a twist that divides people perfectly in half.
You enter a room with a supercomputer and two boxes. One is open, containing a visible $1,000. The other is sealed. The supercomputer has correctly predicted thousands of people's choices in this exact scenario. You can take both boxes or only the mystery box. Here's the twist: before you walked in, the computer predicted your choice. If it predicted you'd take only the mystery box, it put $1 million inside. If it predicted you'd take both, the mystery box is empty. The boxes are already set up. The computer isn't trying to trick you — its only goal is accurate prediction.
When philosopher Robert Nozick introduced this problem, he noted that «to almost everyone, it is perfectly clear and obvious what should be done. The difficulty is that these people seem to divide almost evenly on the problem, with large numbers thinking that the opposite half is just being silly.» A 2016 Guardian poll of over 31,000 people found 53.5% were one-boxers (mystery box only) and 46.5% were two-boxers (both boxes). The Veritasium audience poll of 24,000 responses skewed even more heavily: two-thirds chose one box. Yet each camp is utterly convinced the other is making an obvious mistake.
The One-Boxer Argument: Trust the Evidence
The Two-Boxer Argument: The Boxes Are Already Set
Why Ain'cha Rich?
One-boxers walk away millionaires; two-boxers get $1,000 — yet both claim rationality.
“If someone is very good at predicting behavior and rewards predicted irrationality richly, then irrationality will be richly rewarded.”
The Hidden Fork: Rationality at Different Levels
Rational acts don't always produce rational outcomes; rational societies may require irrational individuals.
Philosophers Gibbard and Harper admitted that two-boxers fare worse, but argued they still made the rational choice — the game is simply rigged to reward irrationality. But this reveals something profound: sometimes being a rational person requires acting irrationally. The parallel with the prisoner's dilemma is instructive. In a single round, the rational act is to defect. But a rational society is full of cooperators. Rationality at one level isn't compatible with rationality at another.
This creates three scenarios where a committed two-boxer would switch to one-boxing: first, if your choice now can actually change the past (like a wormhole to the future). Second, if there are multiple trials where reputation matters. Third, if you can pre-commit before the prediction is made. In all three cases, the commitment itself changes the structure of the game. The question shifts from «what should I choose now?» to «what kind of person should I be?»
Mutually Assured Destruction: When Worse Is Better
Cold War nuclear strategy proves pre-committing to a terrible option can prevent disaster.
On August 29, 1949, the Soviet Union detonated its first nuclear weapon, launching an arms race that left the US with over 30,000 warheads by the mid-1960s. US Secretary of Defense Robert McNamara advocated «mutually assured destruction» — the commitment to retaliate with total annihilation if attacked. This commitment is what kept the peace. But imagine you're the US president and you've just received confirmed word that Soviet missiles have launched. If you retaliate, everyone dies — possibly triggering a nuclear winter that kills nearly everyone on Earth.
The rational act in the moment is not to launch. But the rational pre-commitment is to always launch, because that's what prevents the attack in the first place. As depicted in «Dr. Strangelove,» the Soviets built a doomsday device that automatically triggers if attacked or tampered with — including tampering by the Russians themselves. The whole point is to remove the option of changing your mind. In game theory terms, you visibly throw your steering wheel out the window in a game of chicken, forcing the other driver to swerve. Pre-commitment to a worse option produces the best outcome.
The Answer: Be the Kind of Person Who One-Boxes
Rationality isn't about individual choices but the rules you commit to living by.
The Answer: Be the Kind of Person Who One-Boxes
The real question isn't what to choose in the moment, but what rules you would wire yourself to follow if you could program your own behavior. The best strategy is to become the kind of person who honors ideal pre-commitments — even ones you never explicitly made. When you face Newcomb's problem, ask: «If I could have pre-committed, what would the best pre-commitment have been?» The answer is to be a one-boxer. And if you've already wired yourself to honor such commitments, you're already committed to one-boxing before you walk into the room. Life is iterated. Reputation matters. Building yourself into someone who sticks to ideal commitments pays dividends far beyond a single thought experiment.
Data Points from the Paradox
Key numbers from the problem and its real-world applications.
Люди
Глоссарий
Отказ от ответственности: Это ИИ-сводка видео с YouTube, подготовленная в образовательных и справочных целях. Она не является инвестиционной, финансовой или юридической консультацией. Всегда проверяйте информацию по первоисточникам перед принятием решений. TubeReads не связан с автором контента.