TubeReads

On This Problem Rational People Do Worse

A supercomputer that predicts human choices with near-perfect accuracy offers you a simple choice: take both boxes on the table — one containing $1,000 and one mystery box — or take only the mystery box. If the computer predicted you'd take only the mystery box, it put $1 million inside. If it predicted you'd take both, the mystery box is empty. The catch? The computer has already made its prediction and set up the boxes before you walked in the room. This seemingly simple problem divides people almost exactly in half, with each side convinced the other is obviously wrong. What makes it stranger is that the «rational» choice might actually make you poorer.

Duração do vídeo: 25:40·Publicado 9 de mar. de 2026·Idioma do vídeo: English
6–7 min de leitura·4,725 palavras faladasresumido para 1,395 palavras (3x)·

1

Pontos-chave

1

Two-thirds of people choose only the mystery box (one-boxers), trusting the supercomputer's track record, while one-third take both boxes (two-boxers), reasoning that the boxes are already set and their choice can't change the past.

2

The paradox exposes a fundamental divide between evidential decision theory (using prior evidence of accuracy) and causal decision theory (only considering what you can actually cause), both of which are logically valid but lead to opposite conclusions.

3

One-boxers consistently walk away wealthier, yet two-boxers argue they made the rational choice — revealing that rational acts don't always lead to rational outcomes, and a rational society might be made up of people who choose «irrational» strategies.

4

The only way to guarantee maximum winnings ($1,001,000) would be to convince the predictor you're a one-boxer, then two-box at the last second — but if the predictor is perfect, this becomes impossible, raising questions about whether free will exists at all.

5

Real-world parallels like mutually assured destruction during the Cold War show that pre-committing to a worse option (nuclear retaliation that kills everyone) can produce the best outcome (deterring attack entirely), suggesting rationality is about the rules you live by, not individual choices.

Em resumo

Newcomb's paradox reveals that sometimes being a rational person requires pre-committing to seemingly irrational acts — and that the question isn't what to choose in the moment, but what kind of person you've already decided to be.


2

The Setup: A Choice Between Certainty and Prediction

A supercomputer offers two boxes with a twist that divides people perfectly in half.

You enter a room with a supercomputer and two boxes. One is open, containing a visible $1,000. The other is sealed. The supercomputer has correctly predicted thousands of people's choices in this exact scenario. You can take both boxes or only the mystery box. Here's the twist: before you walked in, the computer predicted your choice. If it predicted you'd take only the mystery box, it put $1 million inside. If it predicted you'd take both, the mystery box is empty. The boxes are already set up. The computer isn't trying to trick you — its only goal is accurate prediction.

When philosopher Robert Nozick introduced this problem, he noted that «to almost everyone, it is perfectly clear and obvious what should be done. The difficulty is that these people seem to divide almost evenly on the problem, with large numbers thinking that the opposite half is just being silly.» A 2016 Guardian poll of over 31,000 people found 53.5% were one-boxers (mystery box only) and 46.5% were two-boxers (both boxes). The Veritasium audience poll of 24,000 responses skewed even more heavily: two-thirds chose one box. Yet each camp is utterly convinced the other is making an obvious mistake.


3

The One-Boxer Argument: Trust the Evidence

🎲
Evidential Decision Theory
One-boxers calculate expected utility using the probability C that the computer predicted correctly. If C is greater than 50.05%, one-boxing yields higher expected value. Since the computer has been accurate thousands of times, C is very high.
💰
The Track Record Matters
Basically everyone who took one box walked away a millionaire, and everyone who took both walked away with only $1,000. This prior evidence is powerful — why would you be the exception?
🧠
Predictable = Profitable
If you convince yourself you're a one-boxer, the machine will have predicted that. The very thought process that leads you to one-box is what causes the million to be there.

4

The Two-Boxer Argument: The Boxes Are Already Set

🔒
Causal Decision Theory
The supercomputer made its prediction before you entered. The boxes are set. Whatever you decide now cannot change the past or alter what's in the mystery box.
📊
Strategic Dominance
No matter what's in the mystery box — $0 or $1 million — you're always $1,000 better off by taking both boxes. This dominant strategy always wins in a direct comparison.
🤔
Wishful Thinking
Two-boxers argue that one-boxers are engaging in magical thinking: believing that your current decision can retroactively influence what the computer put in the box earlier.

5

Why Ain'cha Rich?

One-boxers walk away millionaires; two-boxers get $1,000 — yet both claim rationality.

If someone is very good at predicting behavior and rewards predicted irrationality richly, then irrationality will be richly rewarded.

Gibbard and Harper, 1978


6

The Hidden Fork: Rationality at Different Levels

Rational acts don't always produce rational outcomes; rational societies may require irrational individuals.

Philosophers Gibbard and Harper admitted that two-boxers fare worse, but argued they still made the rational choice — the game is simply rigged to reward irrationality. But this reveals something profound: sometimes being a rational person requires acting irrationally. The parallel with the prisoner's dilemma is instructive. In a single round, the rational act is to defect. But a rational society is full of cooperators. Rationality at one level isn't compatible with rationality at another.

This creates three scenarios where a committed two-boxer would switch to one-boxing: first, if your choice now can actually change the past (like a wormhole to the future). Second, if there are multiple trials where reputation matters. Third, if you can pre-commit before the prediction is made. In all three cases, the commitment itself changes the structure of the game. The question shifts from «what should I choose now?» to «what kind of person should I be?»


7

Mutually Assured Destruction: When Worse Is Better

Cold War nuclear strategy proves pre-committing to a terrible option can prevent disaster.

On August 29, 1949, the Soviet Union detonated its first nuclear weapon, launching an arms race that left the US with over 30,000 warheads by the mid-1960s. US Secretary of Defense Robert McNamara advocated «mutually assured destruction» — the commitment to retaliate with total annihilation if attacked. This commitment is what kept the peace. But imagine you're the US president and you've just received confirmed word that Soviet missiles have launched. If you retaliate, everyone dies — possibly triggering a nuclear winter that kills nearly everyone on Earth.

The rational act in the moment is not to launch. But the rational pre-commitment is to always launch, because that's what prevents the attack in the first place. As depicted in «Dr. Strangelove,» the Soviets built a doomsday device that automatically triggers if attacked or tampered with — including tampering by the Russians themselves. The whole point is to remove the option of changing your mind. In game theory terms, you visibly throw your steering wheel out the window in a game of chicken, forcing the other driver to swerve. Pre-commitment to a worse option produces the best outcome.


8

The Answer: Be the Kind of Person Who One-Boxes

Rationality isn't about individual choices but the rules you commit to living by.

💡

The Answer: Be the Kind of Person Who One-Boxes

The real question isn't what to choose in the moment, but what rules you would wire yourself to follow if you could program your own behavior. The best strategy is to become the kind of person who honors ideal pre-commitments — even ones you never explicitly made. When you face Newcomb's problem, ask: «If I could have pre-committed, what would the best pre-commitment have been?» The answer is to be a one-boxer. And if you've already wired yourself to honor such commitments, you're already committed to one-boxing before you walk into the room. Life is iterated. Reputation matters. Building yourself into someone who sticks to ideal commitments pays dividends far beyond a single thought experiment.


9

Data Points from the Paradox

Key numbers from the problem and its real-world applications.

Guardian Poll (2016) — One-Boxers
53.5%
Over 31,000 people were surveyed on Newcomb's paradox.
Guardian Poll (2016) — Two-Boxers
46.5%
Nearly an even split, showing how perfectly divisive the problem is.
Veritasium Audience Poll — One-Boxers
66.7%
Over 24,000 responses; two-thirds chose only the mystery box.
Critical Accuracy Threshold
50.05%
If the supercomputer is more accurate than this, one-boxing has higher expected utility.
US Nuclear Warheads (mid-1960s)
30,000+
Peak of the Cold War arms race, enabling mutually assured destruction strategy.
USSR Nuclear Warheads (mid-1960s)
6,000+
Both superpowers had more than enough to annihilate each other.

10

Pessoas

Derek
Host
host
Casper
Co-host/Interviewer
host
Gregor
Co-host/Interviewer
host
Henry
Interview Subject
guest
Robert Nozick
American Philosopher
mentioned
William Newcomb
Inventor of Newcomb's Paradox
mentioned
Robert McNamara
US Secretary of Defense (1960s)
mentioned
Gibbard and Harper
Philosophers
mentioned

Glossário
Evidential Decision TheoryA decision-making framework that uses prior evidence and correlations to calculate expected utility, even if those correlations aren't causal.
Causal Decision TheoryA decision-making framework that only considers outcomes you can actually cause through your actions, ignoring non-causal correlations.
Strategic DominanceA strategy that always performs better than alternatives regardless of what the other party does.
Expected UtilityThe weighted average of all possible outcomes of a decision, calculated by multiplying each outcome's value by its probability.
Mutually Assured Destruction (MAD)Cold War nuclear deterrence strategy based on the certainty that any attack would trigger total retaliation, ensuring both sides' destruction.

Aviso: Este é um resumo gerado por IA de um vídeo do YouTube para fins educacionais e de referência. Não constitui aconselhamento de investimento, financeiro ou jurídico. Verifique sempre as informações com as fontes originais antes de tomar decisões. O TubeReads não é afiliado ao criador do conteúdo.