You Should Betray Your Friend. Here's Why You Probably Won't.
It's 1950. Two men are arrested on the outskirts of town on suspicion of a serious crime. The police have only enough evidence for a minor charge, but they suspect the men are guilty of something much bigger.
So they split them up. Put them in separate rooms. No phones, no lawyers, no way to communicate.
Then a detective walks into each room and lays out the same deal.
"Here's what happens. If you stay quiet and your partner also stays quiet, we charge you both with the minor offense: one year each. But if you talk and he doesn't, you go free. He gets ten years. If he talks and you don't, he walks and you get ten. And if you both talk... five years each."
The detective closes the door. The two men sit alone with their thoughts.
What should you do?
The Matrix
Before we answer that, let's lay it out. The payoff structure looks like this:
Every cell represents a pair of choices. Hover over them to understand what each outcome actually means.
Now ask yourself: what's the rational move?
Let's think through it. Suppose you have no idea what your partner will do.
If they cooperate (stay silent): you get 1 year if you also cooperate, or go free if you defect. Defecting is better.
If they defect (betray you): you get 10 years if you cooperate, or 5 years if you defect. Defecting is better.
No matter what your partner does, defecting gives you a better outcome. Defection is what game theorists call a dominant strategy: it beats every alternative regardless of the opponent's choice.
So both players, reasoning identically, choose to defect. And they both get 5 years.
Which is worse than if they had both stayed quiet.
That's the dilemma.
Why Rational Thinking Breaks Down
Two smart people, each making the individually correct decision, produce a collectively terrible outcome.
This isn't a quirk or a failure of logic. It's what happens when individual incentives are structurally misaligned with collective outcomes. Economists have a name for where both players defect: the Nash equilibrium, the point where neither player can improve their result by changing strategy alone. No one has an incentive to switch.
But there's another outcome, mutual cooperation, where both players are better off. Economists call that Pareto-optimal: no change exists that helps one person without hurting the other. The tragedy is that the Pareto-optimal outcome and the Nash equilibrium are different cells.
The game is rigged to pull everyone toward a worse outcome. Not because people are stupid or selfish, but because of the structure of the choices.
The Prisoner's Dilemma isn't really about prisoners. The prison scenario is a frame, a narrative wrapper. The real subject is the tension between individual rationality and collective welfare, which shows up everywhere.
The Experiment Nobody Talks About
Here's where history gets interesting.
Most people know the Prisoner's Dilemma as a thought experiment. Few know that in January 1950 (the same year the puzzle was named) two mathematicians actually ran it.
Merrill Flood and Melvin Dresher worked at RAND Corporation in Santa Monica, a think tank funded by the US Air Force to apply mathematics to military strategy. They were interested in game theory, which was still a young field. They decided to test what real people would actually do.
Their subjects were two colleagues: economist Armen Alchian and mathematician John Williams. The two men played 100 rounds of the game. They could see each other's previous moves as the rounds accumulated.
The result was not what theory predicted.
Instead of defecting every round (as a rational agent should), the players cooperated 60% of the time. Mutual defection (the Nash equilibrium) occurred only 14% of the time.
Williams, writing notes during the game, described his approach: "This is like toilet training a child, you have to be very patient." He was deliberately signaling cooperation, punishing defection, rewarding good behavior. He was, without knowing the name for it, playing Tit-for-Tat.
Alchian's notes across the rounds are revealing. Early on: "What is he doing?!!" By the end: a grudging mutual accommodation. Not friendship, but a working equilibrium.
Theory said they should defect. Reality said otherwise.
The name "Prisoner's Dilemma" came later, coined by mathematician Albert Tucker, who created the prison story to explain the Flood-Dresher experiment to an audience of Stanford psychologists. The colorful frame was meant to be a teaching aid. Instead, it eclipsed the original experiment in popular memory.
The Tournament
Thirty years passed. The dilemma remained mostly academic, a puzzle that demonstrated limits of rational-choice theory.
Then in 1980, a political science professor at the University of Michigan named Robert Axelrod had a different idea. What if we ran a tournament?
He invited game theorists from around the world to submit computer programs: strategies for playing the iterated Prisoner's Dilemma. Not once, but again and again, against every other strategy in a round-robin. Strategies would be ranked by total points accumulated.
The submissions came from mathematicians, psychologists, economists, and computer scientists. Some strategies were elaborate. Some ran hundreds of lines of code and tried to infer the opponent's behavior from statistical patterns.
The winner was submitted by Anatol Rapoport, a professor of mathematical psychology at the University of Toronto. His strategy was four lines long.
Cooperate on the first move. Then do whatever your opponent did last round.
That's Tit-for-Tat. Simple enough to explain in one sentence. It won.
Axelrod ran a second tournament after publishing the results of the first, giving everyone a chance to beat Tit-for-Tat knowing exactly how it worked.
It won again.
Why Tit-for-Tat Wins
Axelrod identified four properties that made it so robust:
- Nice: It never defects first. It doesn't start conflict.
- Retaliatory: It punishes defection immediately in the next round.
- Forgiving: As soon as the opponent cooperates, it returns to cooperating.
- Clear: Its behavior is simple and predictable. Opponents learn quickly what to expect.
None of these are moral claims. They're strategic ones. Being nice, retaliatory, forgiving, and clear is simply what works over repeated interactions.
Try It Yourself
Pick two strategies and run the simulation. Watch how cooperation emerges or collapses, round by round.
A few experiments worth trying:
- Tit-for-Tat vs. Always Defect: TFT punishes defection immediately but pays early. Over 20 rounds, who ends up ahead?
- Tit-for-Tat vs. Always Cooperate: both cooperate every round, but Always Cooperate earns slightly fewer points because it's exploitable.
- Grim Trigger vs. Tit-for-Tat: what happens when one slip triggers permanent retaliation?
The key insight from Axelrod's tournament: defectors do well in the short term. If you're playing a "nice" opponent, you can exploit them. But defectors suffer in the long run because they drive away cooperative partners, destroy trust, and eventually only face other defectors, which is a terrible situation for everyone.
The Real World
The Prisoner's Dilemma is a mathematical abstraction. But the logic it captures is everywhere.
The Arms Race
For four decades, the United States and the Soviet Union faced a classic Prisoner's Dilemma. Both countries would have been better off with fewer nuclear weapons: less spending, less risk, equal security. But neither side could unilaterally disarm. If you cut your weapons and they don't, you're vulnerable. So both kept building.
The result was tens of thousands of warheads (roughly 70,000 at the peak), stockpiled by two countries who both understood they'd be better off without them. Neither player defected from the arms race because the downside of being the only one to disarm was catastrophic.
What broke the stalemate wasn't a change in strategy. It was economics. The Soviet economy couldn't sustain the arms race. The shadow of the future changed.
Climate Change
The Paris Agreement is a global iterated Prisoner's Dilemma with 195 players.
Each country benefits from the others reducing emissions. Each country also bears the cost of reducing their own. Any individual nation could defect, pollute freely, let others pay for the transition, and gain a short-term economic edge.
The agreement works on voluntary commitments. There is no enforcement mechanism. Countries that defect face diplomatic pressure, not consequences. And yet many countries have, so far, not defected. Why?
Because climate negotiations happen repeatedly, in international forums, over years. Reputations accumulate. Trust is built or broken in public. The logic of iterated games gradually shapes behavior even without rules.
It's fragile. Every major emitter knows what every other emitter is doing. But it's also more robust than the one-shot game predicts.
Coca-Cola vs. Pepsi
Here's the same structure in boardroom form.
Imagine both companies are considering a price cut to steal market share.
| Pepsi keeps price | Pepsi cuts price | |
|---|---|---|
| Coke keeps price | $500M each | Coke: 750M |
| Coke cuts price | Coke: 250M | $250M each |
The Nash equilibrium is both cutting prices, a race to the bottom where both earn 500M. Individual rationality produces collective ruin.
This is why oligopolies behave so strangely. Companies sometimes maintain high prices without explicit coordination. Not because they're colluding (which would be illegal), but because repeat players in a game learn that mutual defection is self-defeating. The shadow of the future does the work.
This is also why price wars are so destructive. Once one player cuts, the other must respond. The resulting equilibrium leaves everyone worse off, and it can take years to rebuild the cooperative pricing that everyone preferred.
The Lesson
The Prisoner's Dilemma was invented to expose the limits of individual rationality. But Axelrod's tournament revealed something more optimistic buried inside it.
Cooperation isn't naive. It's what wins: over time, among repeated players, in a world where you'll meet each other again.
The math is unambiguous about when cooperation breaks down: when the game is played once, when players won't meet again, when there's no memory of past behavior, when the future doesn't matter. Defect when there's no tomorrow.
But most human problems don't work that way. Climate, trade, security, business: these are games that repeat. The same nations negotiate in Davos every year. The same companies compete in the same market for decades. The same neighbors live next to each other for years.
When the future casts a long enough shadow, cooperation becomes rational.
And Tit-for-Tat offers a surprisingly practical template for navigating it: Start by cooperating. Retaliate when crossed. Forgive quickly. Stay predictable.
It won the tournament. Twice.
That's not a guarantee of anything. But it's an interesting start.
What is so beautiful about Tit-for-Tat is that it is both firm and forgiving. It reacts to defection with defection, but it never holds a grudge longer than it needs to.
If this was worth sharing, send it to someone on 𝕏 or LinkedIn. Got a question or a thought? Drop me a message — I read everything. If this was worth your time, .