How Repeated Games Build Cooperation and Stop Defection
The Shadow of the Future: Repeated Games and Cooperation
Why Repetition Changes Everything
The prisoner's dilemma looks like a trap with no escape — rational players locked into mutual defection. But there's a way out, and it's deceptively simple: play again. If Ada and Bob know they'll interact tomorrow, and the day after, and the day after that, everything shifts. Defecting today might feel good right now, but it burns the bridge for tomorrow. The possibility of future retaliation — or future cooperation — starts to matter. A lot.
Game theorists call this the shadow of the future — how much your future payoffs influence what you do right now. When that shadow is long (you'll keep playing this game for years), cooperation can become rational even for selfish players. When it's short (this is the last time we're meeting), defection comes roaring back.
Here's a practical way to think about it: a taxi driver visiting a city they'll never return to has no real incentive to treat customers fairly. They can overcharge, take the scenic route, behave badly. The driver who works the same neighborhood for twenty years? Totally different calculation. One bad review spreads. Reputation matters. They'll cooperate not because they're nice people, but because the math works out.
Formally, in an infinitely repeated game — or even a finitely repeated game where players can't see the end coming — something remarkable becomes possible. The Folk Theorem, one of game theory's most important results, says that basically any reasonable outcome can be a Nash equilibrium if players are patient enough. Cooperation. Defection. Almost anything. On one hand, it's inspiring: cooperation isn't impossible, it's inevitable under the right conditions. On the other hand, it's frustratingly vague: if almost anything could work, which outcome actually does emerge? That's where psychology, institutions, and how societies actually organize themselves come in.
Measuring the Shadow: Discount Factors and Patience
To pin down the shadow of the future mathematically, game theorists use something called a discount factor, written as δ (delta). It's a number between 0 and 1 that answers one simple question: "How much do I care about money next round compared to money today?"
- If δ = 0.9, next round's payoff is worth 90% of today's.
- If δ = 0.5, it's worth only 50%.
- If δ = 1, you're infinitely patient — tomorrow matters as much as today.
The shadow of the future is literally the discount factor. High δ (close to 1) means a long shadow — the future weighs heavy on your decisions. Low δ (close to 0) means a short shadow — you care almost only about what happens right now.
Here's what this reveals: cooperation demands patience. An impatient person — someone with a low discount factor — will struggle to cooperate in repeated games. The threat of future punishment just doesn't scare them enough to offset the immediate payoff of defecting. Real-world example: a start-up founder desperate to hit quarterly targets might exploit customers in ways a century-old company never would. Both are "rational," but they're operating on different timescales. The founder's discount factor is in free fall; the established company's is high and stable.
Axelrod's Famous Tournament
In the early 1980s, political scientist Robert Axelrod orchestrated what might be the most elegant experiment in social science history. He sent out a call to game theorists, economists, and psychologists: submit a computer program to play repeated prisoner's dilemma. Each program would face every other program for 200 rounds. Whoever scored the most points won.
Fourteen entries arrived. Some were sophisticated — programs packed with algorithms designed to detect weaknesses, exploit them, use randomness to stay unpredictable. Some tried to model what their opponent was doing and stay one step ahead. A few were just... weird.
The winner was the simplest entry of all: Tit-for-Tat, submitted by psychologist Anatol Rapoport. Here's the entire strategy:
- On the first move, cooperate.
- After that, copy whatever your opponent just did.
Done. That's it.
If they cooperated, you cooperate. If they stabbed you in the back, you stab them back once. Then you forgive and offer cooperation again. Why did this beat programs orders of magnitude more complex? Because complexity is a liability. A clever program playing against another clever program can accidentally lock both of them into mutual defection — neither can escape because they're both trying to outthink each other. Tit-for-Tat sidesteps this trap. It never starts a fight, and it's always the first to make peace.
Axelrod ran a second tournament, this time broadcasting Tit-for-Tat's victory. Now researchers knew what to beat. Sixty-two entries poured in, many specifically engineered to exploit Tit-for-Tat. Some figured they could take advantage of its niceness by occasionally defecting, betting it would forgive and they could keep milking it. Others tried to detect its behavior and outsmart it. Tit-for-Tat won again.
Why Tit-for-Tat Works: Four Essential Properties
When Axelrod dug into what made Tit-for-Tat so bulletproof, he found four qualities that separated winners from losers:
-
Nice: Never defect first. Offer cooperation right away and give others the chance to cooperate too. Niceness prevents unnecessary conflict and lets relationships start on a positive note.
-
Retaliatory: Don't let people walk over you. When someone defects, you respond immediately. Retaliation teaches opponents that betrayal has consequences — they learn not to exploit you.
-
Forgiving: Don't hold eternal grudges. After you punish a defection, go back to cooperation the next turn. Forgiveness breaks the death spiral where both players just keep punishing each other forever and never recover.
- Clear: Make your behavior obvious. Be predictable. When someone figures out they're dealing with a Tit-for-Tat player, they understand the deal and can plan accordingly. Clarity lets good reputations develop because people can count on you.
These four properties interlocking create something like a social contract. Niceness stops unnecessary fighting. Retaliation prevents people from freeloading. Forgiveness prevents permanent breakdown. Clarity lets information about trustworthy partners spread.
Here's the wild part: anthropologists studying human societies have found these same four properties in cultures all over the world. Small communities that cooperate well tend to welcome newcomers (nice), punish people who cheat the system (retaliatory), don't exile someone permanently after one mistake (forgiving), and maintain clear rules everyone understands (clear). We might've stumbled onto something fundamental about how humans cooperate.
Axelrod's Broader Insights: Live and Let Live in the Trenches
In The Evolution of Cooperation (1984), Axelrod applied his tournament insights to real history. One of his most striking examples: World War I trenches and a phenomenon called "live and let live."
Soldiers staring at enemy soldiers across no man's land were ordered to kill. But in many sectors, something different happened. Informal truces emerged. Troops fired at inconvenient times and places where bullets were unlikely to connect. They avoided areas where the other side slept. A rough, unspoken equilibrium developed: we won't really try to kill you if you don't really try to kill us.
Officers were baffled. But the soldiers understood the game perfectly. If I try to kill your guys, you'll try to kill my guys, and we all die for nothing. Better to maintain the status quo.
This was Tit-for-Tat in the trenches. The shadow of the future was long — soldiers would be stuck in the same trench for months or years. Defection would trigger retaliation. Mutual restraint was the rational move. But when units rotated, when new soldiers who didn't know about the truce arrived, the arrangement collapsed. The newcomers didn't understand the reputation equilibrium and started shooting to kill. Or they got shot at by an opposing force testing whether the truce still held. No shared history, no cooperation.
When Cooperation Breaks Down: The Preconditions for Repeated-Game Cooperation
Repetition helps a lot, but it's not a magic wand. For cooperation to actually emerge through repeated interaction, you need four things:
Recognizability: You have to know who you're dealing with. Anonymous interactions kill cooperation because reputation can't form. If you'll never see someone again and they'll never know who you are, the shadow of the future disappears. This is why online anonymity tends to bring out terrible behavior — no consequences, no reputation, no future to worry about. Small towns where everyone knows everyone's history? Those places sustain cooperation naturally.
Memory: People have to remember what happened before. A defection only deters future defection if it's remembered. Same with cooperation — if people forget you helped them, the incentive to help them again weakens. Gossip, historical records, online reviews: these all extend memory beyond what any one person can hold in their head.
Sufficient patience: Players need to care enough about the future. If someone's discount factor is too low — they're desperate for money right now — they'll steal, knowing prison follows, because the immediate gain feels bigger than the future loss. Someone with stable prospects and a long time horizon wouldn't dream of it. The future loss looms too large.
Some possibility of future interaction: If you know for absolute certain this is the last time you'll see someone, cooperation becomes irrational even in games that are technically "repeated." This leads us to a nasty puzzle.
The End-Game Problem: Backward Induction Unravels Cooperation
Here's where things get uncomfortable. Suppose Ada and Bob play prisoner's dilemma for exactly 10 rounds. On round 10 — the last round — Ada knows there's no future. The shadow disappears entirely. She should defect (earning 5 instead of 3).
Knowing this, Bob realizes Ada will defect on round 10. So on round 9, Bob should defect too (why cooperate if you're about to get stabbed?). By the same logic, Ada should defect on round 8. And if Ada defects on 8, Bob should defect on 7. Keep working backward.
This process, called backward induction, unravels cooperation all the way back to round 1. Theoretically, a finite game with a known endpoint should produce zero cooperation — it's equivalent to playing once.
Yet in real life, people cooperate in finitely repeated prisoner's dilemmas even when they know exactly when the game ends. This gap between what theory predicts and what people actually do is a huge red flag. It tells us that pure rationality — the assumption that people are perfectly logical and only care about themselves — doesn't explain human behavior. People care about fairness. They struggle with deep chains of logical reasoning. They cooperate out of habit or moral intuition even when the numbers say they shouldn't.
This tension between theory and practice becomes even sharper when you look at things like the Ultimatum Game, which we'll tackle in detail later. That's where fairness, morality, and what people actually value start crashing into cold rationality. It's a reminder that game theory is powerful, but incomplete without psychology.
Only visible to you
Sign in to take notes.