Prisoner’s dilemma: Difference between revisions

From The Jolly Contrarian
Jump to navigation Jump to search
No edit summary
No edit summary
 
(30 intermediate revisions by the same user not shown)
Line 1: Line 1:
An exercise in calculating economic outcomes by means of {{tag|metaphor}}.
{{a|bi|{{prisonersdilemmatable}}}}The [[prisoner’s dilemma]] is the mathematician’s way of articulating the [[commercial imperative]].


Two people are arrested and charged with a conspiracy<ref>Whether or not they are guilty is beside the point. If it helps you empathise with their predicament, assume they’re innocent</ref>.  Each is held separately in solitary confinement. The two cannot communicate. The authorities have enough evidence to convict both on a lesser charge, but not the main charge. The authorities go to each prisoner separately and offer each the same bargain if she agrees to inform on her co-conspirator. The offer is:
An exercise in calculating economic outcomes by means of [[metaphor]], the [[prisoner’s dilemma]] was developed at the RAND corporation in the 1950s by those splendid brainboxes as a way of predicting individuals’ behaviour in situations requiring [[I believe|trust]] among strangers - for very good example, when unacquainted participants buy or sell in an unregulated market. This field developed into [[game theory]].
 
==The original prisoner’s dilemma==
Two people are charged with a conspiracy<ref>Whether or not they are guilty is beside the point. If it helps you empathise with their predicament, assume they’re innocent</ref>.  Each is held separately. They cannot communicate. There  is enough evidence to convict both on a lesser charge, but not the main charge. Each prisoner is separately offered the same plea bargain. The offer is:
*If A informs B but B refuses to inform on A:
*If A informs B but B refuses to inform on A:
**A will not be prosecuted on the lesser charge and will go free  
**A will not be prosecuted at all and will go free  
**B will be prosecuted on the main charge and serve a '''three''' year sentence
**B will be convicted of the main charge and will get '''3''' years in prison.
*If A informs B ''and'' B informs on A:
*If A informs B ''and'' B informs on A:
**A will serve a two year sentence
**A will get 2 years in prison
**B will serve a two year sentence
**B will get 2 years in prison
*If A refuses to inform on B and B refuses to inform on A:
*If A refuses to inform on B and B refuses to inform on A:
**A will serve a one year sentence (on the lesser charge).
**A will get 1 year in prison  (on the lesser charge).
**B will serve a one year sentence (on the lesser charge).
**B will get 1 year in prison (on the lesser charge).
 
Another way of looking at this, for those easily triggered, is as the [[eBayer’s dilemma]].
 
===[[Single round prisoner’s dilemma]]===
If you play the game once, and in isolation — with someone you don’t know and whom you do not expect to meet again, the payoff is grim: those who cooperate will get ''reamed''. Cooperation is a bad strategy. Your best interest is in defecting on the other guy, because ''his'' best interest is defecting on ''you''.
 
This looks like a bad outcome for commerce: if [[homo economicus]] ''should'' weasel on every deal, how can we have any faith in the market? How, come to think of it, has any kind of market ever got off the ground? Why would anyone take on a sure fire losing bet? Is this the smoking gun that homo economicus doesn’t exist?<ref>No. [[Homo economicus]] ''doesn’t'' exist, but this is not the reason why.</ref>
 
Because [[trust]], faith and confidence changes everything. The [[single round prisoner’s dilemma]] stipulates there is ''no consequence'' on a bad actor for reneging. The defector is guaranteed to get away with it: these are the rules.
 
But in real life, one-off interactions with strangers — counterparts whom you are certain never to see again — are extremely rare, especially in our interconnected age. Business is the process of cultivating relationships. Establishing [[trust]].
 
The game theorists found an easy way to replicate that concept of trust: run the same game again. Repeatedly. An indefinite amount of times.  This is the [[iterated prisoner’s dilemma]].
 
===[[Iterated prisoner’s dilemma]]===
The same actors get to observe how each other act, and respond accordingly. If your counterpart defects, you have a means of retaliating: by defecting on the next game, or by refusing to play the game any more with that counterparty.
 
Now, as well as the short-term payoff, there is a longer-term payoff, and it ''dwarfs'' the short term payoff.  If I defect once, I earn £150. If I cooperate a thousand times, I earn £50,000. If I defect first time round, sure: I am £100 up, but at what cost: if my counterparty refuses to play with me again — and if she tells other players in the market — I will struggle to make much money. ''No one will trust me''.
 
===''One prisoner, or ''two''?===
In our view, one correctly places the apostrophe to designate a ''single'' prisoner having the dilemma — so [[prisoner’s dilemma]], not ''prisoners’'' dilemma — because both prisoners if taken together ''have'' no dilemma: it is only where they are acting individually that they have a problem ... ''with each other''.
{{sa}}
*The [[tragedy of the commons]]
*[[Agency problem]]
*[[I believe]]
{{ref}}
{{draft}}

Latest revision as of 13:30, 14 August 2024

The JC’S favourite Big Ideas™

Pay-off table

A cooperates

A defects

B cooperates

A gets 1 year
B gets 1 year

A goes free
B gets 3 years

B defects

A gets 3 years
B goes free

A gets 2 years
B gets 2 years

Index: Click to expand:
Tell me more
Sign up for our newsletter — or just get in touch: for ½ a weekly 🍺 you get to consult JC. Ask about it here.

The prisoner’s dilemma is the mathematician’s way of articulating the commercial imperative.

An exercise in calculating economic outcomes by means of metaphor, the prisoner’s dilemma was developed at the RAND corporation in the 1950s by those splendid brainboxes as a way of predicting individuals’ behaviour in situations requiring trust among strangers - for very good example, when unacquainted participants buy or sell in an unregulated market. This field developed into game theory.

The original prisoner’s dilemma

Two people are charged with a conspiracy[1]. Each is held separately. They cannot communicate. There is enough evidence to convict both on a lesser charge, but not the main charge. Each prisoner is separately offered the same plea bargain. The offer is:

  • If A informs B but B refuses to inform on A:
    • A will not be prosecuted at all and will go free
    • B will be convicted of the main charge and will get 3 years in prison.
  • If A informs B and B informs on A:
    • A will get 2 years in prison
    • B will get 2 years in prison
  • If A refuses to inform on B and B refuses to inform on A:
    • A will get 1 year in prison (on the lesser charge).
    • B will get 1 year in prison (on the lesser charge).

Another way of looking at this, for those easily triggered, is as the eBayer’s dilemma.

Single round prisoner’s dilemma

If you play the game once, and in isolation — with someone you don’t know and whom you do not expect to meet again, the payoff is grim: those who cooperate will get reamed. Cooperation is a bad strategy. Your best interest is in defecting on the other guy, because his best interest is defecting on you.

This looks like a bad outcome for commerce: if homo economicus should weasel on every deal, how can we have any faith in the market? How, come to think of it, has any kind of market ever got off the ground? Why would anyone take on a sure fire losing bet? Is this the smoking gun that homo economicus doesn’t exist?[2]

Because trust, faith and confidence changes everything. The single round prisoner’s dilemma stipulates there is no consequence on a bad actor for reneging. The defector is guaranteed to get away with it: these are the rules.

But in real life, one-off interactions with strangers — counterparts whom you are certain never to see again — are extremely rare, especially in our interconnected age. Business is the process of cultivating relationships. Establishing trust.

The game theorists found an easy way to replicate that concept of trust: run the same game again. Repeatedly. An indefinite amount of times. This is the iterated prisoner’s dilemma.

Iterated prisoner’s dilemma

The same actors get to observe how each other act, and respond accordingly. If your counterpart defects, you have a means of retaliating: by defecting on the next game, or by refusing to play the game any more with that counterparty.

Now, as well as the short-term payoff, there is a longer-term payoff, and it dwarfs the short term payoff. If I defect once, I earn £150. If I cooperate a thousand times, I earn £50,000. If I defect first time round, sure: I am £100 up, but at what cost: if my counterparty refuses to play with me again — and if she tells other players in the market — I will struggle to make much money. No one will trust me.

One prisoner, or two?

In our view, one correctly places the apostrophe to designate a single prisoner having the dilemma — so prisoner’s dilemma, not prisoners’ dilemma — because both prisoners if taken together have no dilemma: it is only where they are acting individually that they have a problem ... with each other.

See also

References

  1. Whether or not they are guilty is beside the point. If it helps you empathise with their predicament, assume they’re innocent
  2. No. Homo economicus doesn’t exist, but this is not the reason why.