Prisoner’s dilemma: Difference between revisions
Amwelladmin (talk | contribs) No edit summary |
Amwelladmin (talk | contribs) No edit summary |
||
(12 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
{{a| | {{a|bi|{{prisonersdilemmatable}}}}The [[prisoner’s dilemma]] is the mathematician’s way of articulating the [[commercial imperative]]. | ||
An exercise in calculating economic outcomes by means of [[metaphor]], the [[prisoner’s dilemma]] was developed at the RAND corporation in the 1950s by those splendid brainboxes as a way of predicting individuals’ behaviour in situations requiring [[I believe|trust]] among strangers - for very good example, when unacquainted participants buy or sell in an unregulated market. This field developed into [[game theory]]. | |||
==The original prisoner’s dilemma== | |||
Two people are charged with a conspiracy<ref>Whether or not they are guilty is beside the point. If it helps you empathise with their predicament, assume they’re innocent</ref>. Each is held separately. They cannot communicate. There is enough evidence to convict both on a lesser charge, but not the main charge. Each prisoner is separately offered the same plea bargain. The offer is: | Two people are charged with a conspiracy<ref>Whether or not they are guilty is beside the point. If it helps you empathise with their predicament, assume they’re innocent</ref>. Each is held separately. They cannot communicate. There is enough evidence to convict both on a lesser charge, but not the main charge. Each prisoner is separately offered the same plea bargain. The offer is: | ||
*If A informs B but B refuses to inform on A: | *If A informs B but B refuses to inform on A: | ||
Line 15: | Line 15: | ||
**B will get 1 year in prison (on the lesser charge). | **B will get 1 year in prison (on the lesser charge). | ||
Another way of looking at this, for those easily triggered, is as the [[eBayer’s dilemma]]. | |||
===Single round=== | ===[[Single round prisoner’s dilemma]]=== | ||
If you play | If you play the game once, and in isolation — with someone you don’t know and whom you do not expect to meet again, the payoff is grim: those who cooperate will get ''reamed''. Cooperation is a bad strategy. Your best interest is in defecting on the other guy, because ''his'' best interest is defecting on ''you''. | ||
This looks like a bad outcome for commerce | This looks like a bad outcome for commerce: if [[homo economicus]] ''should'' weasel on every deal, how can we have any faith in the market? How, come to think of it, has any kind of market ever got off the ground? Why would anyone take on a sure fire losing bet? Is this the smoking gun that homo economicus doesn’t exist?<ref>No. [[Homo economicus]] ''doesn’t'' exist, but this is not the reason why.</ref> | ||
Because trust, faith and confidence changes everything. The single round prisoner’s dilemma stipulates there is ''no consequence'' on a bad actor for reneging. The defector is guaranteed to get away with it: these are the rules. | Because [[trust]], faith and confidence changes everything. The [[single round prisoner’s dilemma]] stipulates there is ''no consequence'' on a bad actor for reneging. The defector is guaranteed to get away with it: these are the rules. | ||
But in real life, one-off interactions with strangers — counterparts whom you are | But in real life, one-off interactions with strangers — counterparts whom you are certain never to see again — are extremely rare, especially in our interconnected age. Business is the process of cultivating relationships. Establishing [[trust]]. | ||
The game theorists found an easy way to replicate that concept of trust: run the same game again. Repeatedly. An indefinite amount of times. | The game theorists found an easy way to replicate that concept of trust: run the same game again. Repeatedly. An indefinite amount of times. This is the [[iterated prisoner’s dilemma]]. | ||
=== | ===[[Iterated prisoner’s dilemma]]=== | ||
The same actors get to observe how each other act, and respond accordingly. If your counterpart defects, you have a means of retaliating: by defecting on the next game, or by refusing to play the game any more with that counterparty. | The same actors get to observe how each other act, and respond accordingly. If your counterpart defects, you have a means of retaliating: by defecting on the next game, or by refusing to play the game any more with that counterparty. | ||
Now, as well as the short-term payoff, there is a longer-term payoff, and it ''dwarfs'' the short term payoff. If I defect once, I earn £150. If I cooperate a thousand times, I earn £50,000. If I defect first time round, sure: I am £100 up, but at what cost: if my counterparty refuses to play with me again — and if she tells other players in the market — I will struggle to make much money. ''No one will trust me''. | Now, as well as the short-term payoff, there is a longer-term payoff, and it ''dwarfs'' the short term payoff. If I defect once, I earn £150. If I cooperate a thousand times, I earn £50,000. If I defect first time round, sure: I am £100 up, but at what cost: if my counterparty refuses to play with me again — and if she tells other players in the market — I will struggle to make much money. ''No one will trust me''. | ||
{{ | ===''One prisoner, or ''two''?=== | ||
In our view, one correctly places the apostrophe to designate a ''single'' prisoner having the dilemma — so [[prisoner’s dilemma]], not ''prisoners’'' dilemma — because both prisoners if taken together ''have'' no dilemma: it is only where they are acting individually that they have a problem ... ''with each other''. | |||
{{sa}} | |||
*The [[tragedy of the commons]] | |||
*[[Agency problem]] | |||
*[[I believe]] | *[[I believe]] | ||
{{ref}} | {{ref}} | ||
{{draft}} | {{draft}} |
Latest revision as of 13:30, 14 August 2024
The JC’S favourite Big Ideas™
|
The prisoner’s dilemma is the mathematician’s way of articulating the commercial imperative.
An exercise in calculating economic outcomes by means of metaphor, the prisoner’s dilemma was developed at the RAND corporation in the 1950s by those splendid brainboxes as a way of predicting individuals’ behaviour in situations requiring trust among strangers - for very good example, when unacquainted participants buy or sell in an unregulated market. This field developed into game theory.
The original prisoner’s dilemma
Two people are charged with a conspiracy[1]. Each is held separately. They cannot communicate. There is enough evidence to convict both on a lesser charge, but not the main charge. Each prisoner is separately offered the same plea bargain. The offer is:
- If A informs B but B refuses to inform on A:
- A will not be prosecuted at all and will go free
- B will be convicted of the main charge and will get 3 years in prison.
- If A informs B and B informs on A:
- A will get 2 years in prison
- B will get 2 years in prison
- If A refuses to inform on B and B refuses to inform on A:
- A will get 1 year in prison (on the lesser charge).
- B will get 1 year in prison (on the lesser charge).
Another way of looking at this, for those easily triggered, is as the eBayer’s dilemma.
Single round prisoner’s dilemma
If you play the game once, and in isolation — with someone you don’t know and whom you do not expect to meet again, the payoff is grim: those who cooperate will get reamed. Cooperation is a bad strategy. Your best interest is in defecting on the other guy, because his best interest is defecting on you.
This looks like a bad outcome for commerce: if homo economicus should weasel on every deal, how can we have any faith in the market? How, come to think of it, has any kind of market ever got off the ground? Why would anyone take on a sure fire losing bet? Is this the smoking gun that homo economicus doesn’t exist?[2]
Because trust, faith and confidence changes everything. The single round prisoner’s dilemma stipulates there is no consequence on a bad actor for reneging. The defector is guaranteed to get away with it: these are the rules.
But in real life, one-off interactions with strangers — counterparts whom you are certain never to see again — are extremely rare, especially in our interconnected age. Business is the process of cultivating relationships. Establishing trust.
The game theorists found an easy way to replicate that concept of trust: run the same game again. Repeatedly. An indefinite amount of times. This is the iterated prisoner’s dilemma.
Iterated prisoner’s dilemma
The same actors get to observe how each other act, and respond accordingly. If your counterpart defects, you have a means of retaliating: by defecting on the next game, or by refusing to play the game any more with that counterparty.
Now, as well as the short-term payoff, there is a longer-term payoff, and it dwarfs the short term payoff. If I defect once, I earn £150. If I cooperate a thousand times, I earn £50,000. If I defect first time round, sure: I am £100 up, but at what cost: if my counterparty refuses to play with me again — and if she tells other players in the market — I will struggle to make much money. No one will trust me.
One prisoner, or two?
In our view, one correctly places the apostrophe to designate a single prisoner having the dilemma — so prisoner’s dilemma, not prisoners’ dilemma — because both prisoners if taken together have no dilemma: it is only where they are acting individually that they have a problem ... with each other.
See also
References
- ↑ Whether or not they are guilty is beside the point. If it helps you empathise with their predicament, assume they’re innocent
- ↑ No. Homo economicus doesn’t exist, but this is not the reason why.