#jsDisabledContent { display:none; } My Account | Register | Help

# Ultimatum game

Article Id: WHEBN0000801679
Reproduction Date:

 Title: Ultimatum game Author: World Heritage Encyclopedia Language: English Subject: Collection: Game Theory Publisher: World Heritage Encyclopedia Publication Date:

### Ultimatum game

Extensive form representation of a two proposal ultimatum game. Player 1 can offer a fair (F) or unfair (U) proposal; player 2 can accept (A) or reject (R).

The ultimatum game is a game in economic experiments. The first player (the proposer) receives a sum of money and proposes how to divide the sum between himself and another player. The second player (the responder) chooses to either accept or reject this proposal. If the second player accepts, the money is split according to the proposal. If the second player rejects, neither player receives any money. The game is typically played only once so that reciprocation is not an issue.

## Contents

• Equilibrium analysis 1
• Experimental results 2
• Explanations 3
• Neurological explanations 3.1
• Evolutionary game theory 3.2
• Sociological applications 4
• History 5
• Variants 6
• Notes 8
• References 9

## Equilibrium analysis

For illustration, we will suppose there is a smallest division of the good available (say 1 cent). Suppose that the total amount of money available is x.

The first player chooses some amount p he will keep for himself in the interval [0,x], the second player will then receive x-p. The second player chooses some function f: [0, x] → {"accept", "reject"} (i.e. the second chooses which divisions to accept and which to reject). We will represent the strategy profile as (p, f), where p is the proposal and f is the function. If f(p) = "accept" the first receives p and the second xp, otherwise both get zero.

(p, f(p)) is a Nash equilibrium of the ultimatum game if f(p) = "accept" and there is no y > p such that f(y) = "accept" (i.e. player 2 would reject all proposals in which player 1 receives more than p). The first player would not want to unilaterally increase his/her demand since the second would reject any higher demand. The second would not want to reject the demand, since he/she would then get nothing.

There is one other Nash equilibrium where p = x and f(y) = "reject" for all y>0 (i.e. the second rejects all demands that gives the first any amount at all). Here both players get nothing, but neither could get more by unilaterally changing his/her strategy.

However, only one of these Nash equilibria satisfies a more restrictive equilibrium concept, subgame perfection. Suppose that the first demands a large amount that gives the second some (small) amount of money. By rejecting the demand, the second is choosing nothing rather than something. So, it would be better for the second to choose to accept any demand that gives him/her any amount whatsoever. If the first knows this, he/she will give the second the smallest (non-zero) amount possible.[1]

## Experimental results

When carried out between members of a shared social group (e.g, a village, a tribe, a nation, humanity)[2] people offer "fair" (i.e., 50:50) splits, and offers of less than 30% are often rejected.[3]

One limited study on monozygotic and dizygotic twins claims that genetic variation can affect reactions to unfair offers, though the study failed to employ actual controls for environmental differences.[4] It has also been found that delaying the responder's decision makes people accept "unfair" offers more often.[5] Common chimpanzees behaved similarly to human by proposing fair offers in one version of the ultimatum game involving direct interaction between the chimpanzees. [6] However, another study also published in November 2012 showed that both kinds of chimpanzees, common chimpanzees and bonobos did not reject unfair offers, using a mechanical apparatus.[7] As of February 2015, bonobos have not been studied using the protocol involving direct interaction.

## Explanations

The highly mixed results (along with similar results in the Dictator game) have been taken to be both evidence for and against the so-called "Homo economicus" assumptions of rational, utility-maximizing, individual decisions. Since an individual who rejects a positive offer is choosing to get nothing rather than something, that individual must not be acting solely to maximize his economic gain, unless one incorporates economic applications of social, psychological, and methodological factors (such as the observer effect). Several attempts have been made to explain this behavior. Some suggest that individuals are maximizing their expected utility, but money does not translate directly into expected utility.[8] Perhaps individuals get some psychological benefit from engaging in punishment or receive some psychological harm from accepting a low offer. It could also be the case that the second player, by having the power to reject the offer, uses such power as leverage against the first player, thus motivating him to be fair.

The classical explanation of the ultimatum game as a well-formed experiment approximating general behaviour often leads to a conclusion that the rational behavior in assumption is accurate to a degree, but must encompass additional vectors of decision making. Behavioral economic and psychological accounts suggest that second players who reject offers less than 50% of the amount at stake do so for one of two reasons. An altruistic punishment account suggests that rejections occur out of altruism: people reject unfair offers to teach the first player a lesson and thereby reduce the likelihood that the player will make an unfair offer in the future. Thus, rejections are made to benefit the second player in the future, or other people in the future. By contrast, a self-control account suggests that rejections constitute a failure to inhibit a desire to punish the first player for making an unfair offer. Morewedge, Krishnamurti, and Ariely (2014) found that intoxicated participants were more likely to reject unfair offers than sober participants.[9] As intoxication tends to exacerbate decision makers' prepotent response, this result provides support for the self-control than the altruistic punishment account. Other research from social cognitive neuroscience supports this finding.[10]

However, several competing models suggest ways to bring the cultural preferences of the players within the optimized utility function of the players in such a way as to preserve the utility maximizing agent as a feature of microeconomics. For example, researchers have found that Mongolian proposers tend to offer even splits despite knowing that very unequal splits are almost always accepted.[11] Similar results from other small-scale societies players have led some researchers to conclude that "reputation" is seen as more important than any economic reward.[11] Others have proposed the social status of the responder may be part of the payoff.[12] Another way of integrating the conclusion with utility maximization is some form of inequity aversion model (preference for fairness). Even in anonymous one-shot settings, the economic-theory suggested outcome of minimum money transfer and acceptance is rejected by over 80% of the players.

An explanation which was originally quite popular was the "learning" model, in which it was hypothesized that proposers’ offers would decay towards the sub game perfect Nash equilibrium (almost zero) as they mastered the strategy of the game; this decay tends to be seen in other iterated games. However, this explanation (bounded rationality) is less commonly offered now, in light of subsequent empirical evidence.[13]

It has been hypothesised (e.g. by James Surowiecki) that very unequal allocations are rejected only because the absolute amount of the offer is low. The concept here is that if the amount to be split were ten million dollars a 90:10 split would probably be accepted rather than spurning a million dollar offer. Essentially, this explanation says that the absolute amount of the endowment is not significant enough to produce strategically optimal behaviour. However, many experiments have been performed where the amount offered was substantial: studies by Cameron and Hoffman et al. have found that higher stakes cause offers to approach closer to an even split, even in a 100 USD game played in Indonesia, where average per-capita income for is much lower than in United States. Rejections are reportedly independent of the stakes at this level, with 30 USD offers being turned down in Indonesia, as in the United States, even though this equates to two week's wages in Indonesia.[14]

### Neurological explanations

Generous offers in the ultimatum game (offers exceeding the minimum acceptable offer) are commonly made. Zak, Stanton & Ahmadi (2007)[15] showed that two factors can explain generous offers: empathy and perspective taking. They varied empathy by infusing participants with intranasal oxytocin or placebo (blinded). They affected perspective-taking by asking participants to make choices as both player 1 and player 2 in the ultimatum game, with later random assignment to one of these. Oxytocin increased generous offers by 80% relative to placebo. Oxytocin did not affect the minimum acceptance threshold or offers in the dictator game (meant to measure altruism). This indicates that emotions drive generosity.

Rejections in the ultimatum game have been shown to be caused by adverse physiologic reactions to stingy offers.[16] In a brain imaging experiment by Sanfey et al., stingy offers (relative to fair and hyperfair offers) differentially activated several brain areas, especially the anterior insular cortex, a region associated with visceral disgust. If Player 1 in the ultimatum game anticipates this response to a stingy offer, they may be more generous.

An increase in rational decisions in the game has been found among experienced Buddhist meditators. fMRI data show that meditators recruit the posterior insular cortex (associated with interoception) during unfair offers and show reduced activity in the anterior insular cortex compared to controls.[17]

People whose serotonin levels have been artificially lowered will reject unfair offers more often than players with normal serotonin levels.[18]

This is true whether the players are on placebo or are infused with a hormone that makes them more generous in the ultimatum game.[19][20]

People who have ventromedial frontal cortex lesions were found to be more likely to reject unfair offers.[21] This was suggested to be due to the abstractness and delay of the reward, rather than an increased emotional response to the unfairness of the offer.[22]

### Evolutionary game theory

Other authors have used evolutionary game theory to explain behavior in the ultimatum game.[23] Simple evolutionary models, e.g. the replicator dynamics, cannot account for the evolution of fair proposals or for rejections. These authors have attempted to provide increasingly complex models to explain fair behavior.

## Sociological applications

The ultimatum game is important from a sociological perspective, because it illustrates the human unwillingness to accept injustice. The tendency to refuse small offers may also be seen as relevant to the concept of honour.

The extent to which people are willing to tolerate different distributions of the reward from "cooperative" ventures results in inequality that is, measurably, exponential across the strata of management within large corporations. See also: Inequity aversion within companies.

Some see the implications of the ultimatum game as profoundly relevant to the relationship between society and the free market, with Prof. P.J. Hill, (Wheaton College, Illinois) saying:

I see the [ultimatum] game as simply providing counter evidence to the general presumption that participation in a market economy (capitalism) makes a person more selfish.[24]

## History

The first ultimatum game was developed in 1982 as a stylized representation of negotiation, by Güth, Schmittberger, and Schwarze.[25] It has since become a popular economic experiment, and was said to be "quickly catching up with the Prisoner's Dilemma as a prime showpiece of apparently irrational behavior" in a paper by Martin Nowak, Karen M. Page, and Karl Sigmund.[26]

## Variants

In the "competitive ultimatum game" there are many proposers and the responder can accept at most one of their offers: With more than three (naïve) proposers the responder is usually offered almost the entire endowment[27] (which would be the Nash Equilibrium assuming no collusion among proposers).

In the "ultimatum game with tipping", a tip is allowed from responder back to proposer, a feature of the trust game, and net splits tend to be more equitable.[28]

The "reverse ultimatum game" gives more power to the responder by giving the proposer the right to offer as many divisions of the endowment as they like. Now the game only ends when the responder accepts an offer or abandons the game, and therefore the proposer tends to receive slightly less than half of the initial endowment.[29]

The pirate game illustrates a variant with more than two participants with voting power, as illustrated in Ian Stewart's "A Puzzle for Pirates".[30]

## Notes

1. ^ Technically, making a zero offer to the responder and accepting this offer is also a Nash Equilibrium, as the responder's threat to reject the offer is no longer credible, since he/she now gains nothing (materially) by refusing the zero amount offered. Normally, when a player is indifferent between various strategies, the principle in Game Theory is that the strategy with an outcome which is Pareto optimally better for the other players is chosen (as a sort of tie-breaker to create a unique NE). However, it is generally assumed that this principle should not apply to ultimatum game players offered nothing; they are instead assumed to reject the offer although accepting it would be an equally subgame perfect NE. For instance, the University of Wisconsin summary: Testing Subgame Perfection Apart From Fairness in Ultimatum Games from 2002 admits the possibility that the proposer may offer nothing but qualifies the subgame perfect NE with the words (almost nothing) throughout the Introduction.
2. ^ Sanfey, Alan; Rilling, Aronson, Nystrom, Cohen (13 June 2003). "The Neural Basis of Economic Decision-Making in the Ultimatum Game". Science 300 (5626): 1755–1758.
3. ^ See Joseph Henrich et al. (2004) and Oosterbeek et al. (2004).
4. ^ http://www.pnas.org/content/105/10/3721.full.pdf+html
5. ^ See Grimm and Mengel (2011)
6. ^ Proctor, Darby; Williamson, de Waal, Brosnan (2013). "Chimpanzees play the ultimatum game". PNAS 110 (6): 2070–2075.
7. ^ Kaiser, Ingrid; Jensen, Call, Tomasello (2012). "Theft in an ultimatum game: chimpanzees and bonobos are insensitive to unfairness". Biology Letters 8 (6): 942–945.
8. ^ See Bolton (1991), and Ochs and Roth, A. E. (1989).
9. ^ Morewedge, Carey K.; Krishnamurti, Tamar; Ariely, Dan (2014-01-01). "Focused on fairness: Alcohol intoxication increases the costly rejection of inequitable rewards". Journal of Experimental Social Psychology 50: 15–20.
10. ^ Tabibnia, Golnaz; Satpute, Ajay B.; Lieberman, Matthew D. (2008-04-01). "The Sunny Side of Fairness Preference for Fairness Activates Reward Circuitry (and Disregarding Unfairness Activates Self-Control Circuitry)". Psychological Science 19 (4): 339–347.
11. ^ a b Mongolian/Kazakh study conclusion from University of Pennsylvania.
12. ^ Social Role in the Ultimate Game
13. ^ A forthcoming paper "On the Behavior of Proposers in Ultimatum Games" Journal of Economic Behaviour and Organization has the thesis that learning will not cause NE-convergence: see the abstract.
14. ^ See "Do higher stakes lead to more equilibrium play?" (page 18) in 3. Bargaining experiments, Professor Armin Falk's summary at the Institute for the Study of Labor.
15. ^ Zak PJ, Stanton AA, Ahmadi S (2007), Oxytocin Increases Generosity in Humans. PloSONE 2(11):e1128. [4]
16. ^ Sanfey, et al. (2002)
17. ^ Kirk; et al. (2011). "Interoception Drives Increased Rational Decision-Making in Meditators Playing the Ultimatum Game". Frontiers in Neuroscience 5:49: 49.
18. ^ Crockett, Molly J.; Luke Clark, Golnaz Tabibnia,
19. ^ Neural Substrates of Decision-Making in Economic Games Scientific Journals International [5]
20. ^ Oxytocin Increases Generosity in Humans PloSONE 2(11):e1128 [6]
21. ^ Koenigs, Michael; Daniel Tranel (January 2007). "Irrational Economic Decision-Making after Ventromedial Prefrontal Damage: Evidence from the Ultimatum Game". Journal of Neuroscience 27 (4): 951–956.
22. ^ Moretti, Laura; Davide Dragone; Giuseppe di Pellegrino (2009). "Reward and Social Valuation Deficits following Ventromedial Prefrontal Damage". Journal of Cognitive Neuroscience 21 (1): 128–140.
23. ^ See, for example, Gale et al. (1995), Güth and Yaari (1992), Huck and Oechssler (1999), Nowak & Sigmund (2000) and Skyrms (1996)
24. ^ See The Ultimatum game detailed description as a class room plan from EconomicsTeaching.org. (This is a more thorough explanation of the practicalities of the game than is possible here.)
25. ^ Güth et al. (1982), page 367: the description of the game at Neuroeconomics cites this as the earliest example.
26. ^ Nowak, M. A.; Page, K. M.; Sigmund, K. (2000). "Fairness Versus Reason in the Ultimatum Game". Science 289 (5485): 1773–1775.
27. ^ Ultimatum game with proposer competition by the GameLab.
28. ^ Ruffle (1998), p. 247.
29. ^ The reverse ultimatum game and the effect of deadlines is from Gneezy, Haruvy, & Roth, A. E. (2003).
30. ^ Stewart, Ian (May 1999). "A Puzzle for Pirates" (PDF). Scientific American 05: 98–99. Retrieved 3/11/2011.

## References

• Alvard, M. (2004). "The Ultimatum Game, Fairness, and Cooperation among Big Game Hunters". In
• Bearden, J. Neil (2001). "Ultimatum Bargaining Experiments: The State of the Art".
• Bicchieri, Cristina and Jiji Zhang (2008). "An Embarrassment of Riches: Modeling Social Preferences in Ultimatum games", in U. Maki (ed) Handbook of the Philosophy of Economics, Elsevier
• Bolton, G.E. (1991). "A comparative Model of Bargaining: Theory and Evidence". American Economic Review 81: 1096–1136.
• Gale, J., Binmore, K.G., and Samuelson, L. (1995). "Learning to be Imperfect: The Ultimatum Game". Games and Economic Behavior 8: 56–90.
• Gneezy, Haruvy, and
• Grimm, Veronika and F. Mengel (2011). "Let me sleep on it: Delay reduces rejection rates in Ultimatum Games'', Economics Letters, Elsevier, vol. 111(2), pages 113-115, May.
• Güth, W., Schmittberger, and Schwarze (1982). "An Experimental Analysis of Ultimatum Bargaining". Journal of Economic Behavior and Organization 3 (4): 367–388.
• Güth, W. and Yaari, M. (1992). "An Evolutionary Approach to Explain Reciprocal Behavior in a Simple Strategic Game". In U. Witt. Explaining Process and Change – Approaches to Evolutionary Economics. Ann Arbor. pp. 23–34.
• Huck, S. and Oechssler, J. (1999). "The Indirect Evolutionary Approach to Explaining Fair Allocations". Games and Economic Behavior 28: 13–24.
• Ochs, J. and
• Oosterbeek, Hessel, Randolph Sloof, and Gijs van de Kuilen (2004). "Cultural Differences in Ultimatum Game Experiments: Evidence from a Meta-Analysis". Experimental Economics 7 (2): 171–188.
• Ruffle, B.J. (1998). "More is Better, but Fair is Fair: Tipping in Dictator and Ultimatum Games". Games and Economic Behavior 23 (2): 247. .
• Sanfey; Rilling, JK; Aronson, JA; Nystrom, LE; Cohen, JD; et al. (2002). "The neural basis of economic decision-making in the ultimatum game". Science 300 (5626): 1755–1758. Category:CS1 maint: Explicit use of et al.).
• Skyrms, B. (1996). Evolution of the Social Contract. Cambridge University Press.
• Zak, P.J., Stanton, A.A., Ahmadi, S. (2007). Brosnan, Sarah, ed. "Oxytocin Increases Generosity in Humans" (PDF). PLoS ONE 2 (11): e1128.
• Angela A. Stanton (2007). "Neural Substrates of Decision-Making in Economic Games" (PDF). Scientific Journals International 1 (1): 1–64.