Professional Documents
Culture Documents
Juuso Valimaki
November 5, 2001
1
These notes are based on to a large extent on the notes of Dirk Bergemann for a similar course at Yale
University. I am grateful for Dirk for the permission to use and edit those notes.
Contents
i
CONTENTS ii
4 Repeated Games 48
4.1 Finitely Repeated Games . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
4.2 Infinitely Repeated Games . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
4.2.1 The prisoner’s dilemma . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
4.3 Collusion among duopolists . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
4.3.1 Static Game . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
4.3.2 Best Response . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
4.3.3 Collusion with trigger strategy . . . . . . . . . . . . . . . . . . . . . . . . . 53
4.3.4 Collusion with two phase strategies . . . . . . . . . . . . . . . . . . . . . . . 53
5 Sequential Bargaining 55
5.1 Bargaining with Complete Information . . . . . . . . . . . . . . . . . . . . . 56
5.1.1 Extending the Basic Model . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
5.2 Dynamic Programming . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
5.2.1 Additional Material: Capital Accumulation . . . . . . . . . . . . . . . . . . 62
6.6.1 Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
6.6.2 Equilibrium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
6.6.3 Equilibrium Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
8 Theoretical Complements 81
8.1 Mixed Strategy Bayes Nash Equilibrium . . . . . . . . . . . . . . . . . . . . . . . . 81
8.2 Sender-Receiver Games . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.2.1 Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.3 Perfect Bayesian Equilibrium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
8.3.1 Informal Notion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
8.3.2 Formal Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
Chapter 1
Game theory is the study of multi-person decision problems. The focus of game theory is interde-
pendence, situations in which an entire group of people is affected by the choices made by every
individual within that group. As such they appear frequently in economics. Models and situations
of trading processes (auction, bargaining), labor and financial markets, oligopolistic markets and
models of political economy all involve game theoretic reasoning. There are multi-agent decision
problems within an organization, many person may compete for a promotion, several divisions
compete for investment capital. In international economics countries choose tariffs and trade
policies, in macroeconomics, the central bank attempts to control price level.
2. What if each individual is uncertain about the characteristics of the other players?
1
CHAPTER 1. GAME THEORY: AN INTRODUCTION 2
In all game theoretic models, the basic entity is a player. In noncooperative games the individ-
ual player and her actions are the primitives of the model, whereas in cooperative games coalition
of players and their joint actions are the primitives. In other words, cooperative game theory can
be interpreted as a theory that allows for binding contracts between individuals. These notes will
consider only non-cooperative game theory.
• J. Nash (1950) Nonzero sum games and the concept of Nash equilibrium.
To describe a strategic situation we need to describe the players, the rules, the outcomes, and
the payoffs or utilities.
CHAPTER 1. GAME THEORY: AN INTRODUCTION 3
Example 3 Meeting in New York with unknown meeting place (Non Zero Sum Game, Coordi-
nation Game, Imperfect Information).
Chapter 2
We start with games that have finitely many strategies for all players and where players choose
their strategies simultaneously at the start of the game. In these games, the solution of the game
can simply be obtained by searching (in a clever way) among all possible solutions.
Bruce
Art Ballet
Art 2, 2 0, 0
Alice
Ballet 0, 0 1, 1
is the payoff of the “row player”. The second entry is the payoff of the “column player”. How
should this game be played? What should each individual player do?
But first we need to give a definition of a game. In order to describe a game, the set of players
must be specified:
I = {1, 2, ..., I} . (2.1)
In order to describe what might happen in the interaction, we need to specify the set of possible
choices, called strategies for each player
∀i, si ∈ Si , (2.2)
4
CHAPTER 2. NORMAL FORM GAMES 5
where each individual player i has a set of pure strategies Si available to him and a particular
element in the set of pure strategies is denoted by si ∈ Si . Finally there are payoff functions for
each player i:
ui : S1 × S2 × · · · × SI → R. (2.3)
A profile of pure strategies for the players is given by
I
s = (s1 , ..., sI ) ∈ × Si
i=1
or alternatively by separating the strategy of player i from all other players, denoted by −i:
A strategy profile is said to induce an outcome in the game. Hence for any profile we can deduce
the payoffs received by the players. This representation of the game is known as normal or
strategic form of the game.
Definition 4 (Normal Form Representation) The normal form ΓN represent a game as:
2.2 Dominance
The most primitive solution concept to games does not require knowledge of the actions taken
by the other players. A dominant strategy is the best choice for a player in a game regardless
of what the others are doing. The next solution concepts assumes that rationality of the players
is common knowledge. Common knowledge of rationality is essentially a recursive notion, so
that every player is rational, every player knows that every player is rational, every player knows
that every player knows that every player is rational..., possibly ad infinitum. Finally, the Nash
equilibrium concept requires that each player’s choice be optimal given his belief about the other
players’ behavior, and that this belief be correct in equilibrium.
Consider next a truly famous game which figures prominently in the debate on the possibility
of cooperation in strategic situations, the prisoner’s dilemma game: Besides its somewhat puzzling
Cooperate Defect
Cooperate 3, 3 0, 4
Defect 4, 0 1, 1
conclusion this game has another interesting property which comes to light when we analyze the
payoff of each individual player, for the row player and for the column player In both cases, one
strategy is better the other strategy independent of all choices made by the opponent.
CHAPTER 2. NORMAL FORM GAMES 6
C D
C 3 0
D 4 1
C D
C 3 4
D 0 1
The most straightforward games to analyze are the ones that have dominant strategies for all
players. In those games, there is little need to think about the strategic choices of others as the
dominant strategy can never be beaten by another choice of strategy.
Example 8 Second price auction (with complete information) is an example of a game with a
dominant strategy equilibrium.
Now consider a game Γ0 obtained from the original game Γ after eliminating all strictly domi-
nated actions. Once again if a player knows that all the other players are rational, then he should
not choose a strategy in Γ0 that is never-best response to in Γ0 . Continuing to argue in this way
leads to the outcome in Γ should survive an unlimited number of rounds of such elimination:
Definition 11 (Iterated Pure Strict Dominance) The process of iterated deletion of purely
strictly dominated strategies proceeds as follows: Set Si0 = Si . Define Sin recursively by
Sin = si ∈ Sin−1 @s0i ∈ Sin−1 , s.th. ui s0i , s−i > ui (si , s−i ) , ∀s−i ∈ S−i
n−1
,
Set
∞
\
Si∞ = Sin .
n=0
The set Si∞ is then the set of pure strategies that survive iterated deletion of purely strictly dom-
inated strategies.
Observe that the definition of iterated strict dominance implies that in each round all strictly
dominated strategies have to be eliminated. It is conceivable to define a weaker notion of iterated
elimination, where in each step only a subset of the strategies have to be eliminated.
Here is an example of iterated strict dominance:
(3) (1)
α2 β2 γ2
α1 2, 3 3, 2 0, 1
(2)
β1 0, 0 1, 3 4, 2
which yields a unique prediction of the game, and the superscript indicates the order in which
dominated strategies are removed. We might be tempted to suggest a similar notion of iteration
for weakly dominated strategies, however this runs into additional technical problems as order
and speed in the removal of strategies matters for the shape of the residual game as the following
example shows: If we eliminate first β 1 , then α2 is weakly dominated, similarly if we eliminate
α2 β2
α1 3, 2 2, 2
β1 1, 1 0, 0
γ1 1, 1 0, 0
γ 1 , then β 2 is weakly dominated, but if we eliminate both, then neither α2 nor β 2 is strictly
CHAPTER 2. NORMAL FORM GAMES 8
L R
T 1, 1 0, 0
M 1, 1 2, 1
B 0, 0 2, 1
dominated anymore. A similar example is The discussion of domination demonstrated that less
restrictive notion of play may often give good prediction.. However as the following example shows,
it clearly is not enough for analyzing all strategic interactions. The next subsection develops a
L M R
T 3, 0 0, 2 0, 3
M 2, 0 1, 1 2, 0
B 0, 3 0, 2 3, 0
solution concept that singles out (M, M ) as the relevant solution of the above game.
Art Ballet
Art 2 0
Ballet 0 1
for all s0i ∈ Si . Strategy si is never a best response if there is no s−i for which si is a best response.
CHAPTER 2. NORMAL FORM GAMES 9
We could now proceed and ask that the best response property holds for every individual engaged
in this play.
Definition 13 (Pure Strategy Nash Equilibrium) A strategy profile s = (s1 , ..., sn ) consti-
tutes a pure strategy Nash equilibrium of the game ΓN if for every i ∈ I,
In other words, a strategy profile is a Nash equilibrium if each player’s strategy si is a best
response to the strategy profile s−i of all the remaining players. In the language of a best response,
it says that
s ∈ BR (s) ,
and hence a pure strategy Nash equilibrium s is a fixed-point under the best response mapping.
Let us now consider some more examples to test our ability with these new notions.
Hawk Dove
Hawk 3, 3 1, 4
Dove 4, 1 0, 0
There are two conflicting interpretations of solutions for strategic and extensive form games:
• deductive (“eductive”).
The latter approach treats the game in isolation and attempts to infer the restrictions that
rationality imposes on the outcome. The games we are considering next are economic applications.
They are distinct from our previous examples insofar as the strategy space is now a continuum
of strategies. The first problem is indeed very continuous and regular optimization instruments
apply directly. The second example has continuous but not differentiable, payoff functions and
we shall proceed directly without using the regular optimization calculus.
CHAPTER 2. NORMAL FORM GAMES 10
q2
q1 (q2 )
q2 (q1 )
q1
P (q1 , q2 ) = a − q1 − q2
Πi = (a − q1 − q2 )qi
⇔
a − q2 a − q1
q1 (q2 ) = and q2 (q1 ) = , (2.7)
2 2
where q1 (q2 ) and q2 (q1 ) are the reaction functions or best response functions of the players, which
can be described graphically as:
CHAPTER 2. NORMAL FORM GAMES 11
for a = 1, and where the steeper function is q1 (q2 ) and the flatter function is q2 (q1 ). While
we can solve (2.7) directly, we realize the symmetry of the problem, which reduces the system of
equation to a single equation, and a symmetric equilibrium, so that
a−q
q=
2
and
a
q=
3
and hence the profits are
a a a2
Πi q 1 = , q 2 = = . (2.8)
3 3 9
We shall consider two different auction mechanisms: the first and the second price auction. The
object is given to the bidder with the highest index among all those who submit the highest bid.
In the first price auction, the winner pays his own bid, in the second price auction, the winner
pays the second highest bid.
The payoff functions in the first price auction are
vi − bi if bi > bj , ∀j 6= i,
ui (vi , b1 , ..., bn ) = vi − bi if bi ≥ bj , and if bi = bj ⇒ i > j. (2.9)
0, otherwise
bi ≥ bk ≥ bj , ∀j 6= i, k.
We can then show that the equilibrium in the first-price auction has the property that the bidder
with the highest value always wins the object. Notice that this may not be true in the second
price auction. However we can show that in the second price auction, the bid bi = vi weakly
dominates all other strategies.
Reading: Chapter 0 in [2] is a splendid introduction into the subject. Chapter b of [8] and
Chapter 1-1.2 in [4] covers the material of the last two lectures as does Chapter 1-1.2 in [3].
CHAPTER 2. NORMAL FORM GAMES 12
Heads Tails
Heads 1, −1 −1, 1
Tails −1, 1 1, −1
It is immediate to verify that no pure strategy equilibrium exists. A similar game is “Rock,
Paper, Scissors”:
R P S
R 0, 0 −1, 1 1, −1
P 1, −1 0, 0 −1, 1
S −1, 1 1, −1 0, 0
2.4.1 Basics
Let us now introduce the notation for mixed strategies. Besides playing a pure strategy si each
player is also allowed to randomize over the set of pure strategies. Formally, we have
Definition 14 A mixed strategy for player i, σ i : Si → [0, 1] assigns to each pure strategy si ∈ Si ,
a probability σ i (si ) ≥ 0 that it will be played, where
X
σ i (si ) = 1
si ∈Si
The mixed strategy can be represented as a simplex over the pure strategies:
n
( )
X
∆ (Si ) = (σ i (s1 ) , ..., σ i (sn )) ∈ Rn : σ i (sik ) ≥ 0, ∀j, and σ i (sik ) = 1 .
k=1
This simplex is also called the mixed extension of Si . For |Si | = 3, it can be easily represented as
a two-dimensional triangle sitting in a three dimensional space. σ i (si ) is formally a probability
CHAPTER 2. NORMAL FORM GAMES 13
mass function and σ i is a probability vector. Such a randomization is called a mixed strategy σ i
which is a element of set of distributions over the set of pure strategies:
σ i ∈ ∆ (Si ) .
An alternative notation for the set of mixed strategies, which you may find frequently is
σ i ∈ Σi .
In order to respect the idea that the players are choosing their strategies simultaneously and
independently of each other, we require that the randomized strategies chosen by the different
players satisfy statistical independence.1 In other words, we require that the probability of a pure
strategy profile s0 = (s01 , ..., s0N ) is chosen is given by
N
Y
σ i s0i .
i=1
The expected utility of any pure strategy si when some of the remaining players choose a
mixed strategy profile σ −i is
X Y
ui (σ 1 , ..., σ i−1 , si , σ i+1 , ..., σ I ) = σ j (sj ) ui (s1 , ..., si−1 , si , si+1 , ..., sI ) .
s−i ∈S−i j6=i
As another example, consider the famous “battle of the sexes” game: This game has clearly
Opera Football
Opera 2, 1 0, 0
Football 0, 0 1, 2
two pure strategy equilibria. To see if it has a mixed strategy equilibrium as well, calculate the
payoff to Sheila if Bruce is choosing according to mixed strategy σ B (·):
uS (O, σ B ) = 2σ B (O) + 0σ B (F ) ,
and similarly
uS (F, σ B ) = 0σ B (O) + 1σ B (F ) .
Again this allows for an illuminating graphical representation by the best-response function.
Consider the best-response mapping of the two players, where we plot the probability of attending
the football game for Sheila on the x-axis and for Bruce on the y-axis. Every point of intersection
indicates a Nash equilibrium. An equilibrium mixed strategy by Sheila would be written as
2 1
σ S = σ S (O) = , σ S (F ) = (2.11)
3 3
and if the ordering of the pure strategies is given, then we can write (2.11) simply as
2 1
σS = , .
3 3
Before we give a systematic account of the construction of a mixed strategy equilibrium, let us
extend our definitions for pure strategies to those of mixed strategies.
ui (σ i , σ −i ) ≥ ui σ 0i , σ −i
for all σ 0i ∈ Σi .
ui σ ∗i , σ ∗−i ≥ ui σ i , σ ∗−i
(2.12)
for all σ i ∈ Σi .
CHAPTER 2. NORMAL FORM GAMES 15
ui σ ∗i , σ ∗−i ≥ ui si , σ ∗−i ,
(2.13)
Since this is the first proof, we will be somewhat pedantic about it.
Proof. Suppose first that (2.12) holds. Then, it is sufficient (why?) to show that for every
si ∈ Si , there exists σ 0i ∈ Σi such that
0
But since the set of pure strategies is a strict subset of the set of mixed strategies, or Si ⊂ Σi , we
can simply set σ 0i = s0i , and hence it follows that (2.13) has to hold.
Next suppose that (2.13) holds and we want to show that this implies that (2.12) holds. It is
then sufficient (again, why) to show that for every σ 0i ∈ Σi , there exists s0i ∈ Si such that
But since every mixed strategy σ i is composed out of pure strategies, in particular since
X
ui σ i , σ ∗−i = σ i (si ) ui si , σ ∗−i ,
si ∈Si
consider the pure strategy si among all those with σ i (si ) > 0 which achieves the highest payoff
ui si , σ ∗−i . Then it must be the case that
ui si , σ ∗−i ≥ ui σ i , σ ∗−i ,
then Si+ (σ i ) is the set of all those strategies which receive positive probability under the mixed
strategy σ i . As Si+ (σ i ) contains all points for which σ i assigns positive probability, it is mathe-
matically referred to as the support of Si .
Proposition 19 (Composition) Let Si+ (σ ∗i ) ⊂ Si denote the set of pure strategies that player
i plays with positive probability in a mixed strategy profile σ ∗ = (σ ∗1 , ..., σ ∗I ). Strategy profile σ ∗ is
a Nash equilibrium in game ΓN = I, (Σi , ui )i∈I if and only if for all i = 1, ..., I
Bruce
Opera Football
Opera 2, 1 0, 0
Sheila
Football 0, 0 1, 2
uS (σ B , O) = 2σ B (O) + 0σ B (F ) ,
and similarly
uS (σ B , F ) = 0σ B (O) + 1σ B (F ) .
We want to create a mixed strategy for Sheila and Bruce. If Sheila is supposed to use a mixed
strategy, it means that she picks either alternative with a positive probability. As the strategy is
supposed to be a best-response to her, it has to be the case that the expected payoffs are identical
across the two alternative, for otherwise Sheila would clearly be better off to use only the pure
strategy which gives her the strictly higher payoff. In short, it has to be that
uS (σ B , O) = uS (σ B , F )
or explicitly
2σ B (O) + 0σ B (F ) = 0σ B (O) + 1σ B (F ) . (2.14)
Thus when we consider whether Sheila will randomize, we examine conditions on the mixing
behavior of Bruce under which Sheila is indifferent. Let
σ B (F ) = σ B
and consequently
σ B (O) = 1 − σ B .
when is this the case. Well, we have to investigate in turn how the indifference condition of
Bruce determines the randomizing behavior of Sheila. (The construction of the mixed strategy
equilibrium highlights the interactive decision problem, which is essentially solved as a fixed point
problem.) A similar set of conditions as before gives us
uB (σ S , O) = uB (σ S , F )
or explicitly
σ S (O) + 0σ S (F ) = 0σ S (O) + 2σ S (F ) . (2.16)
Thus as we consider whether Bruce is willing to randomize, we examine conditions on the mixing
behavior of Sheila to make Bruce indifferent. Let
σ S (F ) = σ S
and consequently
σ S (O) = 1 − σ S .
The payoff to the two players is then 32 in this mixed strategy equilibrium. This examples presents
the general logic behind the construction of a mixed strategy Nash equilibrium. The only issue
which we still have to discuss is the support of the mixed strategy equilibrium.
The following game is a helpful example for this question:
α β γ
α −1, 1 1, −1 x, x
β 1, −1 −1, 1 x, x
γ x, x x, x x, x
For x < 0, γ cannot be part of any mixed strategy equilibrium, and for x > 0, the original
matching pennies equilibrium cannot be a mixed strategy equilibrium anymore. However for
CHAPTER 2. NORMAL FORM GAMES 18
x = 0, there can be many mixed strategy equilibria, and they differ not only in the probabilities,
but in the support as well.2
3. mixed strategies as pure strategies in a game where the randomness is introduced through
some random private information (Harsayni’s purification argument)
Remark 20 In the battle of sexes, the mixed strategy equilibrium used private randomization by
each player.
Suppose instead they would have a common (or public) randomization device, say a coin, then
they could realize a payoff of ( 32 , 32 ) by e.g. both going to opera if heads and both going to football if
tails. Observe that this payoff is far superior to the mixed strategy equilibrium payoff The notion
of public randomization is used in the correlated equilibrium.
Finally we consider games where each player has a continuum of actions available. The first
one is an example of an “all pay auction”.
2
Verify that there can’t be an equilibrium in which randomization occurs between α and γ only, or alternatively
between β and γ.
CHAPTER 2. NORMAL FORM GAMES 19
The mixed (or pure) strategy by player i can be represented by the distribution function of the
bid of i:
Fi (bi ) ,
where we recall that the distribution function (or cumulative distribution function) Fi (bi ) is
defining the event
Fi (bi ) = Pr (b ≤ bi )
and
F b+ = lim F (a) .
a↑b
Before we can analyze and derive the equilibrium, we have to ask what is the expected payoff
from submitting a bid b1 given that player 2 uses an arbitrary distribution function F2 . From the
payoff functions, we see that it can be calculated as:
As agent i chooses his bids optimally in equilibrium, his bid must maximize his payoff. In other
words, his optimal bidding strategy is characterized by the first order conditions which represent
the marginal benefit equals marginal cost condition
f2 (b1 ) v − 1 = 0, (2.19)
which states that the increase in probability of winning which occurs at the rate of f (b1 ) multiplied
by the gain of v must equal the marginal cost of increasing the bid which is 1. As we attempt to
CHAPTER 2. NORMAL FORM GAMES 20
construct a mixed strategy equilibrium, we know that for all bids with f1 (b1 ) > 0, the first order
condition has to be satisfied. Rewriting (2.19) we get
1
f2 (b1 ) = . (2.20)
v
As we look for a symmetric equilibrium, in which agent i and j behave identically, we can omit
the subscript in (6.4) and get a uniform (constant) density
1
f ∗ (b) =
v
with associated distribution function:
b
F ∗ (b) = . (2.21)
v
As F (b = 0) = 0 and F (b = v) = 1, it follows that F ∗ (b) in fact constitutes a mixed equilibrium
strategy.
We went rather fast through the derivation of the equilibrium. Let us now complete the
analysis. Observe first that every pure strategy can be represented as a distribution function. We
start by explaining why there can’t be a pure strategy equilibrium in the all-pay auction, or in
other words why there can be no atoms with probability 1.
Proof. We argue by contradiction. Suppose therefore that there is a pure strategy equilibrium.
Then either (i) bi > bj or (ii) bi = bj . Suppose first that
bi > bj
then bi would lower his bid and still win the auction. Suppose then that
bi = bj .
v − bi − ε.
As ε → 0, and when taking the difference, we see that i would have a jump in his net utility of
1
2 v as
1 1 1
v − bi − ε − v − bi = v − ε → v,
2 2 2
CHAPTER 2. NORMAL FORM GAMES 21
Proof. Suppose ro the contrary that the strategy of player j contains an atom so that for some
b,
Fj (b) − Fj b− > 0,
Fi (b) − Fi (b − ε) > 0,
otherwise j could lower the bids at the atom without changing the probability of winning. But
consider the winnings of i at b0 < b. They are at best
vFj b− − b−
00
but by increasing the bid to b > b, i would receive
0
vFj b − b0 .
0 00
But letting b ↑ b and b ↓ b, we can calculate the expected payoff difference between
00 00
0 0
vFj b − b − vFj b − b = v Fj (b) − Fj b− > 0,
lim00 (2.22)
b0 ↑b, b ↓b
{x : f (x) 6= 0} ,
F(b)
b
b∗
CHAPTER 2. NORMAL FORM GAMES 23
equilibrium which has an interval as its support. Suppose Fi (bi ) is the distribution functions
which defines the mixed strategy by player i. For her to mix over any interval
b i ∈ bi , b i (2.23)
where c is the expected payoff from any mixed strategy equilibrium. Argue that c = 0, then we
solve the equation to
b
Fj (b) = , (2.25)
v
which tells us how player j would have to randomize to make player i indifferent. The argument
is actually already made by the derivation of the first order conditions and the observation that
there can’t be any atom in the interior of [0, v]. By symmetry, (2.25) then defines the equilibrium
pricing strategy by each player.
Suppose we were to increase the number of bidders, then (6.10) changes under the assumption
of symmetry to
v (F (b))n−1 − b = c, (2.26)
where c is the expected payoff from any mixed strategy equilibrium. Argue again that c = 0, then
we solve the equation to
1
b n−1
F (b, n) = (2.27)
v
and the expected payment of each individual bidder is
Z v Z v
bdF (b) = bf (b) db
0 0
and since 1
b
n−1
0 v
F (b) = f (b) =
(n − 1) b
we obtain 1
b
Z v
n−1
v v
db = . (2.28)
0 (n − 1) n
Nash equilibria, which have τ i = 0, τ j ≥ vc , however the only symmetric equilibrium is a mixed
strategy equilibrium, where each player must be indifferent between continuing and dropping out
of the game at any point in time, or
fi (t) fi (t) c
v=c⇔ = ,
1 − Fi (t) 1 − Fi (t) v
where the indifference condition may be interpreted as a first order differential equation
dFi (t) c
= .
(1 − Fi (t)) v
It is very easy to solve this equation once it is observed that the left hand side in the equation is
d
ln (1 − Fi (t)) .
−
dt
Hence integrating both sides and taking exponentials on both sides yields
c c
1 − Fi (t) = e− v t , or Fi (t) = 1 − e− v t .
and the expected time of dropping out of the game, conditional on the other player staying in the
game forever is Z ∞ Z ∞
v
tdF (t) = tf (t) dt = .
0 0 c
The properties of the exponential distribution are illustrated below for λ = 12 .
4
To understand why this is called the hazard rate, think of t as the time a machine fails. Then as time t passes,
and moving from t to t + dt, one finds that the conditional probability that the machine fails in this interval,
conditional on not having failed before is
f (t) dt
.
1 − F (t)
Als recall that the conditional probability is
P (A ∩ B)
P (A |B ) =
P (B)
CHAPTER 2. NORMAL FORM GAMES 25
1
2
f (t)
t
CHAPTER 2. NORMAL FORM GAMES 26
Discrete Case
In the discrete case, the numbers of possible outcomes is either finite or countable and so we can
list them. Consider the following experiment in which the numbers x ∈ {1, 2, 3, 4} are drawn with
probabilities p (x) as indicated. Any point x with p (x) > 0 is called a mass point or an atom.
The function p (·) is a probability mass function
p : X → [0, 1] ,
such that
X
p (x) = 1.
x∈X
This is an example of discrete probabilities. The cumulative distribution function is then given
by
X
F (z) = p (z) ,
x≤z
Continuous Case
Suppose next that we wish to choose any number between x ∈ [0, 4]. The assignment of proba-
bilities occurs via the density function f (x) ≥ 0, so that
Z 4
f (x) dx = 1.
0
F(t)
1
F(t)
t
CHAPTER 2. NORMAL FORM GAMES 28
The likelihood that a number x ∈ [x0 , x00 ] is chosen in the experiment is given by
Z x00
00 0
F x −F x = f (x) dx
x0
In other words, any interval of numbers [x, x0 ] has some positive probability of being chosen. An
example of such a distribution function is then given by:
0,
3
for x < 0
F (x) = 1 (2.29)
1 − 1 − 4 x , for 0 ≤ x ≤ 4
1, for x > 4
The density function then identifies the rate at which probability is added to the interval by taking
the limit as ∆ = x0 − x converges to zero:
Thus we have to distinguish between events which have probability zero, but are possible and
impossible events. The density function associated with the distribution function F (x) is given
by
0,
for x < 0
3 1
2
f (x) = 4 1 − 4 x , for 0 ≤ x ≤ 4
0 for x > 4
Mixed Case
Finally, we can of course have distribution functions with continuous and discrete parts. We saw
an example of this in the case of a mixed strategy with an atom.
CHAPTER 2. NORMAL FORM GAMES 29
F(t)
F(t)
t
CHAPTER 2. NORMAL FORM GAMES 30
F(t)
F(t)
f (t)
t
CHAPTER 2. NORMAL FORM GAMES 31
Notice that the discrete probability mass introduces a discontinuity in the distribution function
so that
F p− = lim F p0 6= F (p) .
0
p ↑p
The point x = 2 is often called an atom as it carries strictly positive probability to the distribution
function.
Expectations
Reading: Recall: A brief review of basic probability concepts, including sample space and
event, independent events, conditional probabilities and Bayes’ formula, discrete and continuous
random variables, probability mass function, density and distribution function, all introduced in
[12], Chapter 1 -2.4 is highly recommended. [10], chapter 3 contains a very nice discussion of the
interpretation and conceptual aspects of the equilibrium notion, especially in its mixed version.
Lemma 26 In a two player game, a strategy si of a player is never-best response if and only if
it is strictly dominated.
The notion of strict dominance has hence a decision-theoretic basis that is independent of the
notion of mixed strategy.5
Now consider a game Γ0 obtained from the original game Γ after eliminating all strictly domi-
nated actions. Once again if a player knows that all the other players are rational, then he should
not choose a strategy in Γ0 that is never-best response to in Γ0 . Continuing to argue in this way
leads to the outcome in Γ should survive an unlimited number of rounds of such elimination:
Definition 27 (Iterated Strict Dominance) The process of iterated deletion of strictly dom-
inated strategies proceeds as follows: Set Si0 = Si , and Σi = Σ0i . Define Sin recursively by
and similarly
Σni = {σ i ∈ Σi |σ i (si ) > 0 only if si ∈ Sin } .
Set
∞
\
Si∞ = Si ,
i=1
and
Σ∞
0 0
∞
i = σ i ∈ Σi @σ i ∈ Σi s.th. ui σ i , s−i > ui (σ i , s−i ) , ∀s−i ∈ S−i
5
The above Lemma extends to more than two players if all probability distributions on S−i and not only the
ones satisfying independence are allowed.
CHAPTER 2. NORMAL FORM GAMES 33
The set Si∞ is then the set of pure strategies that survive iterated deletion of strictly dominated
strategies. The set Σ∞
i is the set of player’s i mixed strategies that survive iterated strict domi-
nance.
Definition 28 (Rationalizable Strategies) Set Σ0i = Σi and for each i define recursively
n o
0 0
Σni = σ i ∈ Σn−1 n−1 n−1
∃σ ∈ × co Σ , s.th. u (σ , σ ) ≥ u σ , σ , ∀σ ∈ Σ .
i −i j6=i j i i −i i i −i i i
(2.31)
The rationalizable strategies for player i are
∞
\
Ri = Σni . (2.32)
n=0
I, (Σi , ui )i∈I , the strategies in ∆ (Si ) that survive the iterated removal of strategies that are
never a best response are known as player i0 s rationalizable strategies.
The reason for having the convex hull operator in the recursive definition is the following.
Each point in the convex hull can be obtained as a randomization over undominated strategies
for j 6= i. This randomization corresponds to i0 s subjective uncertainty over j 0 s choice. It could
fully
well be that this random element is dominated as a mixed strategy for player j. Hence
co Σn−1j 6= Σn−1
j in some cases.
Rationalizability and iterated strict dominance coincide in two player games. The reason for
the coincidence is the equivalence of never-best response and strictly dominated in two player
games which does not hold in three player games as the following example suggests. In the game
below, we have pictured the payoff to player 3, and player 3 chooses either matrix A, B, C or D.
where we can show that action D is not a best response to any mixed strategy of players 1
and 2, but that D is not dominated. If each player i can have a correlated belief about the others’
choices, the equivalence of ‘correlated rationalizability’ and iterated deletion of strictly dominated
strategies can be established.
CHAPTER 2. NORMAL FORM GAMES 34
α2 β2 α2 β2 α2 β2 α2 β2
α1 9 0 α1 0 9 α1 0 0 α1 6 0
β1 0 0 β1 9 0 β1 0 9 β1 0 6
A B C D
Theorem 29 (Existence in Finite Games) Finite games, i.e. games with a finite number of
players and a finite number of strategies for each player, have a mixed strategy Nash equilibrium.
The proof of this result is a standard application of Kakutani’s fixed point theorem, and it is
given in almost all graduate level textbooks, for instance the appendix to Chapter 8 in [8] give
the proof.
Many strategic situations are more easily modeled as games with a continuum of strategies
for each player (e.g. the Cournot quantity competition mode above). For these games, slightly
different existence theorems are needed. The simplest case that was also proved by Nash is the
following (and also covered in the same appendix of [8]).
Theorem 30 (Continuous-Quasiconcave case) Let ΓN = I, (Si , ui )i∈I be a normal form
game where I = {1, ..., N } for some N < ∞ and Si = [ai , bi ] ⊂ R for all i. Then if ui : S → R is
continuous for all i and quasiconcave in si for all i, ΓN has a pure strategy Nash equilibrium.
Observe that this theorem actually guarantees the existence of a pure strategy equilibrium.
This results from the assumption of continuity and quasiconcavity. If either of these requirements
is dropped, pure strategy equilibria do not necessarily exist.
Unfirtunately, the payoffs in many games of interest are discontinuous (e.g. the all pay auction
and the war of attrition examples above show this). For these games, powerful existence results
have been proved, but the mathematical techniques are a bit too advanced for this course. State
of the art results are available in the recent article [11]. The basic conclusion of that paper is
that it will be very hard to find economically relevant games where existence of mixed strategy
equilibria would be doubtful.
CHAPTER 2. NORMAL FORM GAMES 35
An artificial game where no equilibria (in either pure or mixed strategies) exists is obtained by
considering a modification of the Cournot quantity game. Recall that in that game, the payoffs
to the two players are given by
Πi (qi , qj ) = (a − q1 − q2 ) qi .
If the payoffs are modified to coincide with these at all (q1 , q2 ) except a3 , a3 and π i a3 , a3 = 0
for both firms, then one can show that the game has no Nash equilibria. The argument to show
this is that all of the strategies except for qi = a3 are deleted in finitely many rounds of iterated
deletion of strictly dominated strategies. Hence none of those strategies can be in the support of
any mixed strategy Nash equilibrium. But it is also clear that a3 , a3 is not a Nash equilibrium
of the game.
Reading: Chapter 8 in [8], chapter 2 in [3] and to a lesser extent Chapter 1 in [4] cover the
material of this chapter.
Chapter 3
3.1 Introduction
This lecture covers dynamic games with complete information. As dynamic games, we understand
games which extend over many periods, either finitely many or infinitely many. The new concept
introduced here is the notion of subgame perfect Nash equilibrium (SPE), which is closely related
to the principle of backward induction.
The key idea is the principle of sequential rationality: equilibrium strategies should specify
optimal behavior from any point in the game onward. The notion of subgame perfect Nash
equilibrium is a strengthening of the Nash equilibrium to rule out incredible strategies. We start
with finite games, in particular games with a finite horizon.
We then consider two of the most celebrated applications of dynamic games with complete
information: the infinite horizon bargaining game of Rubinstein and the theory of repeated games.
Example 31 Predation. There is an entrant firm and an incumbent firm. Sharing the market
(1, 1) is less profitable than monopoly (0, 2), but it is better than a price war (−3, −1). Depict
the game in the normal form and the extensive form. Characterize the equilibria of the game.
Distinction between subgame perfect and Nash equilibrium. Credible vs. incredible threat. Future
play (plan) affects current action.
Example 32 Cournot Model. Stackelberg Equilibrium. Consider again the quantity setting
duopoly with aggregate demand
P (q1 , q2 ) = a − q1 − q2
Πi (qi , qj ) = (a − q1 − q2 )qi
36
CHAPTER 3. EXTENSIVE FORM GAMES 37
The best response function in the quantity setting game are, as we observed before,
a − qi
qj = R (qi ) =
2
Suppose player i moves first, and after observing the quantity choice of player i , player j chooses
his quantity optimal. Thus i takes into account the reaction (function) of player j, and hence
a − qi a qi
Πi (qi , R (qi )) = (a − qi − )qi = ( − )qi
2 2 2
and the associated first-order conditions are
a
Π0i (qi , R (qi )) = − qi = 0.
2
The Stackelberg equilibrium is then
a a a a2
qi = , Πi , =
2 2 4 8
a a a a2
q j = , Πj , =
4 2 4 16
The sequential solution method employed here is called “backward induction” and the associated
Nash equilibrium is called the “subgame perfect” Nash equilibrium. Comparing with the Cournot
outcome we find
a a
qiS + qjS ≥ qiC + qiC = +
3 3
but
a2
ΠSi > ΠC i = , ΠSj < ΠCj .
9
This example illustrates the value of commitment, which is in this game simply achieved by moving
earlier than the opponent.
Notice that the Cournot equilibrium is still a Nash equilibrium. But is it reasonable? One has to
think about the issue of “non credibility” and “empty threats”.
v1 , v2 , ...., vn
such that vi and vi+1 are adjacent, i.e., form a 2-subset. If all its vertices are distinct, a walk is
called a path. A walk v1 , v2 , ...., vn such that v1 = vn is called a cycle.
A graph T is a tree if it has two properties
(T 1) T is connected,
where connected means that there is a path from every vertex to every other vertex in the graph.
A forest is a graph satisfying (T 2) but not necessarily (T 1).
In order to define formally the extensive form game, we need the following ingredients.
(i) The set of nodes V consists of initial nodes W , decision nodes X, and terminal nodes Y .
T = {W, X, Y } .
(ii) A mapping
p : T \W → T,
which is called the predecessor mapping (it is a point to set mapping). The immediate predecessor
nodes are p1 (x) and the immediate successor nodes of x are
where we require that α is a one-to-one function onto the set s (x) of x. The term α (s (x))
represents the sets of all actions which can be taken at x.
(iv) A function
ι:X→I
assigning each decision node to the player who moves at that node.
H = {h1 , ..., hk } ,
CHAPTER 3. EXTENSIVE FORM GAMES 40
s.th.
hk ∩ h0k = ∅,
and
[
hk = X.
k
The partition H can be further partitioned into Hi such that Hi = ι−1 (i), where Hi is the
set of all decision nodes at which i is called to make a move. A prominent way of identifying
an information set is through the history leading to it. Hence h ∈ H is often called a particular
history of the game.1
(vi) A collection of payoff functions
ui : Z → R.
Observe that since each terminal node can only be reached through a single path, defining payoffs
on terminal nodes is equivalent to defining payoffs on paths of actions.
(vii) An initial assessment ρ as a probability measure over the set of initial nodes
ρ : W → [0, 1] ,
si : Hi → A, (3.2)
Definition 39 (Mixed Strategy) Given player i0 s (finite) pure strategy set Si , a mixed strategy
for player i,
σ i : Si → [0, 1]
assigns to each pure strategy si ∈ Si a probability σ i (si ) ≥ 0 that it will be played where
X
σ i (si ) = 1.
si ∈Si
where |S| represents in general the cardinality of a set S, i.e. the number of members in the set.
If the number of actions available at each information set hi is constant and given by |A (hi )|,
then |Si | simplifies to
|Si | = |A (hi )||Hi | .
In extensive form games, it is often customary to use a slightly different notion for a mixed
strategy. In the definition above, the number of components needed to specify a mixed strategy
is |Si | − 1. This is often a very large number. A simpler way to define a random strategy is the
following. Rather than considering a grand randomization at the beginning of the game for each
player, we could consider a sequence of independent randomizations, one at each information set
for each of the players. This leads to the notion of a behavior strategy.
CHAPTER 3. EXTENSIVE FORM GAMES 42
bi : Hi → ∆ (A)
components must be specified. It is immediately obvious that each behavior strategy generates
a mixed strategy in the sense of the above definition. It is a much deeper insights into extensive
form games that most often it is enough to consider behavior strategies in the sense that each
outcome induced by a mixed strategy profile is also induced by some behavior strategy profile.
This is true as long as the game is one of perfect recall.
The first requirement says intuitively that no player forgets his moves, and the second says
that no player forgets past knowledge.
Theorem 43 (Kuhn’s Theorem) If the extensive form game {T, ≺; A, α; I, ι; H; ui ; ρ} has per-
fect recall and an outcome µ ∈ ∆ (Z) is induced by ρ and σ = (σ 1 , ..., σ N ) , then there is a behavior
strategy profile b = (b1 , ..., bN ) such that µ is induced by ρ and b.
In other words, in games of perfect recall, there is no loss of generality in restricting attention
to behavior strategies.
The following example is due to Rubinstein and Piccione (see the special issue of Games and
Economic Behavior, July 1997, on games of imperfect recall).
Example 44 (Absent Minded Driver) This is a single person game with two decision nodes
for the player. The player must decide whether to keep on driving D on the highway or whether
to exit E. If the driver chooses D twice, he ends in a bad nighborhood and gets payoff of −1. If
he exits at the first node, he gets 0.M If he chooses D followed by E, then his payoff is 2.
The twist in the story is that the driver is absent minded. He cannot remember if he has
already decided or not in the past. This absent mindedness is represented by a game tree where
CHAPTER 3. EXTENSIVE FORM GAMES 43
both of the nodes are in the same information set even though the two nodes are connected by an
action by the single player. Hence the game is not one of perfect recall as defined above.
It is a useful exercise to draw the game tree for this problem and consider the various notions
of strategies. The first step in the analysis of any games is to decide what the possible strategies
are. As usual, a strategy should be an assignment of a (possibly mixed) action to each information
set. With this definition for a strategy, it is easy to show that the optimal strategy in this game
cannot be a pure strategy. Observe also that if the driver were able to commit to a mixed strategy,
then she would choose to exit with probability p = 23 resulting in a payoff of 29 .
If the driver were to adopt the strategy of exiting with probability 23 , then she would assign
probabilities 3/4 and 1/4 respectively to the first and the second node in her information set. In
order to calculate correctly the payoffs to the driver, behavior at other nodes must be kept constant
when calculating the current payoff to exiting or driving on. Thus choosing D and then following
the proposed strategy yields a payoff of
3 2 1 1 1
(2 − ) − = .
4 3 3 4 2
Choosing E yields similarly a payoff of 12 .
The key to understanding the workings of the game lies in the interpretation of a deviation
from the proposed equilibrium strategy. Most game theorists believe that in games of this type, a
deviation should be modeled as a single deviation away from the proposed strategy. In particular,
this would imply that a deviation can take place at a single node in the information set without
necessarily happening at the other. Piccione and Rubinstein offer a different interpretation of the
game, see the discussions in the issue of GEB as cited above.
A proper subform is a subform T̂ consisting solely of some node x and its successors.
Definition 46 (Subgame) Given a proper subform, the associated proper subgame starts at x,
with the payoffs restricted to T̂ ∩ Z, and the initial assessment ρ̂ (x) = 1.
CHAPTER 3. EXTENSIVE FORM GAMES 44
σ = (σ 1 , ..., σ n )
Clearly, every SPE is an NE, but the converse doesn’t hold. Consider first the extensive
form games with perfect information. Chess is a finite extensive form game as is centipede game
introduced above. What is the unique subgame perfect equilibrium in this game?
It is a description of the equilibrium strategy for each player:
on and off the equilibrium path. The equilibrium path is the set of connected edges which lead
from an initial node to a final node. The equilibrium path is thus {s} and the equilibrium
outcome is the payoff realization (1, 1). Thus a description of the equilibrium strategy asks for
the behavior of every player on and off the equilibrium path.
Theorem 48 (Zermelo, Kuhn) Every finite game of perfect information ΓE has a pure strategy
subgame perfect Nash equilibrium.
w n
w r, r D, 2r − D
n 2r − D, D second period
W N
W R, R 2R − D, D
N D, 2R − D D, D
and
s∗1 = {n, W } , s∗2 {n, W }
Consider the following two-period example with imperfect information. The first period game
is a prisoner’s dilemma game and the second game is coordination game: and then Specify all
α β
α 2, 2 −1, 3
β 3, −1 0, 0
γ δ
γ x, x 0, 0
δ 0, 0 x, x
subgame perfect equilibria in this game. Does the class of subgame perfect equilibria coincide
with the class of Nash equilibria? For x ≥ 2, cooperation is sustainable in a subgame perfect
equilibria by switching from the pure to the mixed strategy equilibrium in the second stage game.
(1) −→ (2) −→ (1) −→ (2) −→ (1) .... (2) −→ (1) −→ (100, 100)
↓ ↓ ↓ ↓ ↓ ↓
(1, 1) (0, 3) (2, 2) (1, 4) (98, 101) (100, 100)
The first entry in the payoff vector refers to player 1, the second to player 2. Instead of looking
at the ‘true’ centipede game, let us look at a smaller version of the game:
What is a strategy for player 1 in this small centipede game. A strategy is a complete
description of actions at every information set (or history). Player 1 has two histories at which
he could conceivable make a choice H1 = {∅, cc} and likewise H2 = {c, ccc0 }. Thus a possible
strategy for player 1 is
h1 = ∅ h1 = cc h1 = ccc0 c0
↓ ↓ ↓
s1 = { c , c0 s00 }.
h1 = ∅ h1 = cc h1 = ccc0 c0
↓ ↓ ↓
s1 = { s , c0 s00 }.
The notion of a strategy in the extensive form game is stretched beyond what we would call
a plan. We saw in the centipede game that it indeed requires a player to specify his action after
histories that are impossible if he carries out his plan. A different interpretation of the extensive
form game may help our understanding. We may think of each player having as many agents (or
multiple selves) as information sets. The strategy of each player is to give instruction to each
agents (which he may not be able to perfectly control) as to what he should do if he is called to
make a decision. All agents of the same player have the same payoff function. A plan is then a
complete set of instruction to every agent who acts on behalf of the player.
Notice that we did not discuss notions similar to rationalizability in the context of extensive
form games. The reason for this is that making sense of what common knowledge of rationality
means in extensive form games is much more problematic than in normal form games. A simple
example will illustrate.
Robert Aumann has maintained that common knowledge of rationality in games of perfect
information implies backwards induction. Hence e.g. the centipede game should have a unique
CHAPTER 3. EXTENSIVE FORM GAMES 47
solution where all players stop the game at all nodes. Suppose that this is indeed the case. Then
after observing a move where the first player continues the game, the second player ought to
conclude that the first player might not be rational. But then it might well be in the second
player’s best interest to continue in the game. But if this is the case, then it is not irrational
for the first player to continue at the first node. It should also be pointed out that knowledge
and belief with probability 1 are more or less the same in normal form games, but in extensive
form games these two notions yield very different conclusions. Appropriate models of knowledge
for extensive form games form a very difficult part of interactive logic (or epistemic logic) and
Chapter 14 in [3] is a slightly outdated introduction into this area.
3.5 Reading
The notion of extensive form game is first (extensively) developed in [7] and nicely developed in [5]
and [6]. The relation between mixed and behavior strategies is further discussed in [3], Ch.3.4.3.
Chapter 4
Repeated Games
C D
C 4, 4 0, 5
D 5, 0 1, 1
i in the repeated game is then a mapping from the history of the game H t into the strategies in
the stage game Si :
sti : H t−1 → Si .
A complete specification of repeated game strategy has to suggest a strategy choice after every
possible history of the game, (and not only the histories which will emerge in equilibrium). A
complete specification can be thought of as a complete set of instructions written before playing
the game, so that no matter what happens in the game, once it started a prescription of play is
found the set of instructions. The history of the game is formed by the past observed strategy
choices of the players:
H t−1 = s11 , ..., s1I ; ...; st−1 t−1
1 , ..., sI .
Definition 49 Given a stage game Γ = I, (Si )i∈I , (ui )i∈I , we denote by Γ (T ) the finitely
repeated game in which Γ is played T times.
The outcomes of all preceding plays are observed before the next play begins. The payoffs for
Γ (T ) are simply the sums of the payoffs from the T stage games.
48
CHAPTER 4. REPEATED GAMES 49
Definition 50 In a repeated game Γ (T ) a player’s strategy specifies the action the player will
take in each stage, for each possible history of play through the previous stage:
where:
sti : H t−1 → Si .
Theorem 51 If the stage game Γ has a unique Nash equilibrium then, for any finite T , the
repeated game Γ (T ) has a unique subgame-perfect outcome: the Nash equilibrium of Γ is played
in every stage.
Corollary 52 Every combination of stage game Nash equilibria forms a subgame perfect equilib-
rium.
Consider now the following game with multiple equilibria in the stage game Can coordination
C D E
C 4, 4 0, 5 −3, −3
D 5, 0 1, 1 −3, −3
E −3, −3 −3, −3 -2,-2
Example 53 To understand the complexity of repeated games, even if they are repeated only
very few times, let us consider the prisoner’s dilemma played three times. What are all possible
histories of a game. How many strategy choice have to be made in order for the repeated game
strategy to be a complete specification?
The history H −1 is naturally the empty set:
H −1 = {∅} .
The history H 0 is the set of all possible outcomes after the game has been played once. Each
outcome is uniquely defined by the (pure) strategy choices leading to this outcome, thus
We then observe that the number of possible histories grows rather rapidly. In period 1, it is
already
1
H = |S1 | × |S2 | × |S1 | × |S2 | = 16,
and in general it is
H = (|S1 | × |S2 |)t+1 .
t
However the complexity of strategies grows even faster. A complete strategy for period t specifies
a particular action for every possible history leading to period t. Thus the number of contingent
choices to be made by each player in period t is given by
t+1
Si = |Si ||H t−1 | = |Si ||S1 |×|S2 | .
t
Notice that a strategy even in this two period game is not merely an instruction to act in a specific
manner in the first and second period, but needs to specify the action after each contingency.
In order to sustain cooperation two conditions need to be satisfies: (i) there are rewards from
cooperation and (ii) there are credible strategies to punish deviators. This theme of “carrots
and sticks” is pervasive in the entire theory of repeated games. As the punishment occurs in the
future, agents must value the future enough so that lower payoffs in the future would indeed harm
the agent’s welfare sufficiently. If the discount factor represents the agent’s attitude towards the
future, then higher discount factors make cooperation easier to achieve as lower payoffs in the
future figure more prominently in the agent’s calculation.
Definition 54 Given a stage game Γ, let Γ (δ) denote the infinitely repeated game in which Γ is
repeated forever and where each of the players discounts future at discount factor δ < 1.
The payoff in an infinitely repeated game with stage payoffs π t and discount factor δ < 1 is
given by
∞
X
δtπt.
t=0
Definition 55 Given the discount factor δ, the average payoff of the infinitely repeated sequence
of payoffs π 0 , π 1 , π 2 , ... is given by
∞
X
(1 − δ) δtπt.
t=0
Notice that in infinitely repeated games every subgame is identical to the game itself.
CHAPTER 4. REPEATED GAMES 51
We now want to show that this set of strategies form a subgame perfect equilibrium. To prevent
a deviation the following has to hold:
β 4
5+ ≤ ,
1−β 1−β
or
5 (1 − β) + β ≤ 4,
so that
1
β≥ .
4
Thus if the players are sufficiently patient, cooperation can be sustained in this repeated game.
The equilibrium we suggested implies cooperation forever along the equilibrium path. How-
ever, this is clearly not the only equilibrium. There are many more, and many of them of very
different characteristics. In the limit as δ → 1, one can show, and this is the content of the
celebrated “folk” theorem’s of the repeated games, that the entire convex hull of all individually
rational payoffs of the stage game can be achieved as a subgame perfect equilibrium. Notice, that
this may involve very asymmetric payoffs for the players:
p (q1 , q2 ) = a − q1 − q2 .
The Cournot stage game equilibrium is found by solving the first order conditions of the profit
functions, where we normalize the marginal cost of production to zero,
π i (q1 , q2 ) = qi (a − q1 − q2 ) .
π (q) = q (a − q) ,
a − 2q = 0
and hence
a
qM =
2
and the profits are
a2
π qM = πM = .
4
π i (q, q) = q (a − 2q) = ax − 2q 2 .
and consider whether the symmetric strategy can form a subgame perfect equilibrium? We have
to consider the alternatives available to the players. Consider an history in which
M M M M
q q q q
ht = , , ...., , ,
2 2 2 2
then the resulting payoff would be
1 1 M δ
π ≥ πd + πC . (4.3)
1−δ2 1−δ
Thus to evaluate whether the player follow the equilibrium strategy, we have to evaluate π d .
SPE
(1 − δ) (1 + δ) 1 M
π ≥ π d + δ ax − 2x2
1−δ 2
and finally as:
1 M 1
π − ax − 2x2 ≥ πd − πM .
δ (4.8)
2 2
The inequality (4.8) suggests that the relative gains from deviating in contrast to collusion today
must be smaller than the relative losses tomorrow induced by a deviation today and again in
contrast to continued collusion. Similarly for the punishment to be credible, starting from (4.7),
we get
1 (a − x)2
ax − 2x2 + δ π M ≥ + δ ax − 2x2 .
(4.9)
2 4
and which permits an intuitive interpretation in terms of today’s and tomorrow’s payoff, or alter-
natively as the one shot deviation strategy. By inserting the value of π M we obtain two equalities:
2 9 2 1 a2
δ 12 a4 − ax − 2x2 = 64 a −2 4
2
2
δ 21 a4 − ax − 2x2 = (a−x) − ax − 2x2
4
5
which we can solve and obtain the following results as the harshest level of punishment x = 12 a
9
and the lowest level of discount factor δ = 32 for which cooperation is achievable.
Reading: [4], Ch. 2.3 is a nice introduction in the theory of repeated games. A far more
advanced reading is [3], Ch.5. The characterization of extremal equilibria is first given in [1].
Chapter 5
Sequential Bargaining
The leading example in all of bargaining theory is the following trading problem in the presence of
bilateral monopoly power. A single seller and a single buyer meet to agree on the terms of trade
for a single unit of output. The seller incurs a cost of c to produce the good while the buyer has a
value of v for the good. If the two parties have payoff functions that are quasilinear in money and
the good, then there are positive gains from trade if and only if v > c. The two parties have also
an opportunity to refuse any trade which yields them an outside option value of 0. Both parties
prefer (weakly) trading at price p to not trading at all if p satisfies v ≥ p ≥ c. Which (if any) of
such prices will be agreed upon when the two parties negotiate?
The conventional wisdom up to late 70’s was that the determination of the price depends
largely on psychological and sociological factors determining the bargaining power of the two
parties and as a result economics had little to say on the issue. When the concept of subgame
perfect equilibrium was established as the standard solution concept in games of perfect infor-
mation, the bargaining issue was taken up again. Preliminary work was done on bargaining in
simple two period models, but the real breakthrough was the infinite horizon bargaining model
with alternating offers proposed by Ariel Rubinstein in [13]. There he showed that the infinite
horizon game has a unique subgame perfect equilibrium and that the division of surplus between
the bargainers can be explained in terms of the time discount rates of the bargainers.
It should be pointed out that other approaches to bargaining were also adopted. An entirely
different methodology starts by describing a set of possible outcomes X that can be obtained in
the bargaining process between the two players and an alternative d that represent the outcome
in the case of no agreement. Each of the bargainers has preference relation i defined on X ∪ {d}.
Let the set of possible preference relations of i be denoted by Ri . A solution to the bargaining
problem is a function that associates an outcome to each profile of preferences. In other words,
the solution is a function f such that
f : R1 × R2 → X ∪ d.
55
CHAPTER 5. SEQUENTIAL BARGAINING 56
The method of analysis in this approach is often taken to be axiomatic. The analyst decides
in the abstract a set of criteria that a good solution concept should satisfy. An example of such
criteria could be Pareto efficiency of the solutions, i.e. if x i x0 and x j x0 , then x0 ∈
/ f (i , j ) .
Once the criteria have been established, the analyst determines the set of functions f that satisfy
this criterion. Preferably this set is small (ideally a singleton) so that some clear properties of
the solutions can be established. This approach precedes the noncooperative approach to be
followed in this handout. The first contribution to this cooperative or axiomatic approach to
bargaining was made by John Nash in “The Bargaining Problem” in Econometrica 1950. Various
other solutions were proposed for two player and many player bargaining models. A good survey
of these developments can be found in chapters 8-10 in Roger Myerson’s book “Game Theory:
Analysis of Conflict”. Rubinstein’s solution of the noncooperative bargaining game picks the
same solution as Nash’s original cooperative solution for that game. A recurring theme in the
noncooperative bargaining literature has been the identification of fully strategic models that
yield the classical cooperative bargaining solutions as their equilibrium outcomes.
be that the equilibrium shares in the game are (0, 1) . Hence we could say that player 2 has all
the bargaining power since she is the one in a position to make the last take it or leave it offer.
The situation changes somewhat if the players are less than perfectly patient. Assume now that
player i has a discount factor δ i between the periods in the game. If player 2 refuses the first offer
and gets all of the surplus in the second period, her payoff from this strategy is δ 2 . Hence she
will accept any first period offer x > δ 2 and refuse any offer x < δ 2 . At x = δ 2 , she is indifferent
between accepting and rejecting the offer, but it is again easily established that in the unique
subgame perfect equilibrium of the game, she will accept the offer x = δ 2 and the outcome of the
game is (1 − δ 2 , δ 2 ) . Notice that the introduction of time discounting lessens the advantage from
being in the position to make the take it or leave it offer since the value of that eventual surplus
is diminished.
If there were 3 periods in the game, the unique equilibrium in the subgame that starts following
a rejection of the first offer would be the solution of the 2 period game (by uniqueness of the
subgame perfect equilibrium in the 2 period game with discounting). Hence player 2 could secure
a payoff of 1 − δ 1 from the second period onwards. As a result, player 1 must offer to keep exactly
1 − δ 2 (1 − δ 1 ) to himself in the first period and the outcome is thus (1 − δ 2 (1 − δ 1 ) , δ 2 (1 − δ 1 )) .
At this point, we are ready to guess the form of the equilibrium in a game with T periods.
Proposition 56 Let xT denote the unique subgame perfect equilibrium share of player 1 in the
bargaining game of length T. Then we have
xT +2 = 1 − δ 2 1 − δ 1 xT .
Proof. Consider the game with T + 2 periods. In period T + 1, player 2 must offer a share of
xT to player 1. This results in surplus 1 − δ 1 xT to player 2. Hence player 1 must offer a share
δ1
δ 2 1 − δ 1 xT to player 2 in period T +2 proving the first claim. To see that in equilibrium, the first
period offer must be accepted, assume to the contrary and denote the equilibrium continuation
payoff vector by (v1 , v2 ) after the refusal by player 2. Because of time discounting and feasibility
v1 +v2 < 1. But then it is possible for 1 to offer x0 such that x0 > v1 and (1 − x0 ) > v2 contradicting
the optimality of the original offer.
The proposition gives the equilibrium shares as a solution to a difference equation. The most
interesting question about the shares is what happens as the horizon becomes arbitrarily long,
i.e. as T → ∞. Let x∗ = limT →∞ xT . Then it is easy to see that
1 − δ2
x∗ = .
1 − δ1δ2
Observe that in the case of identical discount factors, this share is monotonically decreasing in
δ and that (by using L’Hopitals rule) limδ→1 x∗ (δ) = 12 . Hence the unique outcome in finite but
long bargaining games with equally patient players is the equal division of surplus.
CHAPTER 5. SEQUENTIAL BARGAINING 58
Notice that finite games of this type always put special emphasis on the effects of last stages.
By the process of backwards induction, these effects have implications on the earlier stages in the
game as well. If the bargaining can effectively take as much time as necessary, it is questionable
if it is appropriate to use a model which features endgame effects. The simplest way around
these effects is to assume that the bargaining process may take infinitely many periods. A simple
argument based on the continuity of the payoff functions establishes that making the offers from
the finite game in all periods of the infinite game continues to be an equilibrium in the infinite
game. The more difficult claim to establish is the one showing that there are no other subgame
perfect equilibria in the game. In order to establish this, we start by defining the strategies a bit
more carefully.
The analysis here is taken from Fudenberg and Tirole, Chapter 4. The infinite horizon alter-
nating offer bargaining game is a game with perfect information, i.e. at all stages, a single player
moves and all of the previous moves are observed by all players. In odd periods, player 1 makes
the offer and player 2 either accepts (thereby ending the game) or rejects and the game moves to
the next period. In even periods, the roles are reverse. Denote by ht the vector of all past offers
and acceptance decisions. A strategy for player i is then a sequence of functions:
At the next round, player 1 must accept all offers above xk+1 ≡ δ 1 (1 − δ 2 ) + δ 1 δ 2 xk . To see
this, consider the implications of refusing an offer by player 2. This has three possible implications.
First, it may be that agreement is never reached. The second possibility is that 2 accepts some
future offer from 1. This has a current value of at most δ 1 1 − δ 2 1 − xk . The third possibility
is that player 1 accepts one of player 2’s future offers. The payoff from this is at most δ 21 xk . The
payoff from the second possibility is the largest of the three. Hence 1 must accept all offers above
xk+1 .
A similar argument shows that player 2 must accept all offers below y k+1 ≡ 1 − δ 2 + δ 1 δ 2 y k .
The sequences xk and y k are monotonic sequences with limits
δ 1 (1 − δ 2 ) ∞ 1 − δ2
x∞ = , y = .
1 − δ1δ2 1 − δ1δ2
1−δ 2
Hence player 2 rejects any offer where 1’s share is above y ∞ = 1−δ 1 δ
2
and accepts all shares below
1−δ 2 1−δ 2
y ∞ . Hence the unique outcome is again for 1 to propose the share 1−δ 1 δ 2 , 1 − 1−δ 1 δ 2 and for 2
to accept immediately.
values, where the values represent the maximal value one can attain in any subtree. When does
this reduction lead to the optimal decision through the entire tree: if and only if we replace each
subtree with the optimal value.
Theorem 57 Every finite decision tree (with finitely many periods and finitely may choices) has
an optimal solution which can be obtained by backwards induction.
We can extend this theorem certainly to an interactive decision problem, provided that it
maintains the structure of single person decision tree, namely perfect information.
Theorem 58 (Zermelo, Kuhn) Every finite game of perfect information ΓE has a pure strategy
subgame perfect Nash equilibrium.
The theorem applies to many parlor games, in particular to chess. Thus in principle we can
apply backward induction to solve for a subgame perfect Nash equilibrium in chess. How does a
computer program approach the problem of evaluating a move? By approximating the value of
a move through a forward looking procedure which assesses the value of the position in three or
four moves from then on. We may apply the same idea of approximation to every problem with
an infinite horizon. In fact, we can obtain in many situations an exact approximation.
We now return to the bargaining problem as defined above. Two players, 1 and 2 bargain over
a prize of value 1.. The first player starts by making an offer to the second player who can either
accept or reject the offer. If he rejects the offer, then in the next period he has the right to make
an offer which can again be accepted or rejected. The process of making offers then starts again.
The agents are impatient and have common discount factor of
δ = 1/ (1 + r) < 1.
Consider first the game with an infinite horizon. The surprising result and the even more surprising
simplicity of the argument is given first. We can adopt the definition of stationarity also for
equilibrium strategies.
is stationary if
sti = si , ∀i, ∀t.
Theorem 60 The unique subgame perfect equilibrium to the alternate move bargaining game is
given by for the offering party i to suggest
δ
v=
1+δ
CHAPTER 5. SEQUENTIAL BARGAINING 61
in every period and for the receiving party i to always accept if v ≥ δ/ (1 + δ) and reject if
v < δ/1 − δ.
Notice that you have to specify the strategy for every period in the infinite time horizon model.
The proof proceeds in two parts. First we show that a stationary equilibrium exists, and then we
show that it is the unique subgame perfect equilibrium, stationary or not. To do the later part
of the argument, first upper and lower bounds are obtained for each player’s equilibrium payoff.
Then it is shown that the upper and lower bounds are equal, which shows that an equilibrium
exists and that it is unique.
Proof. We start with the stationary equilibrium and use directly the payoffs and then later
on derive the strategies which support these payoff. Suppose then that there is a stationary
equilibrium. Then the following has to be true for the person who accepts today:
v = δ (1 − v) ,
v = 1 − δv̄,
or
1 = v + δv̄. (5.1)
Next we claim that v̄ cannot be larger than 1 − δv if the offer is accepted in the first period. But
what about making an offer that is rejected in the first period? As the second player gets at least
v in the next period, this cannot give the player more than δ (1 − v) as he has to wait until the
next period. Hence
v̄ ≤ 1 − δv.
where the last equality has been obtained through (5.1). But rewriting (5.2) we obtain
v̄ (1 − δ) ≤ v (1 − δ) ,
CHAPTER 5. SEQUENTIAL BARGAINING 62
which be the definition of the upper and lower bounds imply that v̄ = v. Let us denote this payoff
by v ∗ . Since v ∗ = 1 − δv ∗ , we find that player 1 must earn
1
v∗ = ,
1+δ
and player 2 must earn
δ
δv ∗ = .
1+δ
The derivation of the equilibrium strategies follows directly from here.
kt+1 = (1 − λ) kt + it
Thus if we knew the values in the future say V (k1 ), we could certainly find the optimal policy
for period 0, as we could formulate the following problem
1 2
V (k0 ) , max αk0 − i0 + δV ((1 − λ) k0 + i0 ) (5.3)
i0 2
Supposing that V is differentiable, we would get as a first order condition
δV 0 ((1 − λ) k0 + i0 ) − i0 = 0.
kt = k
CHAPTER 5. SEQUENTIAL BARGAINING 63
In other words, the strategy is time invariant and clearly we have that
it = i = λk for all t.
or
αk ∗ − 12 (λk ∗ )2
V (k ∗ ) =
1−δ
However, it still remains to verify what the optimal solution is. But we can do this now by
analyzing the first order condition of the Bellman equation
α − λk ∗
δ − λk ∗ = 0
1−δ
Thus we get
αδ
k∗ = .
λ
Chapter 6
6.1 Introduction
This is a very modest introduction into the theory of games with incomplete information or
Bayesian games. We start by defining incomplete information, then suggest how to model incom-
plete information (as a move by nature) to transform it into a game of imperfect information. We
then introduce the associated equilibrium concept of Bayesian Nash equilibrium.
6.1.1 Example
Harsanyi’s insight is illustrated by the following example. Suppose payoffs of a two player two
action game are either:
α β
α 1,1 0,0
β 0,1 1,0
or
α β
α 1,0 0,1
β 0,0 1,1
i.e. either player II has dominant strategy to play H or a dominant strategy to play T . Suppose
that II knows his own payoffs but player I thinks there is probability p that payoffs are given by
the first matrix, probability 1 − p that they are given by the second matrix. Say that player II is
of type 1 if payoffs are given by the first matrix, type 2 if payoffs are given by the second matrix.
Clearly equilibrium must have: II plays H if type 1, T if type 2; I plays H if p > 12 , T if p < 12 .
But how to analyze this problem in general? All payoff uncertainty can be captured by a single
move of “nature” at the beginning of the game.
64
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 65
6.2 Basics
A (finite) static incomplete information game consists of
• Players 1, ..., I
6.2.1 Strategy
A pure strategy is a mapping si : Ti → Ai . Write Si for the set of such strategies, and let
S = S1 × .. × SI . A mixed strategy can also depend on his private information:
σ i : Ti → ∆(Si ).
6.2.2 Belief
A player’s “type” ti is any private information relevant to the player’s decision making. The private
information may refer to his payoff function, his belief about other agent’s payoff function, the
likelihood of some relevant event, etc. We assume that for the players types {ti }Ii=1 there is some
objective distribution function
p (t1 , ..., tI )
where ti ∈ Ti , which is called the prior distribution or prior belief. The conditional probability
p t0−i , ti
0
p t−i |ti = P
t−i ∈T−i p (t−i , ti )
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 66
by Bayes rule denotes the conditional probability over the other player’s types give one’s own
type ti . This expression assumes that Ti are finite. Similar expressions hold for infinite Ti . If the
player’s types are independent then
6.2.3 Payoffs
The description of the Bayesian game is completed by the payoff function
If player i knew the strategy of the other players as a function of his type then he could calculate
the expected payoff from his decision given the conditional probability p (t−i |ti ).
Again, by imperfect information we mean that at some move in the game the player with the
move does not know the complete history of the game thus far. Notice that with this specification,
the only proper subgame of a game of incomplete information is the whole game. Any Nash
equilibrium is then subgame perfect and the subgame perfection has no discriminating power.
where s = (s1 , ..., sI ) and s (t) = (s1 (t1 ) , ..., sI (tI )). Recall:
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 67
is a strategy profile s∗ = (s∗1 , ..., s∗I ) such that, for all i, for all ti ∈ T , condition (6.8) is satisfied.
Notice the differences in the definition between the Bayesian Nash equilibrium and the Nash
equilibrium. In the former the payoff function may depend on the type. Moreover the payoff is
calculated as an expected payoff given his own information, with the conditional distribution over
all other agent’s type.
A Bayesian Nash equilibrium is then a strategy pair {b1 (v1 ) , b2 (v2 )}. The optimal strategy is
found for each player by solving the following problem
1
max (vi − bi ) Pr (bi > bj (vj )) + (vi − bi ) Pr (bi = bj (vj )) . (6.3)
bi 2
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 68
bi = ai + ci vi .
Notice that
bi − aj
Pr (bi ≥ aj + cj vj ) = Pr vj ≤ , (6.4)
cj
and by the uniformity this results in
bi − aj
max (vi − bi ) .
bi cj
(vi + aj ) /2 = ai + ci vi , ∀vi ,
and
(vj + ai ) /2 = aj + cj vj , ∀vj ,
and hence
1
ci = cj = ,
2
and consequently
ai = aj = 0.
where each price pi has an unconditional probability 0 < αi < 1, and ni=1 αi = 1, by which the
P
price pi is chosen. Suppose the buyer is willing to accept all prices, then the expected price to be
paid by the buyer is
Xn
E [p] = α i pi , (6.7)
i=1
which is the unconditional expectation of the price to be paid. Suppose next, that the buyer
adopts a rule to accepts all price higher or equal to pk and reject all other prices. (Recall this is
just an example.) Then the unconditional expectation of the price to be paid is
n
X
E [p] = α i pi , (6.8)
i=k
The expectation in (6.8) is still unconditional since no new information entered in the formation
of the expectation. Suppose we represent the acceptance rule of the buyer by the index k below
the expectations operator E [·], as in Ek [·]. The expected payment is then
k−1
X n
X
Ek [p] = αi 0 + α i pi ,
i=1 i=k
as
n
X n
X k−1
X
E1 [p] − Ek [p] = α i pi − α i pi = αi pi > 0. (6.9)
i=1 i=k i=1
since the buyer is willing to pay pi in more circumstances in (6.7) and (6.8).
Consider now the average price which is paid under the two different policies conditional on
accepting a price and paying it, in other words, we ask for the conditional expectation:
Ek [p |p ≥ pk ]
We consider now exclusively the events, i.e. realization of prices p, where p ≥ pk for some k. The
conditional probability of a price pi being quoted conditional on pi ≥ pk is then given by
Pr (pi )
Pr (pi |pi ≥ pk ) = , (6.10)
Pr (pi ≥ pk )
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 70
since
Pr ({pi } ∩ {pi ≥ pk }) = Pr (pi )
if indeed pi satisfies pi ≥ pk , which is the nothing else than Bayes rule.1 Notice that we can write
more explicitly as:
αi
Pr (pi |pi ≥ pk ) = Pn , (6.12)
j=k αj
and that
n
X
(pi |pi ≥ pk ) = 1.
i=k
The conditional expectation of the prices, conditioned on accepting then price, is then nothing
else but taken the expectation with respect to the conditional probabilities rather than with the
unconditional probabilities, so that
n
α
Pn i
X
Ek [p |p ≥ pk ] = pi ,
j=k αj
i=k
where Ek [p |p ≥ pk ] is now the expected price paid conditional on accepting the price. You may
now verify that the expected price or average price paid, conditional on paying is increasing in k,
as we might have expected it.
Next we generalize the notion of conditional expectation to continuum of prices where the
prices are distributed according to a distribution function.
where the distribution of prices is given by a (continuous) distribution function F (p) and an
associated density function f (p). Again suppose the buyer is willing to accept all prices, then the
expected price to be paid by the buyer is
Z p̄ Z p̄
E [p] = pf (p) dp = pdF (p) (6.13)
p p
1
Recall that Bayes rule says for any two events, A and B, that
Pr (A ∩ B)
Pr (A |B ) = (6.11)
Pr (B)
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 71
which is the unconditional expectation of the price to be paid. Notice that the summation is now
replaced with the integral operation. Suppose next, that the buyer adopts a rule to accept all
prices higher or equal to p̂ and reject all other prices. (Recall this is just an example.) Then the
unconditional expectation of the price to be paid is
Z P̄
Ep̂ [p] = pdF (p) (6.14)
p̂
Again, the expectation in (6.14) is still unconditional since no new information entered in the
formation of the expectation and we represent the acceptance rule of the buyer by the index p̂
below the expectations operator E [·], as in Ep̂ [·]. The expected payment is then
Z p̂ Z p̄
Ep̂ [p] = 0dF (p) + pdF (p)
p p̂
as Z p̂
E [p] − Ep̂ [p] = pdF (p) > 0. (6.15)
p
since the buyer is willing to pay p in more circumstances in (6.13) and (6.14).
Consider now the average price which is paid under the two different policies conditional on
accepting a price and paying it, in other words, we ask for the conditional expectation:
Ep̂ [p |p ≥ p̂ ]
We consider now exclusively the events, i.e. realization of prices p, where p ≥ p̂ for some p̂. Notice
that we have now a continuum of events, each of which has probability zero, but all of which have
associated densities. We then switch to densities. The conditional densities of a price p being
quoted conditional on p ≥ p̂ is then given by
f (p)
f (p |p ≥ p̂ ) = . (6.16)
1 − F (p̂)
Compare this expression to (6.10) and (6.12) above and notice that this follows from
as long as p ≥ p̂, and otherwise the intersection is of course the empty set Notice that, as before
the conditional densities integrate up to 1 since
Z p̂ Z p̂ Z p̂
f (p) 1
f (p |p ≥ p̂ ) dp = dp = f (p) dp = 1, (6.17)
p̂ p̂ 1 − F (p̂) 1 − F (p̂) p̂
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 72
Finally, the conditional expectation of the prices, conditioned on accepting then price, is then
nothing else but taken the expectation with respect to the conditional probabilities rather than
with the unconditional probabilities, so that
Z p̂ Z p̂ Z p̂
f (p) 1
Ep̂ [p |p ≥ p̂ ] = pf (p |p ≥ p̂ ) dp = p dp = pf (p) dp.
p̂ p̂ 1 − F (p̂) 1 − F (p̂) p̂
Again you may verify that the expected price or average price paid, conditional on paying is
increasing in p̂, as we might have expected it. In particular,
You may wish to verify (6.18) and (6.19) to realize the difference between expected value (price)
and conditional expected value (price) in the second example.
The valuations vs and vb are independent. If sale occurs at p, buyer receives vb − p and seller
receives p − vs . Trade occurs at p = 21 (ps + pb ) , if ps ≤ pb . There is no trade if ps > pb . If there
is no trade both, buyer and seller receive zero payoff. Given vb , pb and ps , the buyer’s payoff is:
(
vb − 12 (ps + pb ) if pb ≥ ps
ub (vb , pb , ps ) =
0 if pb < ps
2
Recall the similarity between the second to last term in (6.17) to the term appearing when computing the
expected value of the bid ps (vs ).
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 73
6.6.2 Equilibrium
A strategy for each agent is then a mapping
pi : Vi → R. (6.22)
such that
Ep∗s [ub (vb , p∗b (vb ) , p∗s )] ≥ Ep∗s ub vb , p0b , p∗s , ∀vb , p0b ,
(6.23)
and
Ep∗b [us (vs , p∗s (vs ) , p∗b )] ≥ Ep∗b us vs , p0s , p∗b , ∀vs , p0s .
(6.24)
Remark 65 The Bayes part in the definition of the Bayes-Nash equilibrium through (6.23) and
(6.24) is that the buyer has to find an optimal bid for every value vb while he is uncertainty about
the ask (p∗s ) of the seller. From the buyer’s point of view, he is uncertain about p∗s as the ask,
p∗s = p∗s (vs ) will depend on the value the seller assigns to the object, but as this is the private
information of the seller, the buyer has to make his bid only knowing that the seller will adopt the
equilibrium strategy: p∗s : Vs → [0, 1], but not knowing the realization of vs and hence not knowing
the realization of p∗s = p∗s (vs ).
where
G∗ (x) = Pr (p∗s ≤ x)
is the distribution of the equilibrium asks of the seller. Notice now that the equilibrium ask, p∗s ,
of the seller will depend on the realization of the seller’s value vs .Thus we can equivalently, the
expectation in (6.25) with respect to vs , or
Z
∗ 1 ∗
Evs [ub (vb , pb , ps (vs ))] = vb − (pb + ps (vs )) dF (vs ) (6.26)
{vs |p∗s (vs )≤pb } 2
The advantage of the later operation is that we have direct information about the distribution of
values, whereas we have only indirect information about the distribution of the equilibrium asks.
Using (6.20), we get
Z
∗ 1 ∗
Evs [ub (vb , pb , ps (vs ))] = vb − (pb + ps (vs )) dvs . (6.27)
{vs |p∗s (vs )≤pb } 2
However, as we don’t know the exact relationship between vs and p∗s , we can’t take the expectation
yet. Similar to the auction example, we therefore make the following guess, which we shall verify
eventually about the relationship of vs and p∗s :
The final question then is what is the set of vs over which we have to integrate, or
The expected utility for the buyer is now expressed as a function of vb , pb and the constants in
the strategy of the seller, but independent of p∗s or vs . Simplifiying, we get
which essentially represents the trade-off for the buyer. A higher bid pb increases the probability
of trade but decreases the revenue conditional on trade. The first order conditions are
∂ub (vb , pb , as , cs ) 3pb + as 3
= vb − − (pb − as ) = 0
∂pb 4 4
or
2vb as
p∗b (vb ) =
+ . (6.32)
3 3
We can perform a similar steps of arguments for the seller to obtain
Z 1
1
(ps + ab + cb vb ) − vs dvb
ps −ab 2
cb
∂us (vs , ps , ab , cb )
=0
∂ps
or equivalently
1 1 2
p∗s (vs ) = cb + ab + vs (6.33)
3 3 3
From the solution of the two pricing policies, we can now solve for ab and as as well as
1 1
cb + ab = as
3 3
as
= ab
3
CHAPTER 6. GAMES OF INCOMPLETE INFORMATION 76
and hence
1 1
as = , ab =
4 12
and thus the complete resolution is
1 2
p∗b (vb ) = + vb
12 3
and
1 2
p∗s (vs ) =+ vs
4 3
Next we can evaluate the efficiency of the equilibrium. Since trade occurs if
1 2 1 2
p∗b (vb ) ≥ p∗s (vs ) ⇔ + vb ≥ + vs
12 3 4 3
1
⇔ vb − vs ≥
4
trade occurs if the gains from trade exceed 14 . The following strategies also form an equilibrium
for any x ∈ [0, 1] . ( (
x if vb ≥ x x if vs ≤ x
pb = ps = .
0 otherwise 1 otherwise
Readings. Chapter 3 in [4], [3] Chapter 6, [8], Chapter 8.E.
Chapter 7
where θi represent the marginal willingness to pay for quality qi and ti is the transfer (price)
buyer i has to pay for the quality qi . The taste parameters θi satisfies
The ex-ante (prior) probability that the buyer has a high willingness to pay is given by
p = Pr (θi = θh )
We also observe that the difference in utility for the high and low valuation buyer for any given
quality q
u (θh , q) − u (θl , q)
77
CHAPTER 7. ADVERSE SELECTION (WITH TWO TYPES) 78
is increasing in q. (This is know as the Spence-Mirrlees sorting condition.). If the taste parameter
θi were a continuous variable, the sorting condition could be written in terms of the second cross
derivative:
∂ 2 u (θ, q)
> 0,
∂θ∂q
which states that taste θ and quality q are complements. The profit for the seller from a bundle
(q, t) is given by
π (t, q) = t − c (q)
The efficient solution is the equilibrium outcome if either the monopolist can perfectly dis-
criminate between the types (first degree price discrimination) or if there is perfect competition.
The two outcomes differ only in terms of the distribution of the social surplus. With a perfectly
discriminating monopolist, the monopolist sets
ti = θ i q i (7.4)
using (7.4). Likewise with perfect competition, the sellers will break even, get zero profit and set
prices at
ti = c (qi∗ )
θ i q i − ti = 0 (IRi ) (7.7)
θi qi − ti = θi qj − tj (ICi ) (7.8)
The question is then how to separate. We will show that the binding constraint are IRl and ICh ,
whereas the remaining constraints are not binding. We then solve for tl and th , which in turn
allows us to solve for qh , and leaves us with an unconstrained problem for ql .Thus we want to
show
(i) IRl binding, (ii) ICh binding, (iii) q̂h ≥ q̂l (iv) q̂h = qh∗ (7.9)
θ h q h − th = θ h q l − tl = θ l q l − t l (7.10)
ICh θh >θl
suppose that θl ql −tl > 0, then we could increase tl , th by a equal amount, satisfy all the constraints
and increase the profits of the seller. Contradiction.
Consider (ii) Suppose not, then as
θ h q h − t h > θ h q l − tl = θ l q l − t l = 0 (7.11)
θh >θl (IRl )
and thus th could be increased, again increasing the profit of the seller.
(iii) Adding up the incentive constraints gives us (ICl ) + (ICl )
and since:
θh > θl ⇒ q̂h − q̂l = 0. (7.13)
tl = θ l q l (7.15)
CHAPTER 7. ADVERSE SELECTION (WITH TWO TYPES) 80
and
th − tL = θh (qh − qL ) ⇒ th = θh (qh − qL ) + θl ql .
but as qh∗ is just as constant, the optimal solution is independent of constant terms and we can
simplify the expression to:
Dividing by (1 − p) we get
p
max θl ql − c (ql ) − (θh − θl ) ql
ql 1−p
and the quality supply to the low valuation buyer is inefficiently low (with the possibility of
complete exclusion).
Consider next the information rent for the high valuation buyer, it is
I (ql ) = (θh − θl ) ql
and therefore the rent is increasing in ql which is the motivation for the seller to depress the
quality supply to the low end of the market.
The material is also covered in [14], Chapter 2.
Chapter 8
Theoretical Complements
is a strategy profile σ ∗ = (σ ∗1 , ..., σ ∗I ) such that, for all i, for all ti ∈ T and all si ∈ supp σ ∗i (ti ):
mixed strategy Nash
Bayes z }| {
z X }| { X
s0i , s−i ; (ti , t−i ) σ ∗−i (s−i |t−i )
pi (t−i |ti ) ui
t−i ∈T−i s−i ∈S−i
≥
mixed strategy Nash
Bayes z }| {
z X }| { X
s0i , s−i ; (ti , t−i ) σ ∗−i (s−i |t−i ), ∀s0i
pi (t−i |ti ) ui
t−i ∈T−i s−i ∈S−i
where σ ∗−i (s−i |t−i ) ≡ σ ∗1 (s1 |t1 ) , ..., σ ∗i−1 (si−1 |ti−1 ) , σ ∗i+1 (si+1 |ti+1 ) , ..., σ ∗n (sn |tn ) .
Since the strategy choices are independent across players, we can write that:
Y
σ ∗−i (s−i |t−i ) = σ ∗j (sj |tj ) .
j6=i
81
CHAPTER 8. THEORETICAL COMPLEMENTS 82
Condition 67 At each information set, the player with the move must have a belief about which
node in the information set has been reached by the play of the game. For a non singleton
information set, a belief is a probability distribution over the nodes in the information set; for a
singleton information set, the player’s belief puts probability one on the single decision node.
Condition 68 Given their beliefs, the players’ strategies must be sequentially rational. That is,
at each information set the action taken by the player with the move (and the player’s subsequent
strategy) must be optimal given the player’s belief at that information set and the other players’
subsequent strategies (where a “subsequent strategy” is a complete plan of action covering every
contingency that might arise after the given information set has been reached).
Condition 70 At information sets on the equilibrium path, beliefs are determined by Bayes’ rule
and the players’ equilibrium strategies.
Condition 71 At information sets off the equilibrium path beliefs are determined by Bayes’ rule
and the players’ equilibrium strategies where possible.
Definition 72 A perfect Bayesian equilibrium consists of strategies and beliefs satisfying Require-
ments 1 through 4.
CHAPTER 8. THEORETICAL COMPLEMENTS 83
Condition 74
1. For each mj in M, the Receiver’s action a∗ (mj ) must maximize the Receiver’s expected
utility, given the belief µ (ti |mj ) about which types could have sent mj . That is,
a∗ (mj ) solves
X
max µ (ti |mj )UR (ti , mj , ak ) . (8.2)
ak ∈A
ti ∈T
2. For each ti in T, Sender’s message m∗ (ti ) must maximize the Sender’s utility, given the
Receiver’s strategy a∗ (mj ) . That is, m∗ (ti ) solves:
Condition 75 For each mj in M , if exists ti in T such that m∗ (ti ) = mj , then the Receiver’s
belief at the information set corresponding to mj must follow from Bayes’ rule and the Sender’s
strategy:
p (ti )
µ (ti |mj ) = P . (8.4)
p (ti )
ti ∈T
We can summarize these results to get the definition of a Perfect Bayesian Equilibrium:
Definition 77 (PBE) A pure strategy Perfect Bayesian Equilibrium is a set of strategies {m∗ (ti ) , a∗ (mj )}
and posterior beliefs p∗ (ti |mj ) such that:
1. ∀ti , ∃p∗ (ti |mj ) , s.th. p∗ (ti |mj ) ≥ 0, and p∗ (ti |mj ) = 1;
P
ti ∈T
p (ti )
p (ti |mj ) = P 0 .
{t0i |m∗ (t0i )=mj } p (ti )
Example 79 Cho-Kreps (Beer-Quiche) - Entry Deterrence. Remark the payoffs in [4] have a
problem, they don’t allow for a separating equilibrium, modify)
Definition 81 (Intuitive Criterion) If the information set following mj is off the equilibrium
path and mj equilibrium dominated for type ti , then
µ (ti |mj ) = 0.
This is possible provided that mj is not equilibrium dominated for all types in T .
Readings: [8] Chapter 9.C, [3] Chapter 8.
Bibliography
[3] D. Fudenberg and J. Tirole. Game Theory. MIT Press, Cambridge, 1991.
[4] R. Gibbons. Game Theory for Applied Economists. Princeton University Press, Princeton,
1992.
[6] D.M. Kreps and R. Wilson. Sequential equilibria. Econometrica, 50:863–894, 1982.
[7] H. Kuhn. Extensive games and the problem of information. In H. Kuhn and A. Tucker,
editors, Contributions to the Theory of Games, pages 193–216. Princeton University Press,
Princeton, 1953.
[8] A. Mas-Collel, M.D. Whinston, and J.R. Green. Microeconomic Theory. Oxford University
Press, Oxford, 1995.
[9] J. Nash. Equilibrium points in n-person games. Proceedings of the National Academy of
Sciences, 36:48–49, 1950.
[10] M.J. Osborne and A. Rubinstein. A Course in Game Theory. MIT Press, Cambridge, 1994.
[11] P. Reny. On t he existence of pure and mixed strategy nash equilibria in discontinuous games.
Econometrica, 67:1029–1056, 1999.
[12] S.M. Ross. Probability Models. Academic Press, San Diego, 1993.
85