3 Alternating O ers Bargaining

Similar documents
6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

Oligopoly markets: The price or quantity decisions by one rm has to directly in uence pro ts by other rms if rms are competing for customers.

UCLA. Department of Economics Ph. D. Preliminary Exam Micro-Economic Theory

Problem Set 2 Due Date: March 7, 2013

The Prison S Dilemma and Its Connections

CeDEx Discussion Paper Series. Discussion Paper No J. Arin, V. Feltkamp and M. Montero July 2012

Adverse Selection. Chapter 3

ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY

36 CHAPTER 1. LIMITS AND CONTINUITY. Figure 1.17: At which points is f not continuous?

How To Play A Static Bayesian Game

Cooperation with Network Monitoring

Lecture Notes 10

Exact Nonparametric Tests for Comparing Means - A Personal Summary

Week 7 - Game Theory and Industrial Organisation

Representation of functions as power series

Game Theory and Algorithms Lecture 10: Extensive Games: Critiques and Extensions

FINAL EXAM, Econ 171, March, 2015, with answers

EconS Advanced Microeconomics II Handout on Cheap Talk

CAPM, Arbitrage, and Linear Factor Models

Sequential lmove Games. Using Backward Induction (Rollback) to Find Equilibrium

Climate-Change Treaties: A Game-Theoretic Approach Roy Radner Stern School, New York University

Competition and Regulation. Lecture 2: Background on imperfect competition

Psychology and Economics (Lecture 17)

Backward Induction and Subgame Perfection

Théorie de la décision et théorie des jeux Stefano Moretti

c 2008 Je rey A. Miron We have described the constraints that a consumer faces, i.e., discussed the budget constraint.

Midterm March (a) Consumer i s budget constraint is. c i b i c i H 12 (1 + r)b i c i L 12 (1 + r)b i ;

5.1 Radical Notation and Rational Exponents

University of Oslo Department of Economics

Buying shares and/or votes for corporate control

9 Repeated Games. Tomorrow, and tomorrow, and tomorrow, Creeps in this petty pace from day to day To the last syllable of recorded time Shakespeare

Decentralised Bilateral Trading in a Market with Incomplete Information

Game Theory in Wireless Networks: A Tutorial

On the incentives of an integrated ISP to favor its own content

AN INTRODUCTION TO GAME THEORY

The Binomial Distribution

8 Divisibility and prime numbers

Christmas Gift Exchange Games

Market Power, Forward Trading and Supply Function. Competition

Interlinkages between Payment and Securities. Settlement Systems

Universita Commerciale Luigi Bocconi. Facolta di Economia. Corso di Laurea: Economics and Social Sciences. Titolo Tesi: Agreements in Dynamic Games

Universidad Carlos III de Madrid Game Theory Problem Set - Dynamic Games

Game Theory 1. Introduction

1 Further Pricing Relationships on Options

Table 10.1: Elimination and equilibrium. 1. Is there a dominant strategy for either of the two agents?

Working Paper An exact non-cooperative support for the sequential Raiffa solution. Working papers // Institute of Mathematical Economics, No.

Perfect Bayesian Equilibrium

Tiers, Preference Similarity, and the Limits on Stable Partners

Game Theory: Supermodular Games 1

Advanced Microeconomics

Web Appendix for Reference-Dependent Consumption Plans by Botond Kőszegi and Matthew Rabin

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

Investing without credible inter-period regulations:

Infinitely Repeated Games with Discounting Ù

Risk Aversion. Expected value as a criterion for making decisions makes sense provided that C H A P T E R Risk Attitude

21. Unverifiable Investment, Hold Up, Options and Ownership

Lecture 11: The Design of Trading Platforms (Alos-Ferrer et al. 2010)

Mathematical Induction

ECON Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted.

The Ascending Auction Paradox

Microeconomic Theory Jamison / Kohlberg / Avery Problem Set 4 Solutions Spring (a) LEFT CENTER RIGHT TOP 8, 5 0, 0 6, 3 BOTTOM 0, 0 7, 6 6, 3

Dynamics and Stability of Political Systems

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

Chapter 7. Sealed-bid Auctions

Partial Fractions Decomposition

3. Mathematical Induction

Working Paper Series ISSN Working Paper No Stability for Coalition Formation Games. Salvador Barberà and Anke Gerber

Economics 159: Introduction to Game Theory. Summer 2015: Session A. Online Course

Solutions for Practice problems on proofs

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

Herding, Contrarianism and Delay in Financial Market Trading

A Simpli ed Axiomatic Approach to Ambiguity Aversion

Heterogeneous Expectations, Adaptive Learning, and Evolutionary Dynamics

Capital Structure. Itay Goldstein. Wharton School, University of Pennsylvania

WARWICK ECONOMIC RESEARCH PAPERS

Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products

Dynamic Pricing of Enterprise Software with Value Uncertainty: Motivation for Selling Software as a Service

Introduction to Binomial Trees

Math 4310 Handout - Quotient Vector Spaces

1.2 Solving a System of Linear Equations

I. Noncooperative Oligopoly

Paid Placement: Advertising and Search on the Internet

Normalization and Mixed Degrees of Integration in Cointegrated Time Series Systems

Chapter 13: Strategic Decision Making in Oligopoly Markets

0.0.2 Pareto Efficiency (Sec. 4, Ch. 1 of text)

Class Notes, Econ 8801 Lump Sum Taxes are Awesome

Sample Induction Proofs

Math 55: Discrete Mathematics

Industry profit in an oligopoly (sum of all firms profits) < monopoly profit.

Central Bank Lending and Money Market Discipline

7: The CRR Market Model

Formal Languages and Automata Theory - Regular Expressions and Finite Automata -

The Economics of the Marriage Contract: Theories and Evidence *

Chapter 1: The Modigliani-Miller Propositions, Taxes and Bankruptcy Costs

Bayesian Nash Equilibrium

Full and Complete Binary Trees

Decision Trees. The analysis of complex decisions with signi cant uncertainty can be confusing C H A P T E R Decision Trees

Games of Incomplete Information

On the Importance of Default Breach Remedies

Discussion of Self-ful lling Fire Sales: Fragility of Collateralised, Short-Term, Debt Markets, by J. C.-F. Kuong

Transcription:

3 Alternating O ers Bargaining Consider the following setup following Rubinstein s (982) extension of a model by Stahl (972): Two players bargain over a pie with size normalized to. Given a division (x ; x 2 ) with x 0; x 2 0 and x + x 2 ; the instantaneous utility for the players is x and x 2 respectively. Hence, we assume that i) agents care only about their own slice of the pie, ii) risk neutrality. Payo s are discounted geometrically at (common) rate ; so, from the point of view of the beginning of the game, the utility from agreeing on (x ; x 2 ) at time t is t x and t x 2 Game has n periods (n = ) will be considered. The extensive form is as follows. In every odd period t = ; 3; 5; ::: < n, if the players have still not agreed on a division, rst player makes an o er (x t ; x t 2), which player 2 either accepts or rejects. If player 2 accepts, the game ends with (time t) payo s (x t ; x t 2) [in terms of time 0 utility, the payo s are t x t ; t x2 t ]. If player 2 rejects; the game proceeds to the next period. In even periods t = 2; 4; 6; ::: < n everything is just like odd periods except that 2 makes the o er and accepts or rejects. If t = n the players receive some exogenous division (s ; s 2 ) [in terms of time 0 utility, the payo s are ( n s ; n s 2 )]. It will be useful to translate payo s into time t units. Informally: De nition The continuation payo of a strategy pro le in a subgame starting at time t is the utility in time t units of the outcome induced by the strategy pro le. Below, you will see examples where the continuation payo is the same for all histories, but, in general, this is usually not the case. 25

3. The Backwards Induction Equilibrium with 3 Periods The last non-terminal nodes in the game are the acceptance decisions by player at time 2; where we see that:. Player must accept if o ered x 2 > s 2. Player is indi erent if o ered x 2 = s 3. Player must reject if o ered x 2 < s : However, we note that if player would reject s ; then the best response problem for player 2 at time 2 would be ill-de ned due to an openness issue. It follows: Claim In any backwards induction equilibrium it must be that player accepts an o er x 2 at time 2 if and only if x 2 s : proofsuppose that would reject s with probability r 2 (0; ) : Then, the expected payo for player 2 is ( r) ( s ) + rs 2 ( r) ( s ) + r ( s ) = ( s ) r [ ] : If 2 o ers s + "; player accepts for sure and the expected payo for 2 is s " > ( s ) r [ ] whenever " < r [ ] : Since r ( ) > 0 we may for example let " = Hence, player 2 will o er s and keep s : It follows that: r( ) 2 > 0: Claim In any backwards induction equilibrium it must be that player 2 accepts an o er x 2 at time if and only if x 2 [ s ] The proof is the same as above. Hence, 26

Claim There is a unique backwards induction equilibrium in which the equilibrium outcome is that player o ers division x ; x2 2 = ( [ s ] ; [ s ]) in the rst period and player 2 accepts. 3.2 A Stationary Equilibrium in the In nite Horizon Model Suppose that s solves or s = [ s ] s = + 2 = + : Using on of the claims above we see that player 2 will (in the 3 period game) o er division (s ; s ) = + ; + in period 2 and that player will o er (s ; s ) = + ; + in period. The critical observation is that the equilibrium division/payo s is identical with the exogenous third period payo s. Hence, we can add two, four, six or any even number of period and use the same argument recursively to conclude that the unique backwards induction equilibrium with (ad hoc) last period payo s (odd number of periods) given by ; + + will have player proposing ; + + in every odd period and player 2 proposing + ; + in every even period and each player accepting if and only if they are o ered at least : + In the in nite horizon game, we cannot backwards induct starting at the nal period, but we can easily use the recursive structure to construct a stationary equilibrium where,. In every odd period t after any sequence of rejected o ers player o ers a division (s t ; s t 2) = ; + + and player 2 accepts any o er such that x t 2 : + 27

2. In every even period t after any sequence of rejected o ers player 2 o ers a division (s t ; s t 2) = ; + + and player accepts any o er such that x t 2 : + To verify that this is an equilibrium, we note that the value of the game for player i is V t i = + in the beginning of every period t in which i makes a proposal (odd for and even for 2). Hence, given these continuation payo s, the unique acceptance rule is to accept a period t proposal if and only if x t i + : Finally, given that in every period and after any history of play a proposal is accepted by the agent not making a proposal if and only if x t i is to keep + and give +, the optimal proposal in every period + to the other agent. This veri es that the strategies speci ed are consistent with a backwards induction equilibrium [strictly speaking we would have to say that they are subgame perfect]. 28

4 Subgame Perfection Quit,, 3,3,3 High I Play II Up III Low High 0,0,0 0,0,0 Down Low 2,2,2 Figure : An example where backwards induction cannot be applie, but where the idea of sequential rationality generalizes Consider Figure. Clearly, (Quit, Up, Low) is a Nash equilibrium as a deviation by would give a payo of 0 instead of. Deviations by players 2 and 3 don t change the outcome, so we conclude that the strategy pro le indeed is a Nash equilibrium. Still, there is something implausible about this equilibrium in that if the game started at the node where player 2 decides between Up and Down, then (Up, Low) would not be a Nash equilibrium. That is, there is something non-credible about assuming that players 2 and 3 would play (Up, Low) following play. If the option Quit would be eliminated for player, then this play would not be consistent with Nash equilibrium in the reduced game. However, we cannot strictly speaking apply backwards induction as this game has imperfect information. The generalization of backwards induction for games with imperfect information is called subgame perfection. Intuitively, the equilibrium concept simply requires play to be Nash equilibrium play in any part of the game that can be thought of as a game in itself. 29

De nition 2 A subgame of an extensive form game K is a subset K 0 of the game that could be analyzed as a stand alone game with the properties:. There exists a node t 0 such that t 0 t for all t in the subgame K 0 and an information set h j such that h j = ft 0 g :(root is singleton information set) 2. For every node t in the subgame K 0, if t 2 h j and et 2 h j ; then et is also a node in the subgame (no broken information sets). In a nite game of perfect information, every node is the beginning of a subgame. De nition 3 A strategy pro le s is called subgame perfect if it induces Nash equilibrium play in every subgame K 0 of K: 4. A Slight Problem With Subgame Perfection I " " 2,0 " """ I II 5, 0,0 4,0 I 2,0 II 5, 0,0 4,0 -, -, Figure 2: Equivalent Games with Di erent Predictions of Subgame Perfection Consider the two extensive forms in Figure 2. We see that: 30

Both games have the same reduced normal form L R U 2; 0 2; 0 T 5; 0; 0 B 4; 0 ; The reduced normal form has two Nash equilibria (U,R) and (T,L) In the game to the left, we note that T strictly dominates B in the subgame. The unique Nash in the subgame is thus (T,L). Hence, the unique subgame prefect equilibrium is (AT,L) : In the game to the right, there is no other subgame than the full game. Hence, both (T,L) and (U,R) are subgame perfect. Disturbing example as we think that the (reduced) normal form and the extensive form are just di erent ways of modelling the same strategic situation. 4.2 Repeated Games Example : Consider the following Normal form game, where the Prisoner s dilemma has been extended with a punishment strategy P, C D P C ; ; 2 2; 2 D 2; 0; 0 2; 2 :: The unique Nash equilibrium is (D,D). P 2; 2 2; 2 3; 3 Consider the case where this repeated twice. Consider the following strategy:. In period, play s i = C 3

2. In period 2, play 8 < D if (C; C) in period s 2 i = : P if (a ; a 2) 6= (C; C) in period If both rms play in accordance to these strategies they each get + 0 = 3. Suppose that a player deviates in the rst period, then the best deviation is D and the payo is at most 2 + ( 2) = 0: 4. Also, no second period deviation is pro table as equilibrium play prescribes (C,C) in the second period, which is a static Nash equilibrium. Hence, we have speci ed a Nash equilibrium. 5. Not subgame prefect as the construction relies on play di erent from (D,D) in nal period. Example 2 Now consider C D A B C 5; 5 ; 6 ; ; D 6; 0; 0 ; ; A ; ; 2; 2 0; 0 B ; ; 0; 0 ; We see that (D,D) ; (A,A) ;and (B,B) are all equilibria. Consider the following strategy:. In period, play s i = C 2. In period 2, play A of (s ; s 2) = (C; C) 32

3. In period 2, play D of (s ; s 2) 6= (C; C) If player i follows the candidate equilibrium strategy the payo is 5 + 2 = 7 If player i wold deviate in the rst period, the best deviation is D and the payo is 6 + 0 = 6 < 7: Since play in the second period is Nash after any history we conclude that we ve constructed a subgame perfect equilibrium Remark The key idea, which is used in many contexts, is that multiplicity of equilibria allows us to create credible punishments. If players behave then play a good equilibrium in later stages. I players don t behave, then play a bad equilibrium in the later stages. 5 In nitely Repeated Games Consider a stage game: G = (I; A; u); where A = n i=a i are the action spaces (strategy spaces in one-shot game) and u i : A! R: Also suppose that payo s are discounted by 2 (0; ) : Let G (; ) denote the in nitely repetition of G given discount factor : We now note that:. Under the assumption that we have perfect monitoring so that all actions in previous rounds are observable we have that a history is the collection of all actions up to the last period. That is H t = A t : That is, H 0 = f?g ; H = A; H 2 = A A; ::: An information set is thus identi ed by the history leading up to that information set. 33

2. A strategy is a full contingent plan of action. One way to think of this is as thinking of a strategy as a sequence s i = s i ; s 2 i ; ::::; s t i; ::: s i 2 A i s t i : H t! A i Sometimes it is useful/more elegant to note that we may write H = [ t=h t s i : H! A i 3. Given a strategy s = (s ; ::; s n ) the outcome path is given by a (s) = a ; a 2 ; :::; a t ; ::: a = s ; s 2; :::; sn 2 A a 2 = s 2 a ; s 2 2 a ; :::; s 2 n a :::: a t = s t a ; ::::; a t ; s t 2 a ; ::::; a t ; :::; s t n a ; ::::; a t = s t (h t ) ; s t 2 (h t ) ; :::; s t n (h t ) after introducing the convenient notation h t = a ; ::::; a t 4. Payo s u i (s) = X t u i t= a t (s) A Nash equilibrium is de ned in the obvious/standard way: De nition 4 s is a Nash equilibrium if u i (s ) u i s i ; s i or every i 2 I and si : H! A i : 34

For subgame perfection we have to introduce some notation to be able to say that starting after any sequence of realized actions all players are playing a Nash equilibrium. Given strategy s i write s i jh t to denote the continuation strategy following history h t : Just like s i we have that s i jh t : H! A i : Stack strategies h and h t into h +t = (h ; h t ) and note that s t ijh (h t ) = s t+ i (h ; h t ) De nition 5 s is a subgame perfect equilibrium in G (; ) if for all histories h t we have that s jh t = (s jh t ; :::; s njh t ) is a Nash equilibrium in subgame after history h t : 5. Cooperation in an In nitely Repeated Prisoner s Dilemma Consider the Prisoner dilemma stage game D C D ; 4; 0 C 0; 4 3; 3 5.. GRIM-TRIGGER In any nite repetition we know that (D; D) in every period is the unique (Nash) equilibrium outcome. However, the in nite repetition allows us to consider (for example) the following GRIM-TRIGGER strategies where each players starts by cooperating and cooperates until somebody defects. In case somebody has defected, then all players defect for ever. That is, s i = C 8 < C s t i = : D if h t = (CC; CC; ::::; CC) otherwise 35

Obviously, the outcome from both players playing GRIM-TRIGGER is CC in every period. Now, consider a deviation on the equilibrium path. If player i follows the recommended play, the payo is whereas a deviation gives Hence, u i (s) = X t=0 u i (snd) = 4 + t 3 = 3 ; : 3 3 4 +, 4, 3 4 ( ), 3, 3 Claim s is a Nash equilibrium whenever 3 : We checked deviations on the equilibrium path above. However, o the equilibrium path play is (D; D) which is the unique stage game Nash. Hence, Claim s is a subgame perfect Nash equilibrium whenever 3 : 5..2 Tit For Tat Instead consider, s i = C 8 < s t i (h t ) = : C if a t j D if a t j = C = D The outcome is (C; C) in every period, which gives a payo of u i (s) = 3 : Checking all possible deviations pretty tricky in this case. Will talk about one shot deviation principle next. 36

To see the need/use. Consider deviating once and then following the speci ed strategy. The play will then evolve according too (D; C) ; (C; D) ; (D; C) ; :::: and the payo from such a deviation is u sns i = D = 4 + 0 + 2 4 + 3 0 + 4 4::: u sns i = D = 2 4 Another possibility is to deviate once and then avoid punishing the opponent for the punishment in the next period in which case we get (D; C) ; (C; D) ; (C; C) ; (C; C) :::: This deviation gives a payo of We then see that So: 4 + 2 3 4 + 0 + 2 3 : 2 2 4 = 4 ( ) ( + ) + 3 2 2 4 = 3 = 2 2 [3 ] + The second deviation is better than the rst if 3 The equilibrium candidate is better than second deviation if 3 3 + 3 + 2 3 4 + 2 3 4 + 2 3 3 : 37

We conclude that, under the assumption that 3 ; then:. 4 + 2 3 must be the best deviator payo (since 0 always follows D). Hence, Tit-for-Tat is a Nash equilibrium given the restriction 3. 2. However, Tit-for-Tat is not subgame perfect. After (D; C) the strategy prescribes (C; D) ; (D; C) ; (D; C) :::: and we concluded already that the strategy where would play C (against C) after the punishment in the rst period after the deviation (C; D) does better One way to make the Tit-for-Tat subgame perfect is to amend it so that agents take part in the punishments of themselves. That is s i = C 8 < s t i (h t ) = : C if a t ; a t 2 = (C; C) : 6= (C; C) D if a t ; a t 2 Here, any unilateral deviation prescribes (D; D) in the next period. 38