Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents



Similar documents
Notes on metric spaces

10 Evolutionarily Stable Strategies

Maximum Likelihood Estimation

A Simple Model of Price Dispersion *

TOPIC 4: DERIVATIVES

Choice under Uncertainty

I. Pointwise convergence

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

U = x x x 1 4. What are the equilibrium relative prices of the three goods? traders has members who are best off?

Notes from Week 1: Algorithms for sequential prediction

1 Sufficient statistics

ALMOST COMMON PRIORS 1. INTRODUCTION

ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Lecture Notes on Elasticity of Substitution

x a x 2 (1 + x 2 ) n.

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

The Ergodic Theorem and randomness

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

Practice with Proofs

Separation Properties for Locally Convex Cones

ECON Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted.

2. Information Economics

Online Appendix Feedback Effects, Asymmetric Trading, and the Limits to Arbitrage

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

Bargaining Solutions in a Social Network

Nonparametric adaptive age replacement with a one-cycle criterion

The Real Business Cycle Model

Dynamics and Equilibria

ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE

Schooling, Political Participation, and the Economy. (Online Supplementary Appendix: Not for Publication)

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

1.7 Graphs of Functions

1 if 1 x 0 1 if 0 x 1

A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails

6.207/14.15: Networks Lectures 19-21: Incomplete Information: Bayesian Nash Equilibria, Auctions and Introduction to Social Learning

Lecture 4: BK inequality 27th August and 6th September, 2007

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Multi-variable Calculus and Optimization

Homework # 3 Solutions

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Constrained optimization.

ON PRICE CAPS UNDER UNCERTAINTY

Mathematical finance and linear programming (optimization)

Consumer Theory. The consumer s problem

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

BANACH AND HILBERT SPACE REVIEW

Lecture 5 Principal Minors and the Hessian

Economics 2020a / HBS 4010 / HKS API-111 FALL 2010 Solutions to Practice Problems for Lectures 1 to 4

CPC/CPA Hybrid Bidding in a Second Price Auction

LECTURE 15: AMERICAN OPTIONS

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

Online Appendix for Student Portfolios and the College Admissions Problem

Economics of Insurance

Evolutionary dynamics and backward induction

AN ANALYSIS OF A WAR-LIKE CARD GAME. Introduction

Cost Minimization and the Cost Function

n k=1 k=0 1/k! = e. Example 6.4. The series 1/k 2 converges in R. Indeed, if s n = n then k=1 1/k, then s 2n s n = 1 n

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

A New Interpretation of Information Rate

Lecture Notes 10

36 CHAPTER 1. LIMITS AND CONTINUITY. Figure 1.17: At which points is f not continuous?

A Portfolio Model of Insurance Demand. April Kalamazoo, MI East Lansing, MI 48824

arxiv: v1 [math.pr] 5 Dec 2011

Gambling Systems and Multiplication-Invariant Measures

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

1 Portfolio mean and variance

4 The M/M/1 queue. 4.1 Time-dependent behaviour

A characterization of trace zero symmetric nonnegative 5x5 matrices

Section 1.3 P 1 = 1 2. = P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., =

6.4 Logarithmic Equations and Inequalities

1 The EOQ and Extensions

Web Appendix for Reference-Dependent Consumption Plans by Botond Kőszegi and Matthew Rabin

0 0 such that f x L whenever x a

Chapter 6. Cuboids. and. vol(conv(p ))

Mathematical Methods of Engineering Analysis

MA107 Precalculus Algebra Exam 2 Review Solutions

Lectures 5-6: Taylor Series

RANDOM INTERVAL HOMEOMORPHISMS. MICHA L MISIUREWICZ Indiana University Purdue University Indianapolis

Online Appendix. Supplemental Material for Insider Trading, Stochastic Liquidity and. Equilibrium Prices. by Pierre Collin-Dufresne and Vyacheslav Fos

The Heat Equation. Lectures INF2320 p. 1/88

4 Lyapunov Stability Theory

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan , Fall 2010

Real Roots of Univariate Polynomials with Real Coefficients

Jung-Soon Hyun and Young-Hee Kim

SPARE PARTS INVENTORY SYSTEMS UNDER AN INCREASING FAILURE RATE DEMAND INTERVAL DISTRIBUTION

Principle of Data Reduction

1 Error in Euler s Method

Continuity of the Perron Root

Bayesian logistic betting strategy against probability forecasting. Akimichi Takemura, Univ. Tokyo. November 12, 2012

Equations, Inequalities & Partial Fractions

1 Norms and Vector Spaces

6 EXTENDING ALGEBRA. 6.0 Introduction. 6.1 The cubic equation. Objectives

Transcription:

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents William H. Sandholm January 6, 22 O.. Imitative protocols, mean dynamics, and equilibrium selection In this section, we consider stochastic evolution under two well-known imitative protocols: imitation of success and imitation driven by dissatisfaction. Example O... Imitation of success. Suppose that when an agent receives a revision opportunity, he randomly samples an opponent, switching to the opponent s strategy with probability proportional to the difference between the opponent s payoff and some fixed baseline m R. This protocol is described by the conditional imitation probabilities r i π) λπ m) for i, O.) where λ > and m are chosen to ensure that conditions B) and ) both hold. If there are no mutations, then substituting into equation M ) and simplifying shows that finite-horizon behavior under this protocol is described by the mean dynamic ẋ λx x )F x ) F x )). O.2) This is the replicator dynamic of Taylor and Jonker [4], with the constant λ representing a uniform change in speed. Department of Economics, University of Wisconsin, 8 Observatory Drive, Madison, WI 5376, USA. e-mail: whs@ssc.wisc.edu; website: http://www.ssc.wisc.edu/ whs. Early analyses of these protocols include Börnerstedt and Weibull [], Weibull [5], and Hofbauer [2].

Example O..2. Imitation driven by dissatisfaction. Suppose that a revising agent compares his own payoff to some aspiration level M R. He then opts to switch strategies with probability proportional to the amount by which his payoff is deficient, switching to the strategy of a randomly chosen opponent in this event. This rule is captured by the conditional imitation probabilities r i π) λm π i ) for i, O.3) where λ > and M are such that B) and ) both hold. If there are no mutations, substituting into equation M ) and simplifying shows that this protocol too generates the replicator dynamic O.2) as its mean dynamic. Since protocols O.) and O.3) generate the same mean dynamic, the deterministic approximation theorem described in Section 3 implies that in large populations, the finitehorizon aggregate behavior traectories generated by the two protocols are essentially indistinguishable. Despite this strong agreement, we now use our results from Section 4 to show that infinite-horizon behavior under protocols O.) and O.3) can be quite different: the two protocols can generate different stochastically stable states, even in very simple games. Example O..3. Consider a population of agents matched to play the symmetric normal form game h A b h B with b < h < B, which we call Boar Hunt. The strategies in this coordination game are hunting for Hare strategy ) and hunting for Boar strategy ). Hare yields a certain payoff of h, while coordinating on Boar yields the highest possible payoff of B. population of agents are matched to play Boar Hunt, the resulting population game has payoffs F x ) h and F x ) b x ) + Bx. In addition to the all-hare x ) and all-boar x ) equilibria, there is a mixed equilibrium in which fraction x h b of the B b population chooses Boar. ote that either strategy in Boar Hunt can be risk dominant. If committed agents are present, Theorems 4.2 and 4.3 tell us that the state that is stochastically stable in the large population limit is the one that maximizes the ordinal potential function J. It is easy to check that since F is a coordination game, the ordinal potentials induced by protocols O.) and O.3) are strictly convex. Thus, if J) < J), x is stochastically stable, while if J) >, state x is stochastically stable. If a 2

..5..5..2.4.6.8. Figure : The ordinal potentials J S dashed) and J D solid) when m, b, h 2, B 3, and M 4. This criterion is easy to evaluate. Under imitation of success O.), the value of the ordinal potential function at x is J S ) log F y) m B m dy log F y) m h m + b m B b log B m b m. Under imitation driven by dissatisfaction O.3), this value is J D ) log M F y) M h dy log M F y) M B + M b B b log M B M b +. To make this example more concrete, let us fix m, b, h 2, B 3, and M 4. Then the mixed ash equilibrium of F is x h b, implying that neither strategy is B b 2 strictly risk dominant; obviously, the all-boar equilibrium x is payoff dominant. The functions J S and J D are graphed in Figure. Since J S ) 3 log 3 log 2.452, 2 imitation of success leads to the selection of the payoff dominated all-hare equilibrium, x. But since and J D ) J S ).452, imitation driven by dissatisfaction leads to the selection of the payoff dominant all-boar equilibrium, x. Clearly, the same qualitiative results obtain in any game with similar payoff values, implying that either protocol can select or fail to select a strictly risk dominant equilibrium. How do the differences between the protocols lead to the different equilibrium selections? In general, the identity of the stochastically stable state is determined by the relative unlikelihoods of excursions between equilibria: here, from all-hare to all-boar, and from all-boar to all-hare. The mean dynamic M ), computed as the differences between the probability p x x x )r Fx )) of a switch from to and the probability 3

q x x x )r Fx )) of a switch from to, provides some information about the relative unlikelihoods of excursions. But the stationary distribution 8), being the limiting distribution of the Markov process, is determined by products of ratios of the one-step transition probabilities. Ultimately, the stationary distribution weights depend on the ratios of the imitation probabilities r ) and r ), or, equivalently, on the differences between the logarithms of r ) and r ). This is evident in the definition ) of the ordinal potential function J, which allows us to write J) as J) log r Fy)) dy log r Fy)) dy. O.4) Equation O.4) says that the stochastically stable state is determined by comparing the averages over population states x of log r Fx )) and log r Fx )). Since we are averaging logarithms of imitation probabilities, the manner in which the levels of dispersion of these probabilities depend on the direction of imitation from Hare to Boar or from Boar to Hare plays a basic role in explaining infinite-horizon behavior. As x varies from to, the payoff to Hare is constant at F x ) 2, while the payoff to Boar, F x) 2x +, is uniformly distributed between and 3. Under imitation of success O.), the conditional imitation probabilities are given by r x ) λf x ) m) λf x ) and r x ) λf x ) m) 2λ. 2 If we view F x ) as a random variable with a uniform[, 3] distribution, then r x ) has a uniform[λ, 3λ] distribution. Thus, since the logarithm function is concave, Jensen s inequality implies that the first integral in O.4) is smaller than the second, and hence that the all-hare equilibrium is stochastically stable. Intuitively, the conditional imitation probabilities in each direction have the same mean, but only those from Hare to Boar are variable. This implies that the Hare to Boar transition is less likely than its opposite, and so that the all-hare equilibrium is selected. Under imitation driven by dissatisfaction O.3), the conditional imitation rates are given by r x ) λm F x )) 2λ and r x ) λ4 F x )). This time, the conditional imitation probabilities from Hare to Boar are fixed, while those from Boar to Hare are variable; the latter transition is therefore less likely, and the all-boar equilibrium is stochastically stable. Maruta [3] also considers a stochastic evolutionary model in which decision rules focusing on the payoff of the current strategy favor selection of the payoff dominant equilibrium, while decision rules focusing on the payoff of the alternative strategy favor selection of the safe equilibrium. Maruta s [3] model differs from ours in a number of 2 Recall that the constant λ > was introduced in O.) to ensure that r x ) and r x ) are less than. 4

important respects: for example, choices in his model are not based on imitation, and his analysis concerns the small noise limit rather than the large population limit. O.2. The proof of Theorem 4. When there are no committed agents and a mutation rate of ε as described in protocol 3), the stationary distribution of the stochastic evolutionary process takes the form µ,ε x µ,ε x p,ε )/ q,ε / x + ε) r F )) + ε) ε) r. O.5) F )) + ε) To prove the first statement in part i), we expand equation O.5) with x to obtain µ,ε µ,ε ε ε)r F )) + ε 2 + ε) r F )) + ε ε) r F )) + ε. ε) r F )) + ε ε) r F )) + ε ε)r F )) + ε ε Since the ε terms cancel, none of the terms in the final product converge to zero or infinity. Indeed, we have lim log µ,ε ε µ,ε log r F )) r F )) so bound B) and the bounded convergence theorem imply that lim lim ε µ,ε log µ,ε lim log r F )) r F )), log r F )) J). r F )) To prove the second statement in part i), observe that for x X {, }, µ,ε x µ,ε ε ε)r F )) + ε x 2 + ε) r F )) + ε ε) r F )) + ε by equation O.5). Since all terms except the initial ε approach positive constants as ε approaches zero, µ,ε x /µ,ε is in Θε). The analysis of µ,ε x /µ,ε is similar. To prove part ii) of the theorem, we note from equation O.5) that 5

log µ,ε x µ,ε x log + log + log ε) r F )) + ε. ε) r F )) + ε Comparing this equation to equation 9) and the subsequent arguments, it is easy to verify that the remainder of the proof is identical to that of Theorem 4.3ii). O.3. The proof of the Theorem 4.4) By substituting the committed agents protocol 4) with c i 7), we obtain the one-step transition probabilities γ i into equations 6) and p x x ) x +γ ) +γ T ) r F x )) and q x x x +α ) +γ T ) r F x )), where γ T γ +γ. Inserting these expressions into equation 8), we find that the stationary distribution of the process {X t } is given by µ x µ x p )/ q / x + +γ r +γ T ) F )) +γ r +γ T ) F )). O.6) It follows that log µ x µ x J x ) + log r F )) + log r F )) x log + + +γ +γ T ) +γ +γ T ) log + log +γ +γ T ) log +γ +γ T ) ). O.7) The proof of Theorem 4.2 shows that J converges uniformly to J. Moreover, applying the dominated convergence theorem as in the proof of Theorem 4.3 shows that the second term of O.7) converges uniformly to x log y) log y + log y+γ +γ T log y+γ +γ T ) dy x log x x ) log x ) + x + γ ) logx + γ ) γ log γ + x + γ ) log x + γ ) + γ ) log + γ ) 6

L γ x ). Thus since J γ x ) Jx ) + L γ x ), we find that lim max x X log µ x µ J γ x ). The proof is completed by combining this conclusion with a slight variation on the last part of the proof of Theorem 4.2. References [] J. Börnerstedt, J. W. Weibull, ash equilibrium and evolution by imitation, in: K. J. Arrow, et al. eds.), The Rational Foundations of Economic Behavior, St. Martin s Press, ew York, 996, pp. 55 8. [2] J. Hofbauer, Imitation dynamics for games, unpublished manuscript, University of Vienna 995). [3] T. Maruta, Binary games with state dependent stochastic choice, Journal of Economic Theory 3 22) 35 376. [4] P. D. Taylor, L. Jonker, Evolutionarily stable strategies and game dynamics, Mathematical Biosciences 4 978) 45 56. [5] J. W. Weibull, Evolutionary Game Theory, MIT Press, Cambridge, 995. 7