Evolutionary Game Theory

Similar documents
10 Evolutionarily Stable Strategies

ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY

Computational Learning Theory Spring Semester, 2003/4. Lecture 1: March 2

From the probabilities that the company uses to move drivers from state to state the next year, we get the following transition matrix:

6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

Mistakes Allow Evolutionary Stability in the Repeated Prisoner's Dilemma Game. I. Introduction

ECON Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted.

Chapter 7. Evolutionary Game Theory

How to Solve Strategic Games? Dominant Strategies

1 Nonzero sum games and Nash equilibria

The Basics of Game Theory

FINAL EXAM, Econ 171, March, 2015, with answers

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

Optimization in ICT and Physical Systems

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents

Game Theory 1. Introduction

12 Evolutionary Dynamics

A Note on Best Response Dynamics

Sequential lmove Games. Using Backward Induction (Rollback) to Find Equilibrium

9 Repeated Games. Tomorrow, and tomorrow, and tomorrow, Creeps in this petty pace from day to day To the last syllable of recorded time Shakespeare

COMP310 MultiAgent Systems. Chapter 11 - Multi-Agent Interactions

6.254 : Game Theory with Engineering Applications Lecture 1: Introduction

Week 7 - Game Theory and Industrial Organisation

Infinitely Repeated Games with Discounting Ù

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Linear Programming: Chapter 11 Game Theory

Course Outline. Parental care and sexual conflict. Papers for 22 October. What is sexual conflict? 10/19/2009

Parental care and sexual conflict.

Cournot s model of oligopoly

Principled Negotiation and the Negotiator s Dilemma is the Getting to Yes -approach too soft?

Question 2: How do you solve a matrix equation using the matrix inverse?

When Machine Learning Meets AI and Game Theory

1. Write the number of the left-hand item next to the item on the right that corresponds to it.

Continued Fractions and the Euclidean Algorithm

DATA ANALYSIS II. Matrix Algorithms

A Game Theoretical Framework for Adversarial Learning

Chapter 9 Experience rating

CPC/CPA Hybrid Bidding in a Second Price Auction

3.2 Roulette and Markov Chains

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Stochastic dynamics of invasion and fixation

Oligopoly: How do firms behave when there are only a few competitors? These firms produce all or most of their industry s output.

Bidding Dynamics of Rational Advertisers in Sponsored Search Auctions on the Web

A Mathematical Study of Purchasing Airline Tickets

Compact Representations and Approximations for Compuation in Games

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

A simple analysis of the TV game WHO WANTS TO BE A MILLIONAIRE? R

Bonus Maths 2: Variable Bet Sizing in the Simplest Possible Game of Poker (JB)

If, under a given assumption, the of a particular observed is extremely. , we conclude that the is probably not

A Game Playing System for Use in Computer Science Education

How To Understand And Solve A Linear Programming Problem

Oligopoly. Oligopoly is a market structure in which the number of sellers is small.

Introduction. Bargaining - whether over arms control, the terms of a peace settlement, exchange rate

Linear-Quadratic Optimal Controller 10.3 Optimal Linear Control Systems

1 Portfolio Selection

Equilibrium computation: Part 1

6.1 What is a Game? 156 CHAPTER 6. GAMES

Minimax Strategies. Minimax Strategies. Zero Sum Games. Why Zero Sum Games? An Example. An Example

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Econ 430 Lecture 9: Games on Networks

Local stability under evolutionary game dynamics

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Linear Programming. March 14, 2014

Common Knowledge: Formalizing the Social Applications

Solving Simultaneous Equations and Matrices

Game Theory: Supermodular Games 1

Games and Strategic Behavior. Chapter 9. Learning Objectives

6.042/18.062J Mathematics for Computer Science. Expected Value I

Section 1.3 P 1 = 1 2. = P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., =

Congestion Games with Player-Specific Payoff Functions

9.916 Altruism and Cooperation

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

Answer: The relationship cannot be determined.

SOLVING LINEAR SYSTEMS

1 Introduction to Matrices

Section 8.2 Solving a System of Equations Using Matrices (Guassian Elimination)

Dynamics and Equilibria

6.3 Conditional Probability and Independence

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Oligopoly and Strategic Pricing

Market Structure: Duopoly and Oligopoly

Do not open this exam until told to do so.

0.0.2 Pareto Efficiency (Sec. 4, Ch. 1 of text)

7 Gaussian Elimination and LU Factorization

Algebra 1 Course Title

F Matrix Calculus F 1

Markov Chains for the RISK Board Game Revisited. Introduction. The Markov Chain. Jason A. Osborne North Carolina State University Raleigh, NC 27695

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

Introduction to Matrix Algebra

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison

Data Mining: Algorithms and Applications Matrix Math Review

Power Rankings: Math for March Madness

A Dutch Book for Group Decision-Making?

MATH 340: MATRIX GAMES AND POKER

Extreme cases. In between cases

Intoduction to the Fifth Workshop Game-Theoretic Probability and Related Topics

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

Evolution leads to Kantian morality

Higher Education Math Placement

Transcription:

Evolutionary Game Theory James Holland Jones Department of Anthropology Stanford University December 1, 2008 1 The Evolutionarily Stable Strategy (ESS) An evolutionarily stable strategy (ESS) is a strategy that, if adopted by nearly all members of a population, can not be invaded by a rare mutant strategy. Consider a game in which D is an ESS. If a rare mutant H with very low frequency (p 1) were to enter the population, would it invade? We can write the fitness of the two strategies as follows: W (D) = C + (1 p)ie(d, D, ) + pie(d, H) W (H) = C + (1 p)ie(h, D) + pie(h, H) where C is the fitness of an individual before the encounter. If D is an ESS, this means that W (D) > W (H) Solving the inequality W (D) > W (H), we get: IE(D, D) IE(H, D) IE(H, H) IE(D, H) > p 1 p We assumed that p 1 so the fraction on the right-hand side of the inequality is essentially equal to zero. Thus for the left-hand side to be greater than zero means that the numerator on the right-hand side must be positive. This yields the condition: IE(D, D) > IE(H, D). This is the Nash Equilibrium solution. However, an ESS can arise even when this condition is not met. If IE(D, D) = IE(H, D) and IE(D, H) > IE(H, D), then D is still an ESS. That is, while H mutants are not at a disadvantage when playing D s, D nonetheless shares an advantage in the population because it does better against H than H does against itself. The Hawk-Dove game was originally developed by Maynard Smith and Price to describe animal conflict. Animals playing the Dove strategy (i.e., Doves ) avoid conflict. When faced Correspondence Address: Department of Anthropology, 450 Serra Mall, Building 50, Stanford, CA 94305-2034; phone: 650-723-4824, fax: 650-725-0605; email: jhj1@stanford.edu 1

with competition over a resource, two doves will split the benefit of the resource v. In contrast, Hawks always escalate conflicts and fight over resources. When a Hawk plays a dove, it always takes the full value of the resource, however, when a Hawk plays another Hawk, there is a 50% chance either will win. These fights between Hawks are costly, incurring a cost c on Hawks during each encounter. Figure 1 presents the payoff matrix for the Hawk-Dove game. Each cell of the matrix contains to entries. The first is the payoff to the row player. The second is the payoff to the column player. Frequently, payoff matrices will simply list the payoff to the row player, since if the game is symmetric, the payoffs to the row players and column players will be identical when facing the same type of opponent. H D H (v c)/2, (v c)/2 v, 0 D 0, v v/2, v/2 Figure 1: Maynard Smith and Price s Hawk-Dove game. Row player s payoffs given first in each cell. 2 Evolutionary Game Theory Evolutionary game theory (EGT) focuses on evolutionary dynamics that are frequency dependent. The fitness payoff for a particular phenotype depends on the population composition. Classical game theory focuses largely on the properties of the equilibria of games. One of the central defining features of EGT is the focus on the dynamics of strategies and their composition in a population rather than on the properties of equilibria. The approach is to study replicator equations. The attractors of these dynamical equations are the evolutionary stable strategies (ESSs) or the Nash Equilibria of the game. Consider the Prisoner s Dilemma game. This game is defined by two strategies: Cooperate and Defect. The prisoner s dilemma game is defined such that the payoff to a defector playing against a cooperator is greatest; the payoff of mutual cooperation is next greatest; the payoff for mutual defection is next greatest; and the payoff to cooperating when the opponent defects is least. These features can be economically summarized in a payoff matrix that lists the payoff to the row strategy when playing the column strategy (figure 2). We can name the payoffs in game 2 T > R > P > S, where T is the temptation, R is the reward for mutual cooperation, P is the punishment for mutual defection, and S is the sucker s payoff. It is conventional to also assume that R > (T + P )/2. Without this, then alternating rounds between cooperation and defection leads to a greater payoff than pure cooperation. C D C R S D T P Figure 2: Prisoner s Dilemma. T > R > P > S and R > (T + P )/2. 2

We assume a large population (effectively infinite) so that we can reasonably model the proportions of the two strategies, x c and x d. If the population were sufficiently small, we would need to worry about things like the random extinction of strategies at low frequency (this is the so-called finite population model). Let w c and w d be the fitnesses of cooperators and defectors respectively. Further, let w(s 1, s 2 ) denote the change in fitness for an s 1 individual against an s 2 opponent. Fitness is a function of the proportions of strategies d and c: w c = W c,0 + x c w(c, C) + x d w(c, D) (1) w d = W d,0 + x c w(d, C) + x d w(d, D) w = x c w c + x d w d where W i,0 is the initial fitness of strategy i and w is mean fitness. The change in frequencies from one generation to the next will be a function of the current proportions and the relative fitnesses. For example, the proportion of strategy c in the next generation will be x c = x c w c / w, the proportion of strategy d will be x d = x dw d / w. The change from one generation to the next for the c strategy will be: x c = x c x c = x c(w c w) w For small change, and where reproduction is continuous, we can approximate this equation with the differential equation, ẋ c = x c(w c w). w Since by definition of the Prisoner s Dilemma (w c w)/ w < 0, it is clear that the c strategy will decay exponentially at a rate equal to (w c w)/ w. The general replicator dynamic equation for n strategies is ẋ i = x i [w i (x) w] (2) where ẋ i is the time derivative of the proportion playing the ith strategy, w i (x) is the fitness of strategy i when played against the current state vector x = (x 1, x 2,..., x n ). This fitness of strategy i comes from the payoff matrix W = [w ij ]. Fitness for the i strategy is w i (x) = j x iw i,j. Note that this is just the expected fitness for strategy i: multiply the frequency of strategy i times the fitnesses of i against itself and all other strategies in the population. This sum is its expected fitness. The system of differential equations in 2 is generally much easier to analyze than the corresponding system of difference equations. For continuously breeding, age-structured populations, the assumption of continuous change is not that bad as long as the change per unit time is not too great. Assuming that the population size is very large means that we don t have to worry about stochastic variation in strategies due to drift. 3

2.1 An Example: Hawk-Dove-Retaliator Hawk-Dove-Retaliator is another game described by Maynard Smith and Price. Like Hawk- Dove, Hawks always fight and Doves always concede to a fight. The game introduces a third strategy, Retaliator, which escalate only if its adversary escalates. That is, Retaliator normally plays Dove but plays Hawk against other Hawks. The payoff to Retaliator is thus equal to that of a Hawk when playing a Hawk and a Dove when playing Doves or itself. These payoffs are summarized in figure 3. H D R H (v c)/2 v (v c)/2 D 0 v/2 v/2 R (v c)/2 v/2 v/2 Figure 3: Maynard Smith and Price s Hawk-Dove-Retaliator game. Row player s payoffs given first in each cell. v is the payoff for victory in a conflict. c is the cost of fighting (e.g., wounding). The trajectories of the dynamics of this game are plotted in the three-sided simplex of figure 4. In this plot, each vertex of the triangle represents the pure strategy. A point in the interior represents a mixture of strategies. The sum of the distances from each vertex is always one. 4

Retaliator Dove Hawk Figure 4: Evolutionary trajectories for the Hawk-Dove-Retaliator game with different starting values. The initial values are x (1) 0 = (1/3, 1/3, 1/3) and x (2) 0 = (1/3, 1/2, 1/6). Initial values x (1) 0 leads to a mixed equilibrium that is 2/3 Doves and 1/3 Hawks, while x (2) 0 leads to a mixed equilibrium that is 69% Retaliator and 31% Dove. 5