6.254 : Game Theory with Engineering Applications Lecture 6: Continuous and Discontinuous Games

Similar documents
6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

2.3 Convex Constrained Optimization Problems

Duality of linear conic problems

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

1 if 1 x 0 1 if 0 x 1

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

Date: April 12, Contents

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

6.254 : Game Theory with Engineering Applications Lecture 1: Introduction

A FIRST COURSE IN OPTIMIZATION THEORY

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

Metric Spaces. Chapter Metrics

Follow links for Class Use and other Permissions. For more information send to:

Mathematical finance and linear programming (optimization)

Convex analysis and profit/cost/support functions

Practice with Proofs

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

Separation Properties for Locally Convex Cones

Nonlinear Programming Methods.S2 Quadratic Programming

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

Definition and Properties of the Production Function: Lecture

1. Prove that the empty set is a subset of every set.

Continued Fractions and the Euclidean Algorithm

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

Game Theory: Supermodular Games 1

Fuzzy Differential Systems and the New Concept of Stability

ALMOST COMMON PRIORS 1. INTRODUCTION

Fixed Point Theorems

Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization

Econ 430 Lecture 9: Games on Networks

10.2 Series and Convergence

x a x 2 (1 + x 2 ) n.

Mathematical Methods of Engineering Analysis

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

IMPLEMENTING ARROW-DEBREU EQUILIBRIA BY TRADING INFINITELY-LIVED SECURITIES

1 Norms and Vector Spaces

MA651 Topology. Lecture 6. Separation Axioms.

10 Evolutionarily Stable Strategies

Lecture Notes 10

ECON Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted.

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

Chapter 7. Sealed-bid Auctions

24. The Branch and Bound Method

On the Existence of Nash Equilibrium in General Imperfectly Competitive Insurance Markets with Asymmetric Information

Critical points of once continuously differentiable functions are important because they are the only points that can be local maxima or minima.

(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties

Introduction to Topology

Notes on metric spaces

BANACH AND HILBERT SPACE REVIEW

THE BANACH CONTRACTION PRINCIPLE. Contents

Advanced Microeconomics

Class Meeting # 1: Introduction to PDEs

Lecture 7: Finding Lyapunov Functions 1

Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.

CPC/CPA Hybrid Bidding in a Second Price Auction

4 Lyapunov Stability Theory

Numerical methods for American options

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Sensitivity analysis of utility based prices and risk-tolerance wealth processes

God created the integers and the rest is the work of man. (Leopold Kronecker, in an after-dinner speech at a conference, Berlin, 1886)

0 <β 1 let u(x) u(y) kuk u := sup u(x) and [u] β := sup

Unraveling versus Unraveling: A Memo on Competitive Equilibriums and Trade in Insurance Markets

Adaptive Online Gradient Descent

Systems of Linear Equations

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

15 Limit sets. Lyapunov functions

What is Linear Programming?

Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

Scalar Valued Functions of Several Variables; the Gradient Vector

Inner Product Spaces

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

CHAPTER 9. Integer Programming

Quasi-static evolution and congested transport

Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products

TOPIC 4: DERIVATIVES

When is Reputation Bad? 1

SOLUTIONS TO EXERCISES FOR. MATHEMATICS 205A Part 3. Spaces with special properties

4.6 Linear Programming duality

CONTINUED FRACTIONS AND PELL S EQUATION. Contents 1. Continued Fractions 1 2. Solution to Pell s Equation 9 References 12

Imprecise probabilities, bets and functional analytic methods in Łukasiewicz logic.

Metric Spaces Joseph Muscat 2003 (Last revised May 2009)

Math 4310 Handout - Quotient Vector Spaces

Working Paper Series

Metric Spaces. Chapter 1

Finite dimensional topological vector spaces

Mathematics for Econometrics, Fourth Edition

4.5 Linear Dependence and Linear Independence

E3: PROBABILITY AND STATISTICS lecture notes

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

An Example of a Repeated Partnership Game with Discounting and with Uniformly Inefficient Equilibria

Tiers, Preference Similarity, and the Limits on Stable Partners

LS.6 Solution Matrices

Solving Systems of Linear Equations

Bargaining Solutions in a Social Network

Decentralization and Private Information with Mutual Organizations

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Transcription:

6.254 : Game Theory with Engineering Applications Lecture 6: Continuous and Discontinuous Games Asu Ozdaglar MIT February 23, 2010 1

Introduction Outline Existence of a Mixed Nash Equilibrium in (Glicksberg s Theorem) Existence of a Mixed Nash Equilibrium with Discontinuous Payoffs Construction of a Mixed Nash Equilibrium with Infinite Strategy Sets Uniqueness of a Pure Nash Equilibrium for Reading: Myerson, Chapter 3. Fudenberg and Tirole, Sections 12.2, 12.3. Rosen J.B., Existence and uniqueness of equilibrium points for concave N-person games, Econometrica, vol. 33, no. 3, 1965. 2

Game Theory: Lecture 6 We consider games in which players may have infinitely many pure strategies. Definition A continuous game is a game I, (S i ), (u i ) where I is a finite set, the S i are nonempty compact metric spaces, and the u i : S R are continuous functions. We next state the analogue of Nash s Theorem for continuous games. 3

Existence of a Mixed Nash Equilibrium Theorem (Glicksberg) Every continuous game has a mixed strategy Nash equilibrium. With continuous strategy spaces, space of mixed strategies infinite dimensional, therefore we need a more powerful fixed point theorem than the version of Kakutani we have used before. Here we adopt an alternative approach to prove Glicksberg s Theorem, which can be summarized as follows: We approximate the original game with a sequence of finite games, which correspond to successively finer discretization of the original game. We use Nash s Theorem to produce an equilibrium for each approximation. We use the weak topology and the continuity assumptions to show that these converge to an equilibrium of the original game. 4

Closeness of Two Games Let u = (u 1,..., u I ) and ũ = ( ũ 1,..., ũ I ) be two profiles of utility functions defined on S such that for each i I, the functions u i : S R and ũ i : S R are bounded (measurable) functions. We define the distance between the utility function profiles u and ũ as max sup u i (s) ũ i (s). i I s S Consider two strategic form games defined by two profiles of utility functions: G = I, (S i ), (u i ), G = I, (S i ), (ũ i ). If σ is a mixed strategy Nash equilibrium of G, then σ need not be a mixed strategy Nash equilibrium of G. Even if u and ũ are very close, the equilibria of G and G may be far apart. For example, assume there is only one player, S 1 = [0, 1], u 1 (s 1 ) = ɛs 1, and ũ 1 (s 1 ) = ɛs 1, where ɛ > 0 is a sufficiently small scalar. The unique equilibrium of G is s 1 = 1, and the unique equilibrium of G is s 1 = 0, even if the distance between u and ũ is only 2ɛ. 5

Closeness of Two Games and ɛ-equilibrium However, if u and ũ are very close, there is a sense in which the equilibria of G are almost equilibria of G. Definition (ɛ-equilibrium) Given ɛ 0, a mixed strategy σ Σ is called an ɛ-equilibrium if for all i I and s i S i, u i (s i, σ i ) u i (σ i, σ i ) + ɛ. Obviously, when ɛ = 0, an ɛ-equilibrium is a Nash equilibrium in the usual sense. 6

Continuity Property of ɛ-equilibria Proposition (1) Let G be a continuous game. Assume that σ k σ, ɛ k ɛ, and for each k, σ k is an ɛ k -equilibrium of G. Then σ is an ɛ-equilibrium of G. Proof: For all i I, and all s i S i, we have u i (s i, σ k ) u i (σ k ) + ɛ k, i Taking the limit as k in the preceding relation, and using the continuity of the utility functions (together with the convergence of probability distributions under weak topology), we obtain, establishing the result. u i (s i, σ i ) u i (σ) + ɛ, 7

Closeness of Two Games We next define formally the closeness of two strategic form games. Definition Let G and G be two strategic form games with G = I, (S i ), (u i ), G = I, (S i ), (u i ). Then G is an α approximation to G if for all i I and s S, we have u i (s) u i (s) α. 8

ɛ equilibria of Close Games The next proposition relates the ɛ equilibria of close games. Proposition (2) If G is an α-approximation to G and σ is an ɛ-equilibrium of G, then σ is an (ɛ + 2α)-equilibrium of G. Proof: For all i I and all s i S i, we have u i (s i, σ i ) u i (σ) = u i (s i, σ i ) u i (s i, σ i ) + u i (s i, σ i ) u i (σ) +u i (σ) u i (σ) α + ɛ + α = ɛ + 2α. 9

Approximating a Continuous Game with an Essentially Finite Game The next proposition shows that we can approximate a continuous game with an essentially finite game to an arbitrary degree of accuracy. Proposition (3) For any continuous game G and any α > 0, there exists an essentially finite game which is an α-approximation to G. 10

Proof Since S is a compact metric space, the utility functions u i are uniformly continuous, i.e., for all α > 0, there exists some ɛ > 0 such that u i (s) u i (t) α for all d(s, t) ɛ. Since S i is a compact metric space, it can be covered with finitely many open balls U i j, each with radius less than ɛ (assume without loss of generality that these balls are disjoint and nonempty). Choose an s i j U i j for each i, j. Define the essentially finite game G with the utility functions u i defined as I u i (s) = u i (s 1 j,..., si j ), s U j = U k j. k=1 Then for all s S and all i I, we have u i (s) u i (s) α, since d(s, s j ) ɛ for all j, implying the desired result. 11

Proof of Glicksberg s Theorem We now return to the proof of Glicksberg s Theorem. Let {α k } be a scalar sequence with α k 0. For each α k, there exists an essentially finite α k -approximation G k of G by Proposition 3. Since G k is essentially finite for each k, it follows using Nash s Theorem that it has a 0-equilibrium, which we denote by σ k. Then, by Proposition 2, σ k is a 2α k -equilibrium of G. Since Σ is compact, {σ k } has a convergent subsequence. Without loss of generality, we assume that σ k σ. Since 2α k 0, σ k σ, by Proposition 1, it follows that σ is a 0-equilibrium of G. 12

Discontinuous Games Game Theory: Lecture 6 There are many games in which the utility functions are not continuous (e.g. price competition models, congestion-competition models). We next show that for discontinuous games, under some mild semicontinuity conditions on the utility functions, it is possible to establish the existence of a mixed Nash equilibrium (see [Dasgupta and Maskin 86]). The key assumption is to allow discontinuities in the utility function to occur only on a subset of measure zero, in which a player s strategy is related to another player s strategy. To formalize this notion, we introduce the following set: for any two players i and j, let D be a finite index set and for d D, let f ij d : S i S j be a bijective and continuous function. Then, for each i, we define S (i) = {s S j = i such that s j = f ij d (s i ).} (1) 13

Discontinuous Games Game Theory: Lecture 6 Before stating the theorem, we first introduce some weak continuity conditions. Definition Let X be a subset of R n, X i be a subset of R, and X i be a subset of R n 1. (i) A function f : X R is called upper semicontinuous (respectively, lower semicontinuous) at a vector x X if f (x) lim sup k f (x k ) (respectively, f (x) lim inf k f (x k )) for every sequence {x k } X that converges to x. If f is upper semicontinuous (lower semicontinuous) at every x X, we say that f is upper semicontinuous (lower semicontinuous). (ii) A function f : X i X i R is called weakly lower semicontinuous in x i over a subset X i X i, if for all x i there exists λ [0, 1] such that, for all x i X i, λ lim inf f (x i, x i ) + (1 λ) lim inf f (x i, x i ) f (x i, x i ). x x i x x i i i 14

Discontinuous Games Theorem (2) [Dasgupta and Maskin] Let S i be a closed interval of R. Assume that u i is continuous except on a subset S (i) of the set S (i) defined in Eq. (1). Assume also that n i=1 u i (s) is upper semicontinuous and that u i (s i, s i ) is bounded and weakly lower semicontinuous in s i over the set {s i S i (s i, s i ) S (i)}. Then the game has a mixed strategy Nash equilibrium. The weakly lower semicontinuity condition on the utility functions implies that the function u i does not jump up when approaching s i either from below or above. Loosely, this ensures that player i can do almost as well with strategies near s i as with s i, even if his opponents put weight on the discontinuity points of u i. 15

Bertrand Competition with Capacity Constraints Consider two firms that charge prices p 1, p 2 [0, 1] per unit of the same good. Assume that there is unit demand and all customers choose the firm with the lower price. If both firms charge the same price, each firm gets half the demand. All demand has to be supplied. The payoff functions of each firm is the profit they make (we assume for simplicity that cost of supplying the good is equal to 0 for both firms). 16

Bertrand Competition with Capacity Constraints We have shown before that (p 1, p 2 ) = (0, 0) is the unique pure strategy Nash equilibrium. Assume now that each firm has a capacity constraint of 2/3 units of demand: Since all demand has to be supplied, this implies that when p 1 < p 2, firm 2 gets 1/3 units of demand). It can be seen in this case that the strategy profile (p 1, p 2 ) = (0, 0) is no longer a pure strategy Nash equilibrium: Either firm can increase his price and still have 1/3 units of demand due to the capacity constraint on the other firm, thus making positive profits. It can be established using Theorem 2 that there exists a mixed strategy Nash equilibrium. Let us next proceed to construct a mixed strategy Nash equilibrium. 17

Bertrand Competition with Capacity Constraints We focus on symmetric Nash equilibria, i.e., both firms use the same mixed strategy. We use the cumulative distribution function F ( ) to represent the mixed strategy used by either firm. It can be seen that the expected payoff of player 1, when he chooses p 1 and firm 2 uses the mixed strategy F ( ), is given by u 1 (p 1, F ( )) = F (p 1 ) p 1 + (1 F (p 1 )) 2 p 1. 3 3 Using the fact that each action in the support of a mixed strategy must yield the same payoff to a player at the equilibrium, we obtain for all p in the support of F ( ), F (p) p + 2 p = k, 3 3 for some k 0. From this we obtain: 3k F (p) = 2. p 18

Bertrand Competition with Capacity Constraints Note next that the upper support of the mixed strategy must be at p = 1, which implies that F (1) = 1. Combining with the preceding, we obtain 0, if 0 p 1 2, F (p) = 2 1 p, if 1 2 p 1, 1, if p 1. 19

Uniqueness of a Pure Strategy Nash Equilibrium in We have shown in the previous lecture the following result: Given a strategic form game I, (S i ), (u i ), assume that the strategy sets S i are nonempty, convex, and compact sets, u i (s) is continuous in s, and u i (s i, s i ) is quasiconcave in s i. Then the game I, (S i ), (u i ) has a pure strategy Nash equilibrium. The next example shows that even under strict convexity assumptions, there may be infinitely many pure strategy Nash equilibria. 20

Uniqueness of a Pure Strategy Nash Equilibrium Example Consider a game with 2 players, S i = [0, 1] for i = 1, 2, and the payoffs 2 u 1 (s 1, s 2 ) = s 1 s 2 s 1, 2 2 2 u 2 (s 1, s 2 ) = s 1 s 2 s. 2 Note that u i (s 1, s 2 ) is strictly concave in s i. It can be seen in this example that the best response correspondences (which are unique-valued) are given by B 1 (s 2 ) = s 2, B 2 (s 1 ) = s 1. Plotting the best response curves shows that any pure strategy profile (s 1, s 2 ) = (x, x) for x [0, 1] is a pure strategy Nash equilibrium. 21

Uniqueness of a Pure Strategy Nash Equilibrium We will next establish conditions that guarantee that a strategic form game has a unique pure strategy Nash equilibrium, following the classical paper [Rosen 65]. Notation: Given a scalar-valued function f : R n R, we use the notation f (x) to denote the gradient vector of f at point x, i.e., [ ] f (x) f (x) T f (x) =,...,. x 1 x n Given a scalar-valued function u : I i=1 R m i R, we use the notation i u(x) to denote the gradient vector of u with respect to x i at point x, i.e., [ ] T u( x) u( x) i u(x) = x 1,..., m. (2) i x i i 22

Optimality Conditions for Nonlinear Optimization Problems Theorem (3) (Karush-Kuhn-Tucker conditions) Let x be an optimal solution of the optimization problem maximize f (x) subject to g j (x) 0, j = 1,..., r, where the cost function f : R n R and the constraint functions g j : R n R are continuously differentiable. Denote the set of active constraints at x as A(x ) = {j = 1,..., r g j (x ) = 0}. Assume that the active constraint gradients, g j (x ), j A(x ), are linearly independent vectors. Then, there exists a nonnegative vector λ R r (Lagrange multiplier vector) such that r f (x ) + λ j g j (x ) = 0, j=1 λ j g j (x ) = 0, j = 1,..., r. (3) 23

Optimality Conditions for Nonlinear Optimization Problems For convex optimization problems (i.e., minimizing a convex function (or maximizing a concave function) over a convex constraint set), we can provide necessary and sufficient conditions for the optimality of a feasible solution: Theorem (4) Consider the optimization problem maximize f (x) subject to g j (x) 0, j = 1,..., r, where the cost function f : R n R and the constraint functions g j : R n R are concave functions. Assume also that there exists some x such that g j ( x) > 0 for all j = 1,..., r. Then a vector x R n is an optimal solution of the preceding problem if and only if g j (x ) 0 for all j = 1,..., r, and there exists a nonnegative vector λ R r (Lagrange multiplier vector) such that r f (x ) + λ j g j (x ) = 0, j=1 λ j g j (x ) = 0, j = 1,..., r. 24

Uniqueness of a Pure Strategy Nash Equilibrium We now return to the analysis of the uniqueness of a pure strategy equilibrium in strategic form games. We assume that for player i I, the strategy set S i is given by where h i : R m i R is a concave function. S i = {x i R m i h i (x i ) 0}, (4) Since h i is concave, it follows that the set S i is a convex set (exercise!). Therefore the set of strategy profiles S = I i=1 S i i I =1 R m i, being a Cartesian product of convex sets, is a convex set. Given these strategy sets, a vector x I =1 Rm i is a pure strategy Nash i equilibrium if and only if for all i I, x i is an optimal solution of maximize yi R m i subject to We use the notation u(x) to denote u i (y i, x i ) h i (y i ) 0. (5) u(x) = [ 1 u 1 (x),..., I u I (x)] T. (6) 25

Uniqueness of a Pure Strategy Nash Equilibrium We introduce the key condition for uniqueness of a pure strategy Nash equilibrium. Definition We say that the payoff functions (u 1,..., u I ) are diagonally strictly concave for x S, if for every x, x S, we have (x x ) T u(x ) + (x x ) T u(x ) > 0. Theorem Consider a strategic form game I, (S i ), (u i ). For all i I, assume that the strategy sets S i are given by Eq. (4), where h i is a concave function, and there exists some x i R m i such that h i (x i ) > 0. Assume also that the payoff functions (u 1,..., u I ) are diagonally strictly concave for x S. Then the game has a unique pure strategy Nash equilibrium. 26

Proof Game Theory: Lecture 6 Assume that there are two distinct pure strategy Nash equilibria. Since for each i I, both x i and x i must be an optimal solution for an optimization problem of the form (5), Theorem 4 implies the existence of nonnegative vectors λ = [λ 1,..., λ I ] T and λ = [ λ 1,..., λ I ] T such that for all i I, we have i u i (x ) + λ i h i (x i ) = 0, (7) and λ i h i (x i ) = 0, (8) i u i (x ) + λ i h i (x i ) = 0, (9) λ i h i (x i ) = 0. (10) 27

Proof Game Theory: Lecture 6 Multiplying Eqs. (7) and (9) by ( x i x i ) T and (x i x i ) T adding over all i I, we obtain respectively, and 0 = (x x ) T u(x ) + (x x) T u( x) (11) + x i xi T λ ) T λi hi ( x i ) ( xi i h i (x i ( ) + x i ) i I > x x x i ) T λ ) T λ i h i (x i ( ) + h ( (x i x i ), i I i where to get the strict inequality, we used the assumption that the payoff functions are diagonally strictly concave for x S. Since the h are concave functions, we have i i i I i I h i (x i ) + h i (x i ) T ( x i x i ) h i (x i ). i i 28

Proof Game Theory: Lecture 6 Using the preceding together with λ i > 0, we obtain for all i, λ i h i (x i ) T ( x i x i ) λ i (h i ( x i ) h i (x i )) = λ i h i ( x i ) 0, where to get the equality we used Eq. (8), and to get the last inequality, we used the facts λ i > 0 and h i ( x i ) 0. Similarly, we have λ i h i (x i ) T (x i x i ) 0. Combining the preceding two relations with the relation in (11) yields a contradiction, thus concluding the proof. 29

Sufficient Condition for Diagonal Strict Concavity Let U(x) denote the Jacobian of u(x) [see Eq. (6)]. In particular, if the x i are all 1-dimensional, then U(x) is given by 2 u 1 (x) 2 u 1 (x) x2 1 x 1 x 2 2 u 2 (x) U(x) =.... x 2 x 1. Proposition For all i I, assume that the strategy sets S i are given by Eq. (4), where h i is a concave function. Assume that the symmetric matrix (U(x) + U T (x)) is negative definite for all x S, i.e., for all x S, we have y T (U(x) + U T (x))y < 0, y = 0. Then, the payoff functions (u 1,..., u I ) are diagonally strictly concave for x S. 30

Proof Game Theory: Lecture 6 Let x, x S. Consider the vector x(λ) = λx + (1 λ) x, for some λ [0, 1]. Since S is a convex set, x(λ) S. Because U(x) is the Jacobian of u(x), we have d dx(λ) dλ u(x(λ)) = U(x(λ)) d(λ) = U(x(λ))(x x), or 1 U(x(λ))(x x)dλ = u(x ) u( x). 0 31

Proof Game Theory: Lecture 6 Multiplying the preceding by ( x x ) T yields (x x ) T u(x ) + (x x) T u( x) 1 1 = (x x) T [U(x(λ)) + U T (x(λ))](x x)dλ 2 0 > 0, where to get the strict inequality we used the assumption that the symmetric matrix (U(x) + U T (x)) is negative definite for all x S. 32

MIT OpenCourseWare http://ocw.mit.edu 6.254 Game Theory with Engineering Applications Spring 2010 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.