Théorie de la décision et théorie des jeux Stefano Moretti



Similar documents
6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

Full and Complete Binary Trees

Game Theory and Algorithms Lecture 10: Extensive Games: Critiques and Extensions

0.0.2 Pareto Efficiency (Sec. 4, Ch. 1 of text)

Backward Induction and Subgame Perfection

ECON Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted.

Lecture V: Mixed Strategies

Mathematical Induction

Competition and Regulation. Lecture 2: Background on imperfect competition

How to Solve Strategic Games? Dominant Strategies

10 Evolutionarily Stable Strategies

UCLA. Department of Economics Ph. D. Preliminary Exam Micro-Economic Theory

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

Game Theory in Wireless Networks: A Tutorial

21. Unverifiable Investment, Hold Up, Options and Ownership

Game Theory 1. Introduction

Biform Games: Additional Online Material

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

Week 7 - Game Theory and Industrial Organisation

ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY

AN INTRODUCTION TO GAME THEORY

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents

Sequential lmove Games. Using Backward Induction (Rollback) to Find Equilibrium

6.254 : Game Theory with Engineering Applications Lecture 1: Introduction

1 Representation of Games. Kerschbamer: Commitment and Information in Games

Equilibrium computation: Part 1

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

Solutions for Practice problems on proofs

Playing with Numbers

Microeconomic Theory Jamison / Kohlberg / Avery Problem Set 4 Solutions Spring (a) LEFT CENTER RIGHT TOP 8, 5 0, 0 6, 3 BOTTOM 0, 0 7, 6 6, 3

6.042/18.062J Mathematics for Computer Science. Expected Value I

Simon Fraser University Spring Econ 302 D200 Final Exam Solution Instructor: Songzi Du Tuesday April 21, 2015, 12 3 PM

Computational Learning Theory Spring Semester, 2003/4. Lecture 1: March 2

Game Theory: Supermodular Games 1

Network Flow I. Lecture Overview The Network Flow Problem

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Price competition with homogenous products: The Bertrand duopoly model [Simultaneous move price setting duopoly]

MATHEMATICAL INDUCTION. Mathematical Induction. This is a powerful method to prove properties of positive integers.

Oligopoly and Strategic Pricing

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

Connectivity and cuts

CPC/CPA Hybrid Bidding in a Second Price Auction

Individual security and network design

Chapter 7. Sealed-bid Auctions

Session 7 Fractions and Decimals

(Refer Slide Time: 2:03)

1 if 1 x 0 1 if 0 x 1

A Graph-Theoretic Network Security Game

Operations with positive and negative numbers - see first chapter below. Rules related to working with fractions - see second chapter below

Lecture 5 - CPA security, Pseudorandom functions

Infinitely Repeated Games with Discounting Ù

Limits and Continuity

Metric Spaces. Chapter 1

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

I. Noncooperative Oligopoly

Christmas Gift Exchange Games

Economics 159: Introduction to Game Theory. Summer 2015: Session A. Online Course

Math 4310 Handout - Quotient Vector Spaces

Dynamics and Equilibria

Minimax Strategies. Minimax Strategies. Zero Sum Games. Why Zero Sum Games? An Example. An Example

Oligopoly: Cournot/Bertrand/Stackelberg

MA4001 Engineering Mathematics 1 Lecture 10 Limits and Continuity

Question of the Day. Key Concepts. Vocabulary. Mathematical Ideas. QuestionofDay

i/io as compared to 4/10 for players 2 and 3. A "correct" way to play Certainly many interesting games are not solvable by this definition.

Decision Making under Uncertainty

Tree sums and maximal connected I-spaces

CS 103X: Discrete Structures Homework Assignment 3 Solutions

Games of Incomplete Information

Cournot s model of oligopoly

k, then n = p2α 1 1 pα k

Advanced Game Theory

x a x 2 (1 + x 2 ) n.

2.3 Convex Constrained Optimization Problems

Oligopoly markets: The price or quantity decisions by one rm has to directly in uence pro ts by other rms if rms are competing for customers.

ALMOST COMMON PRIORS 1. INTRODUCTION

6.3 Conditional Probability and Independence

Lecture Note on Auctions

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Adaptive Online Gradient Descent

2 Session Two - Complex Numbers and Vectors

Game Theory and Nash Equilibrium

The Basics of Game Theory

A Non-Linear Schema Theorem for Genetic Algorithms

FACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z

Mathematical Induction. Lecture 10-11

Duality of linear conic problems

Graph Theory Problems and Solutions

Game Theory. An introduction to the concepts of dominant strategies, Nash equilibrium and strategic commitment

Advanced Microeconomics

TOPIC 4: DERIVATIVES

Cosmological Arguments for the Existence of God S. Clarke

Lecture 28 Economics 181 International Trade

Answer: (a) Since we cannot repeat men on the committee, and the order we select them in does not matter, ( )

Transcription:

héorie de la décision et théorie des jeux Stefano Moretti UMR 7243 CNRS Laboratoire d'analyse et Modélisation de Systèmes pour l'aide à la décision (LAMSADE) Université Paris-Dauphine email: Stefano.MOREI@dauphine.fr

2 Key concepts Dominant strategies Nash equilibrium Mixed extension Strategic form vs. extensive form Information

Looking for a solution What a player will/should do? will : the descriptive point of view. Aiming at predicting what players will do in the model and hence in the real game should : the normative point of view. Rationality is based on a teleological description of the players. Players have an end (not like apples, stones, molecules). So, they could do the wrong thing. We could give them suggestions. Can we say something on the basis of our assumptions?

Domination among strategies From decision theory we borrow the idea of domination among strategies: x 1 is (obviously) better than x 2 if: h(x 1, y) h(x 2, y) for every yy We shall say that x 1 (strongly) dominates x 2. So, if x 1 dominates any other xx, then x 1 is the solution

Prisoner's dilemma he game is: I B II (3,3) (4,1) (1,4) (2,2) Obviously B and R are dominant strategies (for I and II respectively). So, we have the solution (B,R). Nice and easy. But... the outcome is inefficient! Both players prefer the outcome deriving from (, L). And so? he problem is that players are (assumed to be) rational and intelligent. L R

Strategies to avoid A strategy which is (strongly) dominated by another one will not be played. So we can delete it. But then could appear new (strongly) dominated strategies for the other player. We can delete them. And so on... Maybe players are left with just one strategy each. Well, a new way to get a solution for the game. echnically: solution via iterated elimination of dominated strategies.

Strategies to avoid: example I II L R I II L R (2,1) (1,0) (2,1) (1,0) M (1,1) (2,0) M (1,1) (2,0) B (0,0) (0,1) B (0,0) (0,1) I II L R I II L (2,1) (1,0) (2,1) M (1,1) (2,0) M (1,1) Solution: (, L)

Nash equilibrium Basic solution concept, for games in strategic form. (2 players only) Given G = (X, Y, f, g), (x*,y*)xy is a Nash equilibrium for G if: f(x*,y*) f(x,y*) for all xx g(x*,y*) g(x*,y) for all yy Interpretation: x* is a best reply (max utility f) for player I when player II plays strategy y*, and y* is a best reply (max utility g) for player II when player I plays strategy x*. Existence: Nash's theorem: mixed strategies... See later Difficulties: not uniqueness some Nash equilibria are not sensible inefficiency (Adam Smith was wrong?)

Nash equilibrium: examples (2) N.E. calculation in BoS Best reply for player I: max utility Fix this strategy for II I II L R (2,1) (0,0) B (0,0) (1,2) Best reply for player I: max utility I B Fix this strategy for II II L R (2,1) (0,0) (0,0) (1,2) Fix this strategy for I I B Best reply for player II: max utility II L (2,1) (0,0) R (0,0) (1,2) (0,0) Couples of strategies with both payoffs in red are N.E. Fix this strategy for I I B II L (2,1) Best reply for player II: max utility R (0,0) (1,2)

Nash equilibrium: examples (3) Example (battle of the sexes, BoS): (, L) and (B, R) are Nash Equilibria (N.E.). Not unique! I II L R B (2,1) (0,0) (0,0) (1,2)

Nash equilibrium and dominance heorem: If a game has a unique couple that survives iterated elimination of dominated strategies, that this couple is a Nash equilibrium. In particular, a couple of dominating strategies is a Nash equilibrium. So, in the prisoner's dilemma, (B, R) is the (unique) Nash equilibrium. So, Nash equilibrium can give an outcome which is inefficient.

Nash equilibrium: examples (4) Consider the following game (coordination game): I II L C R (0,0) (1,1) (0,0) M (0,0) (0,0) (1,1) B (1,1) (0,0) (0,0) (B, L), (, C) and (M,R) are N.E.

Nash equilibrium: not unique he battle of the sexes and the coordination game (and many others) have more than one NE. BIG ISSUE. players may have different (opposite) preferences on the equilibrium outcomes (see BoS) it is not possible to speak of equilibrium strategies. In the BoS, is an equilibrium strategy? Or B?

One more problem Example: matching pennies (MP) I II L R B (-1,1) (1,-1) (1,-1) (-1,1) here is no equilibrium? But Nash is famous (also) because of his existence thm (1950). But MP is a zero-sum game. So, even vn (1928) guarantees that it has an equilibrium. Where do we find it? Usual math trick: extend (N to Z, sum to integral, solution to weak solution).

Mixed strategies he basic idea is that the player does not choose a strategy, but a probability distribution on strategies. Example: I have an indivisible object and I must assign it in a fair way to one of my children. It is quite possible that the best solution is to decide to assign it randomly (with a uniform probability distribution).

Mixed extension of a game Let's apply it to games in strategic form. Given a game G = (X, Y, f, g), assume X, Y are finite, and let X ={x 1,, x m }, Y ={y 1,, y n }. he mixed extension of G is Γ=( (X), (Y), f, g ). Where: (X) ( (Y))is the set of probability distributions on X (Y). An element of (X) is p = (p 1,, p m )R m, where p i is the probability to play strategy x i and f (p, q) = i{1, m} j{1, n} p i q j f(x i,y j ) g (p, q) = i{1, m} j{1, n} p i q j g(x i,y j ) Of course, (p,q) (X) (Y) Notice that Γ is itself a game in strategic form. So, no need to redefine concepts (in particular, N.E.).

Interpretation? Of course, there is no mathematical problem in the definition of Γ. But: f and g can still be interpreted as payoffs for the players? he answer is YES if the original f and g are vnm utility functions. Otherwise, we cannot attach a meaning to the operations that brought us from G to Γ.

Mixed extension and equilibria for BoS q 1-q he BoS is: I II L R p (2,1) (0,0) 1-p B (0,0) (1,2) Instead of using ((p 1,p 2 ),(q 1,q 2 )) we use ((p,1-p),(q,1-q)), with p,q[0,1]. So: f ((p,1-p),(q,1-q))=2pq+1(1-p)(1-q)=(3q-1)p+(1-q) Given q, the best reply for player I to q is p* such that p*=0 if 0q<1/3 p*[0,1] if q=1/3 p*=1 if 1/3<q1

Mixed extension and equilibria for BoS (2) Given p, the best reply for player II to p* is q*=0 if 0p<2/3 q*[0,1] if p=2/3 q*=1 if 2/3<p1 From the following picture we see there are 3 N.E.

Nash existence theorem heorem: Given a game (X, Y, f, g), where X and Y are finite sets. hen, the mixed extension ( (X), (Y), f, g ) possesses a Nash equilibrium. A proof of this theorem can be given using the fixed point thorem of Kakutani (1941): heorem: Let X be a non-empty compact convex set of R k. Let f:xx be a set-valued function for which - for all xx the set f(x) is nonempty and convex; - the graph of f is closed (i.e. for all sequences {x n } and {y n } such that y n f(x n ) for all n, x n x, y n y, we have that yf(x)). hen there exists a x*x such that x*f(x*).

Nash existence theorem (2) Note that the mixed extension of a finite game satisfies the assumptions of the Kakutani s fixed point theorem: (X), (Y) are convex and compact Moreover we observed that a Nash eq. is a profile a* of actions in A i such that a* i i (a* -i ) for each player i N. Define the set valued function B:A A (where A= in i is the cartesian product of A i over the set of players) such that B(a)= in i (a* -i ). hen a fixed point a* B(a*) is a Nash eq. One can prove that by continuity of f' and g', the set valued function B has closed graph. hen by Kakutani's theorem it has a fixed point.

Matching pennies revisited Let us redefine the rules first you choose, L or R then I choose, or B Is it ok? Is it the same game? It depends. Essential is not the chronological (physical) time, but the information that I have when I must decide. So, can I see (know) your choice before deciding?

rees We have introduced two important aspects: the dynamic structure of the interaction the role of info on past events (on the history) o represent them it is appropriate a tree structure

Information sets (1) First chooses player I. hen, being informed of the choice made by player I, player II makes his choice. Player I Player II B Player II L R L B R B (-1,1) It means: utility of player I is -1 Utility of player II is 1 (1,-1) (1,-1) (-1,1) It means: utility of player I is 1 Utility of player II is -1

Strategy Consider the tree game on the previous slide Player I must begin choosing between or B. Player II must choose among: L L B, L R B, R L B, R R B... L L B means: chose L if player I has chosen at the first step and L B if player I has chosen B at the first step L R B means: chose L if player I has chosen at the first step and R B if player I has chosen B at the first step So, we have a game in strategic form. Payoffs? Follow the path! I II L L B L R B R L B B (-1,1) (1,-1) (-1,1) (-1,1) (1,-1) (1,-1) R R B (1,-1) (-1,1)

Information sets (2) A trick to take note of what player II does not know about the past (the history). wo nodes are connected with a dashed line. he meaning is that a player will know that has to make a choice knowing that: he knows he is in one of these nodes but he does not know in which. hese sets of nodes connected with dashed lines are called information sets. Player I Player II B Player II L R L R (-1,1) (1,-1) (1,-1) (-1,1)

Strategy Consider the tree game on the previous slide Player I must begin choosing between or B. Player II must choose between L and B. we have the following game in strategic form: I II L R B (-1,1) (1,-1) (1,-1) (-1,1)

Nothing new Every game in extensive form can be converted into a game in strategic form, using the (natural) idea of strategy we have seen. So, we can say that the strategic form is, somehow, fundamental, at least for noncooperative games. We can use the Nash equilibrium also as a solution for games in extensive form. It seems that everything is so easy

Backward induction Consider the very simple game depicted in the following slide Player I must begin choosing between a or b. But there is nothing that obliges him to think locally. He knows that he could be called to play again. So, before the game starts, he can decide his strategy. It means, choose among: agil, ahil, bgil, bhil. Similarly, II can choose among:cem, cfm, dem, dfm. So, we have a game in strategic form.

Player I Player II a b Player II Player I g c d e f Player I Player I h i l (-1,5) m Player II (3,2) (1,5) (2,4) (2,0)

Backward induction (2) Look at the penultimate" nodes. here the choice is easy, it is a single DM that has all of the power to enforce the outcome he prefers. having done this, look at the pre-penultimate nodes. aking into account the choices that will be made by the player who follows, the choice for the player at the prepenultimate node becomes obvious too. And so on, till we reach the root of the tree. he method we followed is called backward induction and works for games with perfect information (i.e.: information sets are all singletons). Well, we get a strategy profile that has good chances to be considered a solution! Actually, it can be proved that it is a Nash equilibrium.

Backward induction (3) Player I Player II a b Player II Player I c d e f g h (1,5) (2,4) (-1,5) (2,0) (3,2)

Backward induction (4) Player I Player II a b Player I c d (-1,5) (3,2) (1,5) (2,4) Player I a b (1,5) (-1,5) (1,5)

Summing up Player I Player II a b Player II Player I g c d e f Player I Player I h i l (-1,5) m Player II (3,2) (1,5) (2,4) Backward induction provides the strategy profile (a h i l; d f m) (2,0)

Game in strategic form I II c e m c f m d e m a g i l (2,0) (2,0) (1,5) a h i l (3,2) (3,2) (1,5) d f m (1,5) (1,5) b g i l (2,4) (-1,5) b h i l (2,4) (-1,5) (2,4) (2,4) (-1,5) (-1,5) (a h i l; d f m) is a Nash equilibrium of the game (actually there is already another Nash equilibrium, but with an equivalent outcome )

Let's see a game, very small: Player I Player II B L R (1,2) 2,1) (0,0) I II L R B (2,1) (1,2) (0,0) (1,2)

A small problem Player I Player II B L R 2,1) (0,0) I II (1,2) L R (2,1) (0,0) B (1,2) (1,2) Backward induction gives (; L). But the strategic form has two Nash equilibria: (; L) and (B; R)! And so?

Subgame Perfect Equilibrium (SPE) he key point is that the strategy profile determined by backward induction is more than a NE. It is a SPE. hat is, it is not only a NE, but it is also a NE for all of the subgames. What are subgames? For games with perfect information they coincide with subtrees. More interesting the general case, but the idea is obvious: we want a subtree that can be seen sensibly as a game (subgame of the given game). Games with perfect information: there is coincidence between SPE and strategy profiles found by backward induction. SPE can be defined for any game in extensive form. It is the first example of refinement of NE.