Planning, Learning, Prediction, and Games Two Player Zero Sum Games and von Neumann s Minimax Theorem: 12/18/09-01/15/10

Size: px
Start display at page:

Download "Planning, Learning, Prediction, and Games Two Player Zero Sum Games and von Neumann s Minimax Theorem: 12/18/09-01/15/10"

Transcription

1 Planning, Learning, Prediction, and Games Two Player Zero Sum Games and von Neumann s Minimax Theorem: 12/18/09-01/15/10 Lecturer: Patrick Briest, Peter Pietrzyk Scribe: Philipp Brandes, Peter Pietrzyk, Patrick Briest 4 Two Player Zero Sum Games and von Neumann s Minimax Theorem Definition 4.1 A normal form game is specified by I a set of players, A i a set of strategies for each player i I and u i : A i R a payoff function for each player i I. i I Normal form games with 2 players can be written in matrix form, such that the elements of A 1 index rows, the elements of A 2 index columns, and the entry in row r and column c is (u 1 (r, c), u 2 (r, c)). We call player 1 the row player and player 2 the column player. We can also think of the game as being described by two matrices R and C specifying the payoffs of the row and column players, respectively. Example 4.2 (Rock, Paper, Scissors) Rock beats Scissors, Scissors beats Paper, Paper beats Rock. Strategies: R, P, S Payoffs: 1 for winning, 1 for losing, 0 for draw. Payoff matrix: R P S R (0, 0) ( 1, 1) (1, 1) P (1, 1) (0, 0) ( 1, 1) S ( 1, 1) (1, 1) (0, 0) Example 4.3 (The Prisoners Dilemma) Two players are accused of a crime. If both admit, both go to jail for 2 years. If they both keep quiet, they will go to jail for only one year. If only one of them admits, he becomes a principal witness and goes free, while the other goes to jail for 3 years. 36

2 Define the payoff of x years in jail as (3 x). Strategies of both players: A admit, Q keep quiet Payoff matrix: Q A Q (2, 2) (0, 3) A (3, 0) (1, 1) Definition 4.4 For a set of strategies A, let { (A) = p : A 0, 1 } p(a) = 1 a A denote the set of all probability distributions on A. For some player i I, we call the elements of (A i ) her mixed strategies. Mixed strategies are rules for randomly picking a strategy. Picking a strategy deterministically is called a pure strategy. Elements of i I A i are called pure strategy profiles. Elements of i I (A i ) are mixed called mixed strategy profiles. The expected payoff of player i I given a mixed strategy profile (p 1, p 2,..., p I ) i I (A i ) is u i (p 1, p 2,..., p I ) = (a 1,a 2,...,a I ) i I A i ( j I ) p j (a j ) u i (a 1, a 2,..., a I ), i.e., the expected payoff if every player samples according to their random strategy independently. For strategy profile a = (a 1, a 2,..., a I ) and a i A i, let (a i, a i ) = (a 1,... a i 1, a i, a i+1,..., a I ), i.e., the strategy profile obtained from a by changing player i s strategy from a i to a i. We will use the same notation for mixed strategy profiles. Definition 4.5 A mixed strategy profile p = (p 1,..., p I ) is a mixed Nash equilibrium, if for all i I and q i (A i ), u i (q i, p i ) u i (p i, p i ). If each p i is a pure strategy (assigning probability 1 to a single element from A i ), we call p a pure Nash equilibrium. 37

3 Example 4.6 The Prisoners Dilemma game has one pure Nash equilibrium (A, A). Rock, Paper, Scissors has no pure Nash equilibrium, but a mixed equilibrium in which both players choose P, R, S with probability 1 3 each. (Easy to check: If one player randomizes uniformly, every mixed strategy performs equally well for the other player, so deviating does not increase the payoff.) Example 4.7 (Bach or Mozart) Two players want to decide whether to go to a Bach concert or one of Mozart. They want to go together, but prefer different alternatives: There are 2 pure Nash equilibria: P uree 1 : (B, B) P uree 2 : (M, M) B M B (2, 1) (0, 0) M (0, 0) (1, 2) There is also a third mixed Nash equilibrium MixedE: row player: P r(b) = 2 3, P r(m) = 1 3 column player: P r(b) = 1 3, P r(m) = 2 3 Some critiques of the Nash equilibrium concept: 1. Idea of a Nash equilibrium is that player 1 plays her side of the equilibrium, because she believes that player 2 plays her side of the equilibrium, because she thinks that player 1 plays her side of the equilibrium, because... But if there are multiple equilibria, why should we believe that players will be able to coordinate their beliefs? 2. Look at the payoff in Bach or Mozart : P uree 1 : (2, 1) P uree 2 : (1, 2) MixedE : ( 2 3, 2 3 ) So different equilibria result in different payoffs. If we can t predict which Nash equilibrium will be reached, we also can t predict the payoffs. In this lecture we will address these critiques, showing that players arrive at an equilibrium by playing a game repeatedly and using learning rules to adopt to their opponent s behavior. Definition 4.8 A two player zero sum game is one in which I = {1, 2} and u 2 (a 1, a 2 ) = u 1 (a 1, a 2 ) for all pure strategy profiles (a 1, a 2 ). Rock, Paper, Scissors is an example of a zero sum game (actually, even a win/lose game with payoffs in { 1, 1}). 38

4 4.1 Von Neumann s Minimax-Theorem Theorem 4.9 Let a two-player zero-sum game G = (I, (A i ), (u i )) be given. Define and v min 1 = max v max 1 = min min u 1(p, q) max u 1(p, q). It holds that v min 1 = v max 1. We call this value V the game value of G. Intuitively, Theorem 4.9 says the following: The best payoff player 1 can guarantee for herself, if she has to pick a strategy first and player 2 is then allowed to play a best response ( v1 min ), is equal to the minimum payoff she can achieve if player 2 has to go first, and player 1 is allowed to respond optimally (v1 max ). This has some very nice consequences. Corollary 4.10 Let G be a two-player zero-sum game with game value V. In every mixed Nash equilibrium (p, q ), we have u 1 (p, q ) = V. Proof: Since p is a best response to q, u 1 (p, q ) = max u 1 (p, q ) min max u 1(p, q) = v1 max = V. Similary, since q is a best response to p, u 2 (p, q ) V. So u 1 (p, q ) = u 2 (p, q ) V. Corollary 4.11 Let G be a two-player zero-sum game. A mixed strategy profile (p, q ) is a mixed Nash equilibrium, if and only if and p argmax min u 1(p, q) q argmin max u 1(p, q). In particular, the set of mixed Nash equilibria is non-empty. Proof: : Let (p, q ) be a Nash equilibrium. By Corollary 4.10, u 1 (p, q ) = V. Since q is a best response to p, min u 1 (p, q) = u 1 (p, q ) = V = max min u 1 (p, q). Thus, p argmax p (A1 ) min q (A2 ) u 1 (p, q). By symmetry, the same argument applies to player 2, as well. 39

5 : Let strategy profile (p, q ) with p, q from the respective sets of mixed strategies be given. Since p argmax p (A1 ) min q (A2 ) u 1 (p, q), we have u 1 (p, q ) V. On the other hand, for any p (A 1 ), u 1 (p, q ) max u 1 (p, q ) = min max u 1 (p, q) = V where the last line follows because q argmin q (A2 ) max p (A1 ) u 1 (p, q). So player 1 has no incentive to defect to a different strategy. Again, by symmetry, the same argument applies to player 2. Remark Note that Corollary 4.11 resolves one of the critiques of the Nash equilibrium concept. Players don t have to coordinate their actions in order to reach an equilibrium, but can pick strategies from their respective argmax-sets independently. Recall the MaxHedge algorithm from Homework Assignment 4. We showed: ( T ) Theorem 4.12 Algorithm MaxHedge (with n experts, costs in 0, 1) has regret O ln(n) against the class of adaptive adversaries. In particular, for any sequence of reward functions r t : n 0, 1, the sequence of experts x 1,..., x T selected by the algorithm satisfies T E r t (x t ) E max p (n) ( n ) p x r t (x) x=1 O T ln(n). In Theorem 4.12 above we compare our algorithm to the best mixture of experts. However, the bound follows immediately by observing that the maximum on the right hand side is always achieved by a single best expert. Proof of Theorem 4.9: We start with the easy direction and prove that v min 1 v max 1 : For any strategy profile ( ˆp, ˆq), Thus, u 1 (ˆp, ˆq) min u 1 (ˆp, q) and, taking the maximum of both sides, v min 1 = max min u 1 (p, q) max u 1 (p, ˆq). min min max u 1 (p, q) max u 1 (p, q) = v max To prove the other direction v1 max v1 min, we will use the existence of expert learning algorithms with vanishing per-time-step regret (as, e.g., Hedge). We assume w.l.o.g. that u 1 (p, q) 0, 1 (and, thus, u 2 (p, q) 1, 0) for all p (A 1 ), q (A 2 ). This can always be achieved by applying an appropriate linear transformation to the game matrix

6 Note, that Similarly, v max 1 = v min 2. v min 1 = max = max = min = v max 2. min u 1 (p, q) min ( u 2 (p, q)) max u 2 (p, q) Now assume that for T steps, both players use an expert learning algorithm to determine their strategy. Formally, let n = max { A 1, A 2 }. Player 1 applies the algorithm with one expert for each a A 1. If player 2 plays strategy b at time t, the reward of expert a is r t (a) = u 1 (a, b). Player 2 applies the algorithm with an expert for each b A 2 and reward functions r t (b) = u 2 (b, a) + 1 (making sure rewards are in 0, 1). Let a 1,..., a T and b 1,...(, b T be the strategies selected by the 2 players and assume that both T ) algorithms have regret O ln(n). Define mixed strategies p = 1 T a t and q = 1 T b t. (The above is somewhat sloppy notation. We associate each a t with the vector a t (A 1 ) that assigns probability 1 to strategy a t A 1.) Intuitively, strategies p and q mix pure strategies proportional to the frequency with which they have been played. Let p argmax p (A1 ) u 1 (p, q), q argmax q (A2 ) u 2 (p, q) be best responses to q and p, respectively. Note, that u 1 (p, q) = max u 1 (p, q) min max u 1 (p, q) = v max Analogously, u 2 (p, q ) v2 max. By our regret-bound, 1 1 E u 1 (a t, b t ) E u 1 (p, b t ) T T = E u 1 (p, q) O 1. O ln(n)/t ( ln(n)/t ). 41

7 Finally, combining the above, v1 max O ln(n)/t E u 1 (p, q) O ln(n)/t E = E 1 T 1 T u 1 (a t, b t ) u 2 (a t, b t ) E u 2 (p, q ) + O ln(n)/t v2 max + O ln(n)/t = v1 min + O ln(n)/t. Now the claim follows for T by compactness of (A 1 ), (A 2 ). Remark By the last set of inequalities in the proof of Theorem 4.9, running the regret-minimizing algorithms for Ω(ln(n)/(δ/2) 2 ) steps, yields a strategy profile (p, q), such that E max u 1(p, q) = Eu 1 (p, q) V + δ/2. Consequently, Eu 1 (p, q) V + δ/2 and, since the game is zero sum, Eu 2 (p, q) V δ/2. Similarly, for player 2 we have that E max u 2(p, q) = Eu 2 (p, q ) V + δ/2. Thus, Eu 2 (p, q) V + δ/2 and Eu 1 (p, q) V δ/2. Such a strategy profile, in which no player can gain more than δ by deviating unilaterally is called a δ-approximate Nash equilibrium. Since we only run the algorithms for O(log n) steps, the mixed strategies p and q assign positive probability to at most O(log n) pure strategies. We conclude that every two-player zero-sum game possesses δ-approximate Nash equilibria with small support of size O(log n). Remark Using Markov s inequality we can turn the above existence result into a constructive procedure. Running the algorithms for Ω(ln(n)/(δ/2c) 2 ) steps for any c > 0, we obtain a mixed strategy profile (p, q) with ( Prob max u 1(p, q) δ ) = 1 2 c and support of size O(log n). Running the procedure repeatedly yields the desired δ-approximate Nash equilibrium with probability exponentially close to 1. δ 2c δ 2 Remark Assume that a zero-sum game is played repeatedly. By using a regret-minimizing algorithm like Hedge, player 1 comes close to the best possible payoff against whatever distribution of strategies player 2 happens to use. It is not necessary to assume that player 2 is acting rationally. 42

8 4.2 Yao s Minimax Principle Consider a computational problem with I - a finite set of possible inputs and A - a finite set of possbible algorithms. For all i I, a A denote by t(i, a) the running time of algorithm a on input i. We can think of this as a game, in which player 1 gets to select the input, while player 2 is allowed to pick an algorithm. Similar to the payoff functions in the previous sections, we can generalize t to distributions p (I), q (A) as t(p, q) = t(i, a)p(i)q(a). i I a A In words, t(p, q) denotes the expected running time of an algorithm from distribution q on an input from distribution p. Theorem 4.13 In the setting described above, it holds that max min t(p, a) = min p (I) a A max q (A) i I t(i, q). Proof: This follows immediately from von Neumann s Minimax Theorem 4.9 and the observation that for any mixed strategy p, q of one of the players, the other player has a pure strategy that constitutes a best response. Yao s Minimax Principle states that, for finite sets of algorithms and inputs, the best worst-case running time achievable by any randomized algorithm (the right hand side), is equal to the best running time obtainable by a deterministic algorithm on a worst-case distribution of inputs. This is very helpful in proving lower bounds on the performance guarantee of randomized algorithms (e.g., in online computation). If one can construct a distribution on inputs, on which no deterministic algorithm performs well in expectation (which is often much easier than arguing about randomized algorithms), then it follows that there must exist an instance on which no randomized algorithm performs well. 43

6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation 6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation Daron Acemoglu and Asu Ozdaglar MIT November 2, 2009 1 Introduction Outline The problem of cooperation Finitely-repeated prisoner s dilemma

More information

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games 6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games Asu Ozdaglar MIT February 4, 2009 1 Introduction Outline Decisions, utility maximization Strategic form games Best responses

More information

1 Nonzero sum games and Nash equilibria

1 Nonzero sum games and Nash equilibria princeton univ. F 14 cos 521: Advanced Algorithm Design Lecture 19: Equilibria and algorithms Lecturer: Sanjeev Arora Scribe: Economic and game-theoretic reasoning specifically, how agents respond to economic

More information

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015 ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1

More information

Linear Programming: Chapter 11 Game Theory

Linear Programming: Chapter 11 Game Theory Linear Programming: Chapter 11 Game Theory Robert J. Vanderbei October 17, 2007 Operations Research and Financial Engineering Princeton University Princeton, NJ 08544 http://www.princeton.edu/ rvdb Rock-Paper-Scissors

More information

1 Portfolio Selection

1 Portfolio Selection COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture # Scribe: Nadia Heninger April 8, 008 Portfolio Selection Last time we discussed our model of the stock market N stocks start on day with

More information

ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY 100-3 90-99 21 80-89 14 70-79 4 0-69 11

ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY 100-3 90-99 21 80-89 14 70-79 4 0-69 11 The distribution of grades was as follows. ECO 199 B GAMES OF STRATEGY Spring Term 2004 PROBLEM SET 4 B DRAFT ANSWER KEY Range Numbers 100-3 90-99 21 80-89 14 70-79 4 0-69 11 Question 1: 30 points Games

More information

Computational Learning Theory Spring Semester, 2003/4. Lecture 1: March 2

Computational Learning Theory Spring Semester, 2003/4. Lecture 1: March 2 Computational Learning Theory Spring Semester, 2003/4 Lecture 1: March 2 Lecturer: Yishay Mansour Scribe: Gur Yaari, Idan Szpektor 1.1 Introduction Several fields in computer science and economics are

More information

Row Echelon Form and Reduced Row Echelon Form

Row Echelon Form and Reduced Row Echelon Form These notes closely follow the presentation of the material given in David C Lay s textbook Linear Algebra and its Applications (3rd edition) These notes are intended primarily for in-class presentation

More information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1. MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

More information

Lecture V: Mixed Strategies

Lecture V: Mixed Strategies Lecture V: Mixed Strategies Markus M. Möbius February 26, 2008 Osborne, chapter 4 Gibbons, sections 1.3-1.3.A 1 The Advantage of Mixed Strategies Consider the following Rock-Paper-Scissors game: Note that

More information

Minimax Strategies. Minimax Strategies. Zero Sum Games. Why Zero Sum Games? An Example. An Example

Minimax Strategies. Minimax Strategies. Zero Sum Games. Why Zero Sum Games? An Example. An Example Everyone who has studied a game like poker knows the importance of mixing strategies With a bad hand, you often fold But you must bluff sometimes Lectures in Microeconomics-Charles W Upton Zero Sum Games

More information

Game Theory and Nash Equilibrium

Game Theory and Nash Equilibrium Game Theory and Nash Equilibrium by Jenny Duffy A project submitted to the Department of Mathematical Sciences in conformity with the requirements for Math 4301 (Honours Seminar) Lakehead University Thunder

More information

Optimization in ICT and Physical Systems

Optimization in ICT and Physical Systems 27. OKTOBER 2010 in ICT and Physical Systems @ Aarhus University, Course outline, formal stuff Prerequisite Lectures Homework Textbook, Homepage and CampusNet, http://kurser.iha.dk/ee-ict-master/tiopti/

More information

10 Evolutionarily Stable Strategies

10 Evolutionarily Stable Strategies 10 Evolutionarily Stable Strategies There is but a step between the sublime and the ridiculous. Leo Tolstoy In 1973 the biologist John Maynard Smith and the mathematician G. R. Price wrote an article in

More information

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where. Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S

More information

Game Theory 1. Introduction

Game Theory 1. Introduction Game Theory 1. Introduction Dmitry Potapov CERN What is Game Theory? Game theory is about interactions among agents that are self-interested I ll use agent and player synonymously Self-interested: Each

More information

On the Complexity of Two-Player Win-Lose Games

On the Complexity of Two-Player Win-Lose Games On the Complexity of Two-Player Win-Lose Games Tim Abbott, Daniel Kane, Paul Valiant April 8, 005 Abstract The efficient computation of Nash equilibria is one of the most formidable challenges in computational

More information

The Multiplicative Weights Update method

The Multiplicative Weights Update method Chapter 2 The Multiplicative Weights Update method The Multiplicative Weights method is a simple idea which has been repeatedly discovered in fields as diverse as Machine Learning, Optimization, and Game

More information

160 CHAPTER 4. VECTOR SPACES

160 CHAPTER 4. VECTOR SPACES 160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results

More information

Chapter 7. Sealed-bid Auctions

Chapter 7. Sealed-bid Auctions Chapter 7 Sealed-bid Auctions An auction is a procedure used for selling and buying items by offering them up for bid. Auctions are often used to sell objects that have a variable price (for example oil)

More information

THE DIMENSION OF A VECTOR SPACE

THE DIMENSION OF A VECTOR SPACE THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field

More information

Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

More information

4 Learning, Regret minimization, and Equilibria

4 Learning, Regret minimization, and Equilibria 4 Learning, Regret minimization, and Equilibria A. Blum and Y. Mansour Abstract Many situations involve repeatedly making decisions in an uncertain environment: for instance, deciding what route to drive

More information

Sequential lmove Games. Using Backward Induction (Rollback) to Find Equilibrium

Sequential lmove Games. Using Backward Induction (Rollback) to Find Equilibrium Sequential lmove Games Using Backward Induction (Rollback) to Find Equilibrium Sequential Move Class Game: Century Mark Played by fixed pairs of players taking turns. At each turn, each player chooses

More information

Solving Systems of Linear Equations

Solving Systems of Linear Equations LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

More information

Game Theory and Algorithms Lecture 10: Extensive Games: Critiques and Extensions

Game Theory and Algorithms Lecture 10: Extensive Games: Critiques and Extensions Game Theory and Algorithms Lecture 0: Extensive Games: Critiques and Extensions March 3, 0 Summary: We discuss a game called the centipede game, a simple extensive game where the prediction made by backwards

More information

Games Manipulators Play

Games Manipulators Play Games Manipulators Play Umberto Grandi Department of Mathematics University of Padova 23 January 2014 [Joint work with Edith Elkind, Francesca Rossi and Arkadii Slinko] Gibbard-Satterthwaite Theorem All

More information

AN ANALYSIS OF A WAR-LIKE CARD GAME. Introduction

AN ANALYSIS OF A WAR-LIKE CARD GAME. Introduction AN ANALYSIS OF A WAR-LIKE CARD GAME BORIS ALEXEEV AND JACOB TSIMERMAN Abstract. In his book Mathematical Mind-Benders, Peter Winkler poses the following open problem, originally due to the first author:

More information

MATH 340: MATRIX GAMES AND POKER

MATH 340: MATRIX GAMES AND POKER MATH 340: MATRIX GAMES AND POKER JOEL FRIEDMAN Contents 1. Matrix Games 2 1.1. A Poker Game 3 1.2. A Simple Matrix Game: Rock/Paper/Scissors 3 1.3. A Simpler Game: Even/Odd Pennies 3 1.4. Some Examples

More information

Dynamics and Equilibria

Dynamics and Equilibria Dynamics and Equilibria Sergiu Hart Presidential Address, GAMES 2008 (July 2008) Revised and Expanded (November 2009) Revised (2010, 2011, 2012, 2013) SERGIU HART c 2008 p. 1 DYNAMICS AND EQUILIBRIA Sergiu

More information

Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

More information

From the probabilities that the company uses to move drivers from state to state the next year, we get the following transition matrix:

From the probabilities that the company uses to move drivers from state to state the next year, we get the following transition matrix: MAT 121 Solutions to Take-Home Exam 2 Problem 1 Car Insurance a) The 3 states in this Markov Chain correspond to the 3 groups used by the insurance company to classify their drivers: G 0, G 1, and G 2

More information

Equilibrium computation: Part 1

Equilibrium computation: Part 1 Equilibrium computation: Part 1 Nicola Gatti 1 Troels Bjerre Sorensen 2 1 Politecnico di Milano, Italy 2 Duke University, USA Nicola Gatti and Troels Bjerre Sørensen ( Politecnico di Milano, Italy, Equilibrium

More information

I d Rather Stay Stupid: The Advantage of Having Low Utility

I d Rather Stay Stupid: The Advantage of Having Low Utility I d Rather Stay Stupid: The Advantage of Having Low Utility Lior Seeman Department of Computer Science Cornell University lseeman@cs.cornell.edu Abstract Motivated by cost of computation in game theory,

More information

ECON 40050 Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted.

ECON 40050 Game Theory Exam 1 - Answer Key. 4) All exams must be turned in by 1:45 pm. No extensions will be granted. 1 ECON 40050 Game Theory Exam 1 - Answer Key Instructions: 1) You may use a pen or pencil, a hand-held nonprogrammable calculator, and a ruler. No other materials may be at or near your desk. Books, coats,

More information

How to Solve Strategic Games? Dominant Strategies

How to Solve Strategic Games? Dominant Strategies How to Solve Strategic Games? There are three main concepts to solve strategic games: 1. Dominant Strategies & Dominant Strategy Equilibrium 2. Dominated Strategies & Iterative Elimination of Dominated

More information

Mathematical finance and linear programming (optimization)

Mathematical finance and linear programming (optimization) Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may

More information

Bonus Maths 2: Variable Bet Sizing in the Simplest Possible Game of Poker (JB)

Bonus Maths 2: Variable Bet Sizing in the Simplest Possible Game of Poker (JB) Bonus Maths 2: Variable Bet Sizing in the Simplest Possible Game of Poker (JB) I recently decided to read Part Three of The Mathematics of Poker (TMOP) more carefully than I did the first time around.

More information

Compact Representations and Approximations for Compuation in Games

Compact Representations and Approximations for Compuation in Games Compact Representations and Approximations for Compuation in Games Kevin Swersky April 23, 2008 Abstract Compact representations have recently been developed as a way of both encoding the strategic interactions

More information

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

More information

Notes from Week 1: Algorithms for sequential prediction

Notes from Week 1: Algorithms for sequential prediction CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking

More information

Probabilistic Strategies: Solutions

Probabilistic Strategies: Solutions Probability Victor Xu Probabilistic Strategies: Solutions Western PA ARML Practice April 3, 2016 1 Problems 1. You roll two 6-sided dice. What s the probability of rolling at least one 6? There is a 1

More information

1 Representation of Games. Kerschbamer: Commitment and Information in Games

1 Representation of Games. Kerschbamer: Commitment and Information in Games 1 epresentation of Games Kerschbamer: Commitment and Information in Games Game-Theoretic Description of Interactive Decision Situations This lecture deals with the process of translating an informal description

More information

An Introduction to Competitive Analysis for Online Optimization. Maurice Queyranne University of British Columbia, and IMA Visitor (Fall 2002)

An Introduction to Competitive Analysis for Online Optimization. Maurice Queyranne University of British Columbia, and IMA Visitor (Fall 2002) An Introduction to Competitive Analysis for Online Optimization Maurice Queyranne University of British Columbia, and IMA Visitor (Fall 2002) IMA, December 11, 2002 Overview 1 Online Optimization sequences

More information

Math 202-0 Quizzes Winter 2009

Math 202-0 Quizzes Winter 2009 Quiz : Basic Probability Ten Scrabble tiles are placed in a bag Four of the tiles have the letter printed on them, and there are two tiles each with the letters B, C and D on them (a) Suppose one tile

More information

(67902) Topics in Theory and Complexity Nov 2, 2006. Lecture 7

(67902) Topics in Theory and Complexity Nov 2, 2006. Lecture 7 (67902) Topics in Theory and Complexity Nov 2, 2006 Lecturer: Irit Dinur Lecture 7 Scribe: Rani Lekach 1 Lecture overview This Lecture consists of two parts In the first part we will refresh the definition

More information

7 Gaussian Elimination and LU Factorization

7 Gaussian Elimination and LU Factorization 7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method

More information

2.1 Complexity Classes

2.1 Complexity Classes 15-859(M): Randomized Algorithms Lecturer: Shuchi Chawla Topic: Complexity classes, Identity checking Date: September 15, 2004 Scribe: Andrew Gilpin 2.1 Complexity Classes In this lecture we will look

More information

ALMOST COMMON PRIORS 1. INTRODUCTION

ALMOST COMMON PRIORS 1. INTRODUCTION ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type

More information

Infinitely Repeated Games with Discounting Ù

Infinitely Repeated Games with Discounting Ù Infinitely Repeated Games with Discounting Page 1 Infinitely Repeated Games with Discounting Ù Introduction 1 Discounting the future 2 Interpreting the discount factor 3 The average discounted payoff 4

More information

Mechanisms for Fair Attribution

Mechanisms for Fair Attribution Mechanisms for Fair Attribution Eric Balkanski Yaron Singer Abstract We propose a new framework for optimization under fairness constraints. The problems we consider model procurement where the goal is

More information

1.2 Solving a System of Linear Equations

1.2 Solving a System of Linear Equations 1.. SOLVING A SYSTEM OF LINEAR EQUATIONS 1. Solving a System of Linear Equations 1..1 Simple Systems - Basic De nitions As noticed above, the general form of a linear system of m equations in n variables

More information

3.2 Roulette and Markov Chains

3.2 Roulette and Markov Chains 238 CHAPTER 3. DISCRETE DYNAMICAL SYSTEMS WITH MANY VARIABLES 3.2 Roulette and Markov Chains In this section we will be discussing an application of systems of recursion equations called Markov Chains.

More information

Adaptive Online Gradient Descent

Adaptive Online Gradient Descent Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650

More information

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents William H. Sandholm January 6, 22 O.. Imitative protocols, mean dynamics, and equilibrium selection In this section, we consider

More information

A Dutch Book for Group Decision-Making?

A Dutch Book for Group Decision-Making? A Dutch Book for Group Decision-Making? Forthcoming in Benedikt Löwe, Eric Pacuit, Jan-Willem Romeijn (eds.) Foundations of the Formal Sciences VI--Reasoning about Probabilities and Probabilistic Reasoning.

More information

The Ergodic Theorem and randomness

The Ergodic Theorem and randomness The Ergodic Theorem and randomness Peter Gács Department of Computer Science Boston University March 19, 2008 Peter Gács (Boston University) Ergodic theorem March 19, 2008 1 / 27 Introduction Introduction

More information

Lecture 4: AC 0 lower bounds and pseudorandomness

Lecture 4: AC 0 lower bounds and pseudorandomness Lecture 4: AC 0 lower bounds and pseudorandomness Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: Jason Perry and Brian Garnett In this lecture,

More information

A Game Theoretical Framework for Adversarial Learning

A Game Theoretical Framework for Adversarial Learning A Game Theoretical Framework for Adversarial Learning Murat Kantarcioglu University of Texas at Dallas Richardson, TX 75083, USA muratk@utdallas Chris Clifton Purdue University West Lafayette, IN 47907,

More information

Methods for Finding Bases

Methods for Finding Bases Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,

More information

Equilibrium Bids in Sponsored Search. Auctions: Theory and Evidence

Equilibrium Bids in Sponsored Search. Auctions: Theory and Evidence Equilibrium Bids in Sponsored Search Auctions: Theory and Evidence Tilman Börgers Ingemar Cox Martin Pesendorfer Vaclav Petricek September 2008 We are grateful to Sébastien Lahaie, David Pennock, Isabelle

More information

4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

More information

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)! Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following

More information

Game theory and AI: a unified approach to poker games

Game theory and AI: a unified approach to poker games Game theory and AI: a unified approach to poker games Thesis for graduation as Master of Artificial Intelligence University of Amsterdam Frans Oliehoek 2 September 2005 ii Abstract This thesis focuses

More information

Decision Theory. 36.1 Rational prospecting

Decision Theory. 36.1 Rational prospecting 36 Decision Theory Decision theory is trivial, apart from computational details (just like playing chess!). You have a choice of various actions, a. The world may be in one of many states x; which one

More information

1 Solving LPs: The Simplex Algorithm of George Dantzig

1 Solving LPs: The Simplex Algorithm of George Dantzig Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

More information

Microeconomic Theory Jamison / Kohlberg / Avery Problem Set 4 Solutions Spring 2012. (a) LEFT CENTER RIGHT TOP 8, 5 0, 0 6, 3 BOTTOM 0, 0 7, 6 6, 3

Microeconomic Theory Jamison / Kohlberg / Avery Problem Set 4 Solutions Spring 2012. (a) LEFT CENTER RIGHT TOP 8, 5 0, 0 6, 3 BOTTOM 0, 0 7, 6 6, 3 Microeconomic Theory Jamison / Kohlberg / Avery Problem Set 4 Solutions Spring 2012 1. Subgame Perfect Equilibrium and Dominance (a) LEFT CENTER RIGHT TOP 8, 5 0, 0 6, 3 BOTTOM 0, 0 7, 6 6, 3 Highlighting

More information

0.0.2 Pareto Efficiency (Sec. 4, Ch. 1 of text)

0.0.2 Pareto Efficiency (Sec. 4, Ch. 1 of text) September 2 Exercises: Problem 2 (p. 21) Efficiency: p. 28-29: 1, 4, 5, 6 0.0.2 Pareto Efficiency (Sec. 4, Ch. 1 of text) We discuss here a notion of efficiency that is rooted in the individual preferences

More information

Dynamic TCP Acknowledgement: Penalizing Long Delays

Dynamic TCP Acknowledgement: Penalizing Long Delays Dynamic TCP Acknowledgement: Penalizing Long Delays Karousatou Christina Network Algorithms June 8, 2010 Karousatou Christina (Network Algorithms) Dynamic TCP Acknowledgement June 8, 2010 1 / 63 Layout

More information

Basics of Counting. The product rule. Product rule example. 22C:19, Chapter 6 Hantao Zhang. Sample question. Total is 18 * 325 = 5850

Basics of Counting. The product rule. Product rule example. 22C:19, Chapter 6 Hantao Zhang. Sample question. Total is 18 * 325 = 5850 Basics of Counting 22C:19, Chapter 6 Hantao Zhang 1 The product rule Also called the multiplication rule If there are n 1 ways to do task 1, and n 2 ways to do task 2 Then there are n 1 n 2 ways to do

More information

4.6 Linear Programming duality

4.6 Linear Programming duality 4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal

More information

Online Algorithms: Learning & Optimization with No Regret.

Online Algorithms: Learning & Optimization with No Regret. Online Algorithms: Learning & Optimization with No Regret. Daniel Golovin 1 The Setup Optimization: Model the problem (objective, constraints) Pick best decision from a feasible set. Learning: Model the

More information

Game Theory: Supermodular Games 1

Game Theory: Supermodular Games 1 Game Theory: Supermodular Games 1 Christoph Schottmüller 1 License: CC Attribution ShareAlike 4.0 1 / 22 Outline 1 Introduction 2 Model 3 Revision questions and exercises 2 / 22 Motivation I several solution

More information

Online Convex Optimization

Online Convex Optimization E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting

More information

Optimal Online-list Batch Scheduling

Optimal Online-list Batch Scheduling Optimal Online-list Batch Scheduling Jacob Jan Paulus a,, Deshi Ye b, Guochuan Zhang b a University of Twente, P.O. box 217, 7500AE Enschede, The Netherlands b Zhejiang University, Hangzhou 310027, China

More information

Lecture 2: Universality

Lecture 2: Universality CS 710: Complexity Theory 1/21/2010 Lecture 2: Universality Instructor: Dieter van Melkebeek Scribe: Tyson Williams In this lecture, we introduce the notion of a universal machine, develop efficient universal

More information

On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information

On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information Finance 400 A. Penati - G. Pennacchi Notes on On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information by Sanford Grossman This model shows how the heterogeneous information

More information

6.254 : Game Theory with Engineering Applications Lecture 1: Introduction

6.254 : Game Theory with Engineering Applications Lecture 1: Introduction 6.254 : Game Theory with Engineering Applications Lecture 1: Introduction Asu Ozdaglar MIT February 2, 2010 1 Introduction Optimization Theory: Optimize a single objective over a decision variable x R

More information

Oligopoly markets: The price or quantity decisions by one rm has to directly in uence pro ts by other rms if rms are competing for customers.

Oligopoly markets: The price or quantity decisions by one rm has to directly in uence pro ts by other rms if rms are competing for customers. 15 Game Theory Varian: Chapters 8-9. The key novelty compared to the competitive (Walrasian) equilibrium analysis is that game theoretic analysis allows for the possibility that utility/pro t/payo s depend

More information

A Game Theoretical Framework on Intrusion Detection in Heterogeneous Networks Lin Chen, Member, IEEE, and Jean Leneutre

A Game Theoretical Framework on Intrusion Detection in Heterogeneous Networks Lin Chen, Member, IEEE, and Jean Leneutre IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL 4, NO 2, JUNE 2009 165 A Game Theoretical Framework on Intrusion Detection in Heterogeneous Networks Lin Chen, Member, IEEE, and Jean Leneutre

More information

The Basics of Game Theory

The Basics of Game Theory Sloan School of Management 15.010/15.011 Massachusetts Institute of Technology RECITATION NOTES #7 The Basics of Game Theory Friday - November 5, 2004 OUTLINE OF TODAY S RECITATION 1. Game theory definitions:

More information

Solution to Homework 2

Solution to Homework 2 Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if

More information

MTH6120 Further Topics in Mathematical Finance Lesson 2

MTH6120 Further Topics in Mathematical Finance Lesson 2 MTH6120 Further Topics in Mathematical Finance Lesson 2 Contents 1.2.3 Non-constant interest rates....................... 15 1.3 Arbitrage and Black-Scholes Theory....................... 16 1.3.1 Informal

More information

9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1

9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1 9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1 Seffi Naor Computer Science Dept. Technion Haifa, Israel Introduction

More information

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

More information

When is Reputation Bad? 1

When is Reputation Bad? 1 When is Reputation Bad? 1 Jeffrey Ely Drew Fudenberg David K Levine 2 First Version: April 22, 2002 This Version: November 20, 2005 Abstract: In traditional reputation theory, the ability to build a reputation

More information

Unit 18 Determinants

Unit 18 Determinants Unit 18 Determinants Every square matrix has a number associated with it, called its determinant. In this section, we determine how to calculate this number, and also look at some of the properties of

More information

Game Theory in Wireless Networks: A Tutorial

Game Theory in Wireless Networks: A Tutorial 1 Game heory in Wireless Networks: A utorial Mark Felegyhazi, Jean-Pierre Hubaux EPFL Switzerland email: {mark.felegyhazi, jean-pierre.hubaux}@epfl.ch EPFL echnical report: LCA-REPOR-2006-002, submitted

More information

Orthogonal Diagonalization of Symmetric Matrices

Orthogonal Diagonalization of Symmetric Matrices MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

More information

International Doctoral School Algorithmic Decision Theory: MCDA and MOO

International Doctoral School Algorithmic Decision Theory: MCDA and MOO International Doctoral School Algorithmic Decision Theory: MCDA and MOO Lecture 2: Multiobjective Linear Programming Department of Engineering Science, The University of Auckland, New Zealand Laboratoire

More information

Hacking-proofness and Stability in a Model of Information Security Networks

Hacking-proofness and Stability in a Model of Information Security Networks Hacking-proofness and Stability in a Model of Information Security Networks Sunghoon Hong Preliminary draft, not for citation. March 1, 2008 Abstract We introduce a model of information security networks.

More information

Decision Making under Uncertainty

Decision Making under Uncertainty 6.825 Techniques in Artificial Intelligence Decision Making under Uncertainty How to make one decision in the face of uncertainty Lecture 19 1 In the next two lectures, we ll look at the question of how

More information

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α

More information

A Game Theoretic Model of Spam E-Mailing

A Game Theoretic Model of Spam E-Mailing A Game Theoretic Model of Spam E-Mailing Ion Androutsopoulos, Evangelos F. Magirou and Dimitrios K. Vassilakis Department of Informatics Athens University of Economics and Business Patission 76, GR-04

More information

Advanced Game Theory

Advanced Game Theory Advanced Game Theory Herve Moulin Baker Hall 263; moulin@rice.edu Rice University ECO 440 Spring 2007 1 Two person zero sum games 1.1 Introduction: strategic interdependency In this section we study games

More information

1 Norms and Vector Spaces

1 Norms and Vector Spaces 008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)

More information

Notes on Continuous Random Variables

Notes on Continuous Random Variables Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes

More information

Solutions for Practice problems on proofs

Solutions for Practice problems on proofs Solutions for Practice problems on proofs Definition: (even) An integer n Z is even if and only if n = 2m for some number m Z. Definition: (odd) An integer n Z is odd if and only if n = 2m + 1 for some

More information

1 The Line vs Point Test

1 The Line vs Point Test 6.875 PCP and Hardness of Approximation MIT, Fall 2010 Lecture 5: Low Degree Testing Lecturer: Dana Moshkovitz Scribe: Gregory Minton and Dana Moshkovitz Having seen a probabilistic verifier for linearity

More information