A simple analysis of the TV game WHO WANTS TO BE A MILLIONAIRE? R


 Ami Simpson
 2 years ago
 Views:
Transcription
1 A simple analysis of the TV game WHO WANTS TO BE A MILLIONAIRE? R Federico Perea Justo Puerto MaMaEuSch Management Mathematics for European Schools CP DE  COMENIUS  C21 University of Seville This project has been carried out with the partial support of the European Union in the framework of the Sokrates programme. The content does not necessarily reflect the position of the European Union, nor does it involve any responsibility on the part of the European Union. 0
2 1 Introduction This article is about the popular TV game show Who Wants To Be A Millionaire? R. When this paper was written there were 45 versions of Who Wants To Be A Millionaire? R, presented in 71 countries. In more than 100 countries the licence has been bought by TV stations and the show will be broadcasted sooner or later. Who Wants To Be A Millionaire? R debuted in the United Kingdom in September of 1998 and was very successful there. Afterwards, it spread all over the world, coming to Spain in the summer of Here, it was broadcasted by the TV station Telecinco under the title Quiere ser millonario? R. The rules of the game are similar in all countries. In this article, we will regard only the Spanish version. One candidate is chosen out of a pool of ten and has the chance of winning the top prize of Euros. In order to achieve this, she must answer 15 multiplechoicequestions correctly in a row. The contestant may quit at any time by keeping her earnings. In each step, she is shown the question and four possible answers before deciding whether to play or not. Once she has decided to stay in the game the next answer has to be correct to continue playing. Each question has a certain monetary value, given here in Euros. The money that a candidate can win if she answers the question correctly is given in table 1: There are three stages ( guarantee points ) where the money is banked and cannot be lost even if the candidate offers an incorrect answer to one of the next questions: 1800, and Euros. There is no time limit for answering a question. If time runs out in a particular day, the next programme continues with that player s game. At any point, the contestant may use one or more of her three lifelines. These are: 50:50 option: the computer eliminates two possible answers leaving one wrong answer and the correct one. Phone a friend: the contestant may discuss the question with a friend or relative on the phone for 30 seconds. Ask the audience: the audience has the option of choosing the answer they consider correct by pressing the corresponding button on their keypads. The result of this poll will be listed in percentages. In the sequel we will refer to those lifelines as 1
3 question index monetary value Table 1: Immediate rewards lifeline 1 for the 50:50 option, lifeline 2 for Phone a friend, lifeline 3 for Ask the audience. Each lifeline may be used only once during a contestant s entire game. The primary aim of this paper is to show how a difficult real decision making problem can be easily modelled and solved by basic Operations Research tools, in our case by Discrete event Dynamical Programming. In this regard, the paper is quite simple from the mathematical point of view. Nevertheless, the modelling phase is not so straightforward and, moreover, this approach can be used as motivating case analysis when presenting Dynamical Programming in classroom. This aim is achieved performing different phases: 1. modelling, 2. mathematical formulation, 3. simulation of the actual process. 2
4 In the modelling phase we identify the essential building blocks that describe the problem and link them to elements of mathematical models. In the formulation phase we propose a description of the game as a discretetime Markov decision process that is solved by discrete event dynamic programming. Two models are presented that guide the players to find optimal strategies maximizing their expected reward, which will be called maximum expected strategy, and optimal strategies to maximize the probability of reaching a given question, which will be called maximum probability strategy. In doing that we establish two mathematical models of the game and find optimal strategies of a candidate participating in the TVgame. This is achieved through a mathematical description of the game as a discretetime Markov decision process that is solved by discrete event dynamic programming. The rest of the paper is organized as follows: the second section is devoted to show the general mathematical model (states, feasible actions, rewards, transition function, probabilities of answering correctly and their estimations). We present in the third section the description of the first model. In that model we want to maximize the expected reward. Also in this section we show the case where we want to maximize the probability of reaching and answering correctly a given question, starting in any departing state. After that, we present some concluding remark based on simulation of how to play in a dynamic way. 2 The general model The actual game requires that the contestant makes decisions each time a question is answered correctly. The planning horizon is finite, we have N = 16 stages, where the 16 th stage stands for the situation after answering correctly question number 15. To make a decision, the candidate has to know the index of the question she faces and the lifelines she has already used. The history of the game is summarized in this information. We define S as the set of state vectors s = (k, l 1, l 2, l 3 ), where k is the index of the current question and { 1 if lifeline i may be used, l i = 0 if lifeline i was already used in an earlier question. At any state s S let A(s) denote the set of feasible actions in this state. If we suppose we are in state s = (k, l 1, l 2, l 3 ), A(s) depends on the question index and the lifelines left. If k = 16, the game is over and there are no feasible actions. If k 15, the candidate has several possibilities: Answer the question without using lifelines. 3
5 r 0 0 r r r r r r r r r r r r r r r r0 0 r1 0 r2 0 r3 0 r4 0 r r r r r r r r r r r Table 2: Immediate versus ensured rewards Answer the question employing one or more lifelines, if any is left. In this case, the candidate must also specify the lifelines she is going to use. Stop and quit the game. If the player decides not to answer, the immediate reward is the monetary value of the last question answered. If the candidate decides to answer, the immediate reward is a random variable and depends on the probability of answering correctly. If the candidate fails, the immediate reward is the last guarantee point reached before failing. If the candidate decides to answer and chooses the correct alternative, there is no immediate reward. The candidate goes on to the next question, and the reward is the expected (final) reward. Denote r k the immediate reward if the candidate decides to quit the game after answering correctly question k, i.e., if the candidate stops in a state s = (k + 1, l 1, l 2, l 3 ), and denote r k the immediate reward if the candidate fails in a state s = (k + 1, l 1, l 2, l 3 ). See table 2. After a decision is made, the process evolves to a new state. 4
6 If the candidate decides to stop or if she fails in a question, the game is over. If she decides to play and chooses the correct answer, there is a transition to another state t(s, a) = (k, l 1, l 2, l 3) S, where the question index k is equal to k + 1 and the lifeline indicators l i are: l i = { l i 1 l i if the candidate uses lifeline i in this question, otherwise. Answering correctly depends on certain probabilities for each question, being the same for all candidates. We further assume that the probabilities can be influenced by using lifelines, which we suppose to be helpful (i.e. to increase the probability of answering correctly). Denote p a s the probability of answering correctly if in state s S action a A(s) is chosen. Our analysis takes into account the possible skill of the participants. For that, we will divide the participants in four groups, namely A, B, C, D. Belonging to these groups means that the a priori probability p a s is modified according to a skill factor associated with each group. Mathematically speaking this is reflected in a multiplicative factor h G, G {A, B, C, D}, that modifies the probability in the following way: h G p a S, G {A, B, C, D}, where h A = 1, h B = 0.9, h C = 0.8, h D = 0.7. This means that the lower the skill the participant has, the smaller the probabilities of answering correctly are. One of the cornerstones in the resolution of the actual problem is to get a good estimation of the probabilities in the decision process. For a realistic estimation, one would need detailed data: for each question and for each possible combination of lifelines, there would have to be a certain number of candidates who answered correctly and failed, and this number would have to be high enough to estimate the probabilities. As mentioned above, the actual data are only available for approximately 40 games broadcasted on Spanish TV, and, of course, for most combinations of lifelines there are no observations, making not possible to give an estimation of the probabilities. Nevertheless, we had enough information to be able to estimate the probabilities of answering correctly without using any lifelines and using one single lifeline. Therefore, in order to solve the problem we make further assumptions. Let p k denote the probability of answering correctly without using any lifeline. We assume that there exists a multiplicative relationship between the probability of failing in a given state 5
7 using lifeline i and the probability of failure without lifelines. This relation is such that the probability of failing decreases by a fix factor c i, 0 < c i < 1, i = 1, 2, 3, or in other words: p i k = 1 (1 p k)c i k, (1) where p i k is the probability of answering correctly the question number k using the ith lifeline (both p k and p i k are known, for all k, i). We assume further that the combination of several lifelines modifies the original probability (1 p k ) in a multiplicative way, multiplying the different c constants. This simplification allows us to give a heuristic expression of the probabilities, that can be justified because we did not have enough data to give a valid estimation for each combination of lifelines. Under this assumption, we can use the information that we have about the candidates to estimate the probabilities of answering correctly with any feasible combination of lifelines. In the sequel we estimate the probabilities of answering correctly without using any lifelines and the constants c i k from the available data. For every question index k, we consider the candidates who did not use any lifelines and those who used only one lifeline. Then, for each of these groups of candidates, respectively, we consider the number of candidates who answered correctly this question and the number of candidates who failed. These probabilities are estimated by the relative frequencies observed in the data, and they are shown in table 3. Let p k denote the probability of answering correctly the kth question without using lifelines, p 1 k the probability of answering correctly using lifeline 1 (50:50 option), p2 k the probability of answering correctly using lifeline 2 (phone a friend) and p 3 k the probability of answering correctly using lifeline 3 (ask the audience). In table 3 get the probabilities of answering correctly (all the probabilities in %) 1. We use our model in the equation (1) to estimate the values of the c constants. Thus, for each question index k the factor c i k that modifies the probability when lifeline i is used is given by: c i k = 1 pi k 1 p k 1 original value 100% replaced by 99% 6
8 question index k p k p 1 k p 2 k p 3 k Table 3: Estimated probabilities of correct answers 3 Mathematical formulation In this section we present two different models. The first model is thought to maximize the expected reward and the second to maximize the probability of reaching a fixed goal. Both of them, apart from giving us expected reward and maximum probability, also give us optimal strategies to achieve their respective goals. 3.1 Model 1: expected reward Let p a s denote the probability of answering correctly if in state s S action a A(s) is chosen. Suppose that the probabilities p a s depend only on the question index and on the used lifelines. Let f(s) be the maximum expected reward that can be obtained starting at the state s. We can evaluate f(s) in the following way. The maximum expected reward from s will be on the maximum among the expected rewards that can be obtained according with the different courses of actions a A(s). At that point, we can either quit the game, thus ensuring r k 1, or go for the next question (assume indexed by k). In the latter case, if we choose an action a A(s) then we answer correctly with probability p a s and fail with probability (1 p a s). The 7
9 k c 1 k c 2 k c 3 k Table 4: Correction factors reward when failing is given by the prior reward ensured to the question k, i.e. rk 1. On the other hand, answering correctly question k produces a transition to the next question with the remaining lifelines. Denote by t(s, a) the transition function that gives the new state when action a is chosen in state s. Then, from that point on the expected reward is f(t(s, a)). In summary, the expected reward under action a is: p a sf(t(s, a)) + (1 p a s)r k 1. Hence, f(s) = max a A(s) {r k 1, p a sf(t(s, a)) + (1 p a s)r k 1}. In order to get the maximum expected reward we have to evaluate f(departing state). If the candidate starts in question number 1 with the three lifelines we have to compute f(1, 1, 1, 1). The values of f can be computed recursively by backward induction once we know the value of f at any feasible state of the terminal stage. These values are easily computed and they are shown in table 5. 8
10 State f(state) 15,1,1, ,0,0, ,0,1, ,0,1, ,1,1, ,1,0, ,1,0, ,0,0, Table 5: Departing state probabilities. Therefore, using backward induction starting with the data in the table above we obtain f(1, 1, 1, 1) and find optimal strategies. In this process we use the estimated probabilities and constants obtained in Section 2. All computations were performed with a MAPLE computer program. The solution obtained by the program is f(1, 1, 1, 1) = and an optimal strategy is shown in table Model 2: reaching a question In this section we address a different solution approach to our problem. We have seen in Section 3.1 the optimal strategy to be used if we want to maximize the expected reward and how much we win following it. Now we want to find the optimal strategy that we have to follow in order to maximize the probability of reaching and answering correctly a given question. Moreover, we also give the probability of doing that if we follow an optimal strategy. Let us define the new problem. Recall that a state s is defined as a fourdimensional vector, as before: s = (k, l 1, l 2, l 3 ). Let k, e = 1, 2,, 15, be a fix number. Our goal is to answer correctly the question number k. We denote by f(s) the maximum probability of getting and answering correctly the question number k, starting in state s. We evaluate f(s) in the following way. The maximum probability of reaching and answering correctly the question number k starting in state s is the maximum among the possible actions a A(s) of the probability of answering correctly the current question times the 9
11 Question index Strategy 1 No lifelines 2 No lifelines 3 No lifelines 4 No lifelines 5 Audience 6 No lifelines 7 No lifelines 8 No lifelines 9 50:50 Option 10 Phone 11 No lifelines 12 No lifelines 13 Stop expected reward Table 6: Solution of Model 1. maximum probability of getting our goal from the state t(a, s), a A(s), where t(a, s) is the transition state after choosing the action a in the state s if answering correctly. Then, we have: f(k, l 1, l 2, l 3 ) = max 0 g i l i g i Z, i {p k,g1,g 2,g 3 f(k + 1, l 1 g 1, l 2 g 2, l 3 g 3 )}, where p k,g1,g 2,g 3 is the probability of answering correctly the k th question using some lifelines. Indeed, i th lifeline is used if g i = 1, i = 1, 2, 3. The function f is a recursive functional, therefore to obtain its evaluation by backward induction we need its value at all the states in the terminal stage. Notice that the goal in this formulation is to reach the stage k. Thus, the probability of having reached the stage k if we are already at k + 1 is clearly 1. Hence, we have f(k + 1, l 1, l 2, l 3 ) = 1 l i {0, 1}, i = 1, 2, 3. Once we have the evaluation of the function at the terminal stage, the solution of this model is the evaluation of f(departing state). If we start from the first question and we have all the lifelines, the departing state is (1,1,1,1). But if we start at the third question and we only 10
12 Question index Goal: 5 Goal: 10 Goal: 13 Goal:15 1 No lifelines No lifelines No lifelines No lifelines 2 No lifelines No lifelines No lifelines No lifelines 3 50:50 No lifelines No lifelines No lifelines 4 Audience No lifelines No lifelines No lifelines 5 Phone No lifelines No lifelines No lifelines 6 Audience No lifelines No lifelines 7 No lifelines No lifelines No lifelines 8 No lifelines No lifelines No lifelines 9 50:50 Audience No lifelines 10 Phone No lifelines No lifelines 11 Phone Phone 12 50:50 No lifelines 13 No lifelines No lifelines 14 Audience 15 50:50 Probability Table 7: Optimal strategies for Model 2. have the 50:50 and the audience lifelines, the departing state would be (3,1,0,1). Anyway, the algorithm we propose solves that problem starting in any departing state and having as goal any level of the game. We use the estimated probabilities and constants c i, calculated as before, in a computer program in MAPLE to evaluate the function f and to find optimal strategies. In this model we do not have a unique solution but 15, because we may have fifteen possible goals: the fifteen questions we have in the game. For the sake of brevity we only show the solutions obtained if we start in state (1,1,1,1) and we want to reach and answer correctly the questions number 5, 10, 13 or 15. The optimal strategies and the probabilities of reaching and answering correctly the goals mentioned above are shown in table 7. The last row of the table above represents the probabilities of getting the proposed goal. 11
13 4 Further analysis of the game We have solved the problem in a static way, because of all the probabilities are determined a priori, that is, without the actual knowledge of each question. Actually the game is played changing the probabilities of answering correctly each time that the player faces the current question. For example, if in the fourth question, once she knows the actual question, she can estimate her probability of answering correctly that question, based on her own knowledge on the question. Then, changing that probability in our scheme, she evaluates again the function starting in the current state and keeping the others probabilities unchanged. This analysis means that the player modifies at each stage k the probability p k of answering correctly according with her own knowledge of the subject. This would be a realistic dynamic way to play the game. This feature has been incorporated in our computer code, so that, at each stage the player can change the probability of answering correctly the current question. Notice that this argument does not modify our recursive analysis of the problem. It only means that we allow to change the probability p k at each step of the analysis. 4.1 Simulation In order to illustrate our analysis of this game we perform a simulation of the process to check the behavior of the winning strategies proposed in our models. As we mentioned in Section 2 we classify the participants in four groups as follows: Players in group A have the original probabilities described previously. Probabilities of answering correctly in group B are the probabilities for group A multiplied by 0.9. Probabilities of answering correctly in group C are the probabilities for group A multiplied by 0.8. Probabilities of answering correctly in group D are the probabilities for group A multiplied by 0.7. In the following, we present two tables (table 8) with the strategy that one participant of each group (A, B, C and D) should take in order to maximize her expected reward (Model 1) and the strategy to maximize the probability of winning, at least, that maximum expected reward (Model 2). For example, the last row in the column of the participant A in Model 1 shows us the expected reward that she would get following the strategy described in that column, and the last row in Model 2 is the probability of winning at least that maximum 12
14 expected reward. To win at least euros we have to answer correctly question number 7. The other cases are analogous. The last row in both tables shows the maximum expected reward, if Model 1, or the probability of being successful if we follow the strategy described on Model 2. To finish this section we are going to show a simulation of Model 1 of the game played in its dynamic version. This is, we will assume that on each actual question the probability of answering correctly are modified once the concrete question is known. Let us assume that the contestant is now facing the question k th. She is deciding whether answering the question and how, or not depending on the degree of difficulty of the actual question. The model assumes that the probabilities of answering correctly the next questions, that is from k + 1 on, are the original ones estimated before. In table 9 the strategies using the lifeline of the 50:50, Phone and Audience are denoted by 50, P and A respectively. In order to simplify the simulation we assume that the probabilities of answering correctly can be: 1 if the contestant knows the right answer. 0.5 if the contestant doubts between two answers if she simply is sure that one of the answers is incorrect if she does not know anything about the answer and the four of them can be possible for her. The reader may notice that any kind of a priori probabilistic information, based on the knowledge of the actual player, can be incorporated into the model. This incorporation is done by computing posterior probabilities using Bayes rule. It is clear that the strategies change depending on the probabilities of the question that the contestant is facing, which have been chosen at random using different probability functions for each question number. The first number in each cell is the actual probability of answering correctly the corresponding question. As can be seen, depending on the simulated probability, the strategies can vary from stopping at the fifth until the twelfth question. 13
15 Skill A Skill B Question Model 1 Model 2 Model 1 Model 2 1 Without Without Without Without 2 Without Without Without Without 3 Without Without Without 50:50 4 Without Without Audience Audience 5 Without Phone Phone Phone 6 Audience 50:50 Without Stop 7 Without Audience Without 8 Without Stop Without 9 50:50 Without 10 Phone 50:50 11 Without Without 12 Without Without 13 Stop Stop E.R / Prob Skill C Skill D Question Model 1 Model 2 Model 1 Model 2 1 Without Without Without Without 2 Without Audience Without Without 3 50:50 50:50 50:50 50:50 4 Audience Phone Audience Audience 5 Phone Stop Phone Phone 6 Without Without Stop 7 Without Without 8 Without Stop 9 Stop E.R / Prob Table 8: Optimal solutions depending on the player s skill 14
16 Question P1 P2 P3 P4 P5 P6 1 1/NL 1/NL 0.5/50A 0.5/50A 0.5/50A 1/NL 2 0.5/50 0.5/P 1/NL 0.33/ P 1/NL 1/NL 3 1/NL 0.33/A 0.5/P 1/NL 1/NL 0.33/50 4 1/NL 0.5/50 0.5/NL 1/NL 0.5/P 1/NL 5 0.5/P 0.25/Stop 0.5/NL 0.33/NL 0.5/NL 1/NL 6 0.5/A 0.33/NL 0.5/NL 1/NL 0.5/A 7 0.5/NL 1/NL 0.5/NL 0.33/NL 1/NL 8 1/NL 0.5/NL 0.5/NL 1/NL 0.5/NL /Stop 0.33/Stop 0.33/Stop 0.25/Stop 1/NL /P /NL /Stop Table 9: Simulation References [1] Chlond M.J. (2001), The Travelling Space Telescope Problem, INFORMS Transactions on Education 2:1 (5860). [2] Cochran J.J. (2001), Who Wants To Be A Millionaire R : The Classroom Edition, INFORMS Transactions on Education 1:3 ( ). [3] Rump C.M. (2001), Who Wants to See a $Million Error?. A Neglected Educational Resource, INFORMS Transactions on Education 1:3 ( ). [4] Heyman D. and Sobel M. (1984), Stochastic Models in Operations Research. Vol 2, McGrawHill, New York. [5] Sniedovich M. (2003), A Neglected Educational Resource, INFORMS Transactions on Education 2:3, [6] Sniedovich M. (2002), Towers of Hanoi, INFORMS Transactions on Education 3:1 (3451). 15
17 [7] Sniedovich M. (2000), Counterfeit Coin Problem INFORMS Transactions on Education 3:2 (3241). [8] Tijms H.C. (1986), Stochastic modeling and analysis. A computational approach. WILEY, New York. 16
An approach to Calculus of Probabilities through real situations
MaMaEuSch Management Mathematics for European Schools http://www.mathematik.unikl.de/ mamaeusch An approach to Calculus of Probabilities through real situations Paula Lagares Barreiro Federico Perea RojasMarcos
More informationPopulation and sample. Sampling techniques
MaMaEuSch Management Mathematics for European Schools http://www.mathematik.unikl.de/ mamaeusch Population and sample. Sampling techniques Paula Lagares Barreiro Justo Puerto Albandoz MaMaEuSch Management
More informationConnection and costs distribution problems
MaMaEuSch Management Mathematics for European Schools http://www.mathematik.unikl.de/ mamaeusch Connection and costs distribution problems Federico Perea Justo Puerto MaMaEuSch Management Mathematics for
More informationVoting power in the European Union
Voting power in the European Union Federico Perea Justo Puerto MaMaEuSch Management Mathematics for European Schools 94342  CP  12001  DE  COMENIUS  C21 University of Seville This project has been
More informationOPTIMIZATION AND OPERATIONS RESEARCH Vol. IV  Markov Decision Processes  Ulrich Rieder
MARKOV DECISIO PROCESSES Ulrich Rieder University of Ulm, Germany Keywords: Markov decision problem, stochastic dynamic program, total reward criteria, average reward, optimal policy, optimality equation,
More informationReinforcement Learning
Reinforcement Learning LU 2  Markov Decision Problems and Dynamic Programming Dr. Martin Lauer AG Maschinelles Lernen und Natürlichsprachliche Systeme AlbertLudwigsUniversität Freiburg martin.lauer@kit.edu
More informationReinforcement Learning
Reinforcement Learning LU 2  Markov Decision Problems and Dynamic Programming Dr. Joschka Bödecker AG Maschinelles Lernen und Natürlichsprachliche Systeme AlbertLudwigsUniversität Freiburg jboedeck@informatik.unifreiburg.de
More information1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let
Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as
More informationOPTIMAL DESIGN OF A MULTITIER REWARD SCHEME. Amir Gandomi *, Saeed Zolfaghari **
OPTIMAL DESIGN OF A MULTITIER REWARD SCHEME Amir Gandomi *, Saeed Zolfaghari ** Department of Mechanical and Industrial Engineering, Ryerson University, Toronto, Ontario * Tel.: + 46 979 5000x7702, Email:
More informationBook Review of Rosenhouse, The Monty Hall Problem. Leslie Burkholder 1
Book Review of Rosenhouse, The Monty Hall Problem Leslie Burkholder 1 The Monty Hall Problem, Jason Rosenhouse, New York, Oxford University Press, 2009, xii, 195 pp, US $24.95, ISBN 9780195#67898 (Source
More informationSection 1.3 P 1 = 1 2. = 1 4 2 8. P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., = 1 2 4.
Difference Equations to Differential Equations Section. The Sum of a Sequence This section considers the problem of adding together the terms of a sequence. Of course, this is a problem only if more than
More informationIs it possible to beat the lottery system?
Is it possible to beat the lottery system? Michael Lydeamore The University of Adelaide Postgraduate Seminar, 2014 The story One day, while sitting at home (working hard)... The story Michael Lydeamore
More informationCreating a NL Texas Hold em Bot
Creating a NL Texas Hold em Bot Introduction Poker is an easy game to learn by very tough to master. One of the things that is hard to do is controlling emotions. Due to frustration, many have made the
More informationAn approach to Descriptive Statistics through real situations
MaMaEuSch Management Mathematics for European Schools http://www.mathematik.unikl.de/ mamaeusch An approach to Descriptive Statistics through real situations Paula Lagares Barreiro 1 Federico Perea RojasMarcos
More informationTetris: Experiments with the LP Approach to Approximate DP
Tetris: Experiments with the LP Approach to Approximate DP Vivek F. Farias Electrical Engineering Stanford University Stanford, CA 94403 vivekf@stanford.edu Benjamin Van Roy Management Science and Engineering
More informationNo Claim Bonus? by John D. Hey*
The Geneva Papers on Risk and Insurance, 10 (No 36, July 1985), 209228 No Claim Bonus? by John D. Hey* 1. Introduction No claim bonus schemes are a prominent feature of some insurance contracts, most
More informationSensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS
Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and
More information1. (a) Multiply by negative one to make the problem into a min: 170A 170B 172A 172B 172C Antonovics Foster Groves 80 88
Econ 172A, W2001: Final Examination, Possible Answers There were 480 possible points. The first question was worth 80 points (40,30,10); the second question was worth 60 points (10 points for the first
More informationAPPLICATION OF HIDDEN MARKOV CHAINS IN QUALITY CONTROL
INTERNATIONAL JOURNAL OF ELECTRONICS; MECHANICAL and MECHATRONICS ENGINEERING Vol.2 Num.4 pp.(353360) APPLICATION OF HIDDEN MARKOV CHAINS IN QUALITY CONTROL Hanife DEMIRALP 1, Ehsan MOGHIMIHADJI 2 1 Department
More informationPerformance Assessment Task Fair Game? Grade 7. Common Core State Standards Math  Content Standards
Performance Assessment Task Fair Game? Grade 7 This task challenges a student to use understanding of probabilities to represent the sample space for simple and compound events. A student must use information
More informationAbout the inverse football pool problem for 9 games 1
Seventh International Workshop on Optimal Codes and Related Topics September 61, 013, Albena, Bulgaria pp. 15133 About the inverse football pool problem for 9 games 1 Emil Kolev Tsonka Baicheva Institute
More informationA ProfitMaximizing Production Lot sizing Decision Model with Stochastic Demand
A ProfitMaximizing Production Lot sizing Decision Model with Stochastic Demand Kizito Paul Mubiru Department of Mechanical and Production Engineering Kyambogo University, Uganda Abstract  Demand uncertainty
More informationThe mathematical branch of probability has its
ACTIVITIES for students Matthew A. Carlton and Mary V. Mortlock Teaching Probability and Statistics through Game Shows The mathematical branch of probability has its origins in games and gambling. And
More informationWelcome to Harcourt Mega Math: The Number Games
Welcome to Harcourt Mega Math: The Number Games Harcourt Mega Math In The Number Games, students take on a math challenge in a lively insect stadium. Introduced by our host Penny and a number of sporting
More informationRISKy Business: An InDepth Look at the Game RISK
RISKy Business: An InDepth Look at the Game RISK Sharon Blatt Advisor: Dr. Crista Coles Elon University Elon, NC 744 slblatt@hotmail.com Introduction Games have been of interest to mathematicians for
More informationLinear Programming Notes VII Sensitivity Analysis
Linear Programming Notes VII Sensitivity Analysis 1 Introduction When you use a mathematical model to describe reality you must make approximations. The world is more complicated than the kinds of optimization
More informationCombinatorics: The Fine Art of Counting
Combinatorics: The Fine Art of Counting Week 7 Lecture Notes Discrete Probability Continued Note Binomial coefficients are written horizontally. The symbol ~ is used to mean approximately equal. The Bernoulli
More informationChapter 2. CASH FLOW Objectives: To calculate the values of cash flows using the standard methods.. To evaluate alternatives and make reasonable
Chapter 2 CASH FLOW Objectives: To calculate the values of cash flows using the standard methods To evaluate alternatives and make reasonable suggestions To simulate mathematical and real content situations
More informationP (A) = lim P (A) = N(A)/N,
1.1 Probability, Relative Frequency and Classical Definition. Probability is the study of random or nondeterministic experiments. Suppose an experiment can be repeated any number of times, so that we
More informationOptimization: Optimal Pricing with Elasticity
Optimization: Optimal Pricing with Elasticity Short Examples Series using Risk Simulator For more information please visit: www.realoptionsvaluation.com or contact us at: admin@realoptionsvaluation.com
More informationReading 13 : Finite State Automata and Regular Expressions
CS/Math 24: Introduction to Discrete Mathematics Fall 25 Reading 3 : Finite State Automata and Regular Expressions Instructors: Beck Hasti, Gautam Prakriya In this reading we study a mathematical model
More informationVideo Poker in South Carolina: A Mathematical Study
Video Poker in South Carolina: A Mathematical Study by Joel V. Brawley and Todd D. Mateer Since its debut in South Carolina in 1986, video poker has become a game of great popularity as well as a game
More informationMarkov Chains for the RISK Board Game Revisited. Introduction. The Markov Chain. Jason A. Osborne North Carolina State University Raleigh, NC 27695
Markov Chains for the RISK Board Game Revisited Jason A. Osborne North Carolina State University Raleigh, NC 27695 Introduction Probabilistic reasoning goes a long way in many popular board games. Abbott
More informationMODELING CUSTOMER RELATIONSHIPS AS MARKOV CHAINS. Journal of Interactive Marketing, 14(2), Spring 2000, 4355
MODELING CUSTOMER RELATIONSHIPS AS MARKOV CHAINS Phillip E. Pfeifer and Robert L. Carraway Darden School of Business 100 Darden Boulevard Charlottesville, VA 22903 Journal of Interactive Marketing, 14(2),
More information9.2 Summation Notation
9. Summation Notation 66 9. Summation Notation In the previous section, we introduced sequences and now we shall present notation and theorems concerning the sum of terms of a sequence. We begin with a
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationLesson 13: Games of Chance and Expected Value
Student Outcomes Students analyze simple games of chance. Students calculate expected payoff for simple games of chance. Students interpret expected payoff in context. esson Notes When students are presented
More information1 Limiting distribution for a Markov chain
Copyright c 2009 by Karl Sigman Limiting distribution for a Markov chain In these Lecture Notes, we shall study the limiting behavior of Markov chains as time n In particular, under suitable easytocheck
More informationDiscrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note 11
CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According
More informationProblem of the Month: Fair Games
Problem of the Month: The Problems of the Month (POM) are used in a variety of ways to promote problem solving and to foster the first standard of mathematical practice from the Common Core State Standards:
More informationMath 728 Lesson Plan
Math 728 Lesson Plan Tatsiana Maskalevich January 27, 2011 Topic: Probability involving sampling without replacement and dependent trials. Grade Level: 812 Objective: Compute the probability of winning
More informationRandomTurn Hex and Other Selection Games
RandomTurn Hex and Other Selection Games YuHan Lyu March 10, 2011 Abstract In this report, we summarize the results in [6]. We define a mathematical formulation for all selection games. When considering
More informationChapter 15 Introduction to Linear Programming
Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2014 WeiTa Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of
More informationChapter 1: The binomial asset pricing model
Chapter 1: The binomial asset pricing model Simone Calogero April 17, 2015 Contents 1 The binomial model 1 2 1+1 dimensional stock markets 4 3 Arbitrage portfolio 8 4 Implementation of the binomial model
More informationMarkov Chains, Stochastic Processes, and Advanced Matrix Decomposition
Markov Chains, Stochastic Processes, and Advanced Matrix Decomposition Jack Gilbert Copyright (c) 2014 Jack Gilbert. Permission is granted to copy, distribute and/or modify this document under the terms
More informationMassachusetts Institute of Technology
n (i) m m (ii) n m ( (iii) n n n n (iv) m m Massachusetts Institute of Technology 6.0/6.: Probabilistic Systems Analysis (Quiz Solutions Spring 009) Question Multiple Choice Questions: CLEARLY circle the
More informationA fairly quick tempo of solutions discussions can be kept during the arithmetic problems.
Distributivity and related number tricks Notes: No calculators are to be used Each group of exercises is preceded by a short discussion of the concepts involved and one or two examples to be worked out
More informationPsychology and Economics (Lecture 17)
Psychology and Economics (Lecture 17) Xavier Gabaix April 13, 2004 Vast body of experimental evidence, demonstrates that discount rates are higher in the shortrun than in the longrun. Consider a final
More information6.231 Dynamic Programming and Stochastic Control Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.231 Dynamic Programming and Stochastic Control Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 6.231
More informationChapter 17: Aggregation
Chapter 17: Aggregation 17.1: Introduction This is a technical chapter in the sense that we need the results contained in it for future work. It contains very little new economics and perhaps contains
More informationWhat is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
More informationLecture 5: Mixed strategies and expected payoffs
Lecture 5: Mixed strategies and expected payoffs As we have seen for example for the Matching pennies game or the RockPaperscissor game, sometimes game have no Nash equilibrium. Actually we will see
More informationMARK SCHEME. Mathematics Unit T5 Paper 2 (With calculator) Foundation Tier [GMT52] 3.00pm 4.00pm. General Certificate of Secondary Education 2014
General Certificate of Secondary Education 2014 Mathematics Unit T Paper 2 (With calculator) Foundation Tier [GMT2] friday 30 may 3.00pm 4.00pm MARK SCHEME 8802.01 F GCSE MATHEMATICS Introduction The mark
More informationECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE YUAN TIAN This synopsis is designed merely for keep a record of the materials covered in lectures. Please refer to your own lecture notes for all proofs.
More information3.2 Roulette and Markov Chains
238 CHAPTER 3. DISCRETE DYNAMICAL SYSTEMS WITH MANY VARIABLES 3.2 Roulette and Markov Chains In this section we will be discussing an application of systems of recursion equations called Markov Chains.
More informationBinomial lattice model for stock prices
Copyright c 2007 by Karl Sigman Binomial lattice model for stock prices Here we model the price of a stock in discrete time by a Markov chain of the recursive form S n+ S n Y n+, n 0, where the {Y i }
More informationProblem of the Month: Game Show
Problem of the Month: The Problems of the Month (POM) are used in a variety of ways to promote problem solving and to foster the first standard of mathematical practice from the Common Core State Standards:
More informationA simple algorithm with no simple verication
A simple algorithm with no simple verication Laszlo Csirmaz Central European University Abstract The correctness of a simple sorting algorithm is resented, which algorithm is \evidently wrong" at the rst
More informationThe Monty Hall Problem
Gateway to Exploring Mathematical Sciences (GEMS) 9 November, 2013 Lecture Notes The Monty Hall Problem Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George
More information11.2 POINT ESTIMATES AND CONFIDENCE INTERVALS
11.2 POINT ESTIMATES AND CONFIDENCE INTERVALS Point Estimates Suppose we want to estimate the proportion of Americans who approve of the president. In the previous section we took a random sample of size
More informationWORKED EXAMPLES 1 TOTAL PROBABILITY AND BAYES THEOREM
WORKED EXAMPLES 1 TOTAL PROBABILITY AND BAYES THEOREM EXAMPLE 1. A biased coin (with probability of obtaining a Head equal to p > 0) is tossed repeatedly and independently until the first head is observed.
More informationInteger Programming Formulation
Integer Programming Formulation 1 Integer Programming Introduction When we introduced linear programs in Chapter 1, we mentioned divisibility as one of the LP assumptions. Divisibility allowed us to consider
More informationMODELING CUSTOMER RELATIONSHIPS AS MARKOV CHAINS
AS MARKOV CHAINS Phillip E. Pfeifer Robert L. Carraway f PHILLIP E. PFEIFER AND ROBERT L. CARRAWAY are with the Darden School of Business, Charlottesville, Virginia. INTRODUCTION The lifetime value of
More informationProbability and Statistics
CHAPTER 2: RANDOM VARIABLES AND ASSOCIATED FUNCTIONS 2b  0 Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute  Systems and Modeling GIGA  Bioinformatics ULg kristel.vansteen@ulg.ac.be
More information6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation
6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation Daron Acemoglu and Asu Ozdaglar MIT November 2, 2009 1 Introduction Outline The problem of cooperation Finitelyrepeated prisoner s dilemma
More informationOptimal Solution Strategy for Games
Optimal Solution Strategy for Games Aman Pratap Singh, Student BTech Computer Science Department, Faculty Of Engineering And Technology Gurukul Kangri Vishvidayala Haridwar, India Abstract In order to
More informationMathematical Induction
Mathematical Induction (Handout March 8, 01) The Principle of Mathematical Induction provides a means to prove infinitely many statements all at once The principle is logical rather than strictly mathematical,
More informationSecretary Problems. October 21, 2010. José Soto SPAMS
Secretary Problems October 21, 2010 José Soto SPAMS A little history 50 s: Problem appeared. 60 s: Simple solutions: Lindley, Dynkin. 7080: Generalizations. It became a field. ( Toy problem for Theory
More informationTHE THREEHAT PROBLEM. 1. Introduction
THE THREEHAT PROBLEM BRIAN BENSON AND YANG WANG 1. Introduction Many classical puzzles involve hats. The general setting for these puzzles is a game in which several players are each given a hat to wear.
More informationLOOKING FOR A GOOD TIME TO BET
LOOKING FOR A GOOD TIME TO BET LAURENT SERLET Abstract. Suppose that the cards of a well shuffled deck of cards are turned up one after another. At any timebut once only you may bet that the next card
More informationGaming the Law of Large Numbers
Gaming the Law of Large Numbers Thomas Hoffman and Bart Snapp July 3, 2012 Many of us view mathematics as a rich and wonderfully elaborate game. In turn, games can be used to illustrate mathematical ideas.
More information4. Joint Distributions
Virtual Laboratories > 2. Distributions > 1 2 3 4 5 6 7 8 4. Joint Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an underlying sample space. Suppose
More informationAN ANALYSIS OF A WARLIKE CARD GAME. Introduction
AN ANALYSIS OF A WARLIKE CARD GAME BORIS ALEXEEV AND JACOB TSIMERMAN Abstract. In his book Mathematical MindBenders, Peter Winkler poses the following open problem, originally due to the first author:
More informationSolving Equations and Inequalities
Solving Equations and Inequalities 59 minutes 53 marks Page 1 of 21 Q1. (a) Solve a 2 = 9 Answer a =... (b) Solve = 5 Answer b =... (c) Solve 2c 3 = 11 Answer c =... (2) (Total 4 marks) Q2. In the magic
More informationQueueing Networks with Blocking  An Introduction 
Queueing Networks with Blocking  An Introduction  Jonatha ANSELMI anselmi@elet.polimi.it 5 maggio 006 Outline Blocking Blocking Mechanisms (BAS, BBS, RS) Approximate Analysis  MSS Basic Notation We
More informationRegular Languages and Finite Automata
Regular Languages and Finite Automata 1 Introduction Hing Leung Department of Computer Science New Mexico State University Sep 16, 2010 In 1943, McCulloch and Pitts [4] published a pioneering work on a
More informationSupplement to Call Centers with Delay Information: Models and Insights
Supplement to Call Centers with Delay Information: Models and Insights Oualid Jouini 1 Zeynep Akşin 2 Yves Dallery 1 1 Laboratoire Genie Industriel, Ecole Centrale Paris, Grande Voie des Vignes, 92290
More informationEvolutionary Game Theory
Evolutionary Game Theory James Holland Jones Department of Anthropology Stanford University December 1, 2008 1 The Evolutionarily Stable Strategy (ESS) An evolutionarily stable strategy (ESS) is a strategy
More informationSolution of Linear Systems
Chapter 3 Solution of Linear Systems In this chapter we study algorithms for possibly the most commonly occurring problem in scientific computing, the solution of linear systems of equations. We start
More informationLecture notes: singleagent dynamics 1
Lecture notes: singleagent dynamics 1 Singleagent dynamic optimization models In these lecture notes we consider specification and estimation of dynamic optimization models. Focus on singleagent models.
More informationWeek 5 Integral Polyhedra
Week 5 Integral Polyhedra We have seen some examples 1 of linear programming formulation that are integral, meaning that every basic feasible solution is an integral vector. This week we develop a theory
More informationWe can express this in decimal notation (in contrast to the underline notation we have been using) as follows: 9081 + 900b + 90c = 9001 + 100c + 10b
In this session, we ll learn how to solve problems related to place value. This is one of the fundamental concepts in arithmetic, something every elementary and middle school mathematics teacher should
More informationMincost flow problems and network simplex algorithm
Mincost flow problems and network simplex algorithm The particular structure of some LP problems can be sometimes used for the design of solution techniques more efficient than the simplex algorithm.
More informationGambling Systems and MultiplicationInvariant Measures
Gambling Systems and MultiplicationInvariant Measures by Jeffrey S. Rosenthal* and Peter O. Schwartz** (May 28, 997.. Introduction. This short paper describes a surprising connection between two previously
More informationSCHOOL OF MATHEMATICS MATHEMATICS FOR PART I ENGINEERING. Self Study Course
SCHOOL OF MATHEMATICS MATHEMATICS FOR PART I ENGINEERING Self Study Course MODULE 17 MATRICES II Module Topics 1. Inverse of matrix using cofactors 2. Sets of linear equations 3. Solution of sets of linear
More information1 Error in Euler s Method
1 Error in Euler s Method Experience with Euler s 1 method raises some interesting questions about numerical approximations for the solutions of differential equations. 1. What determines the amount of
More informationLET S MAKE A DEAL! ACTIVITY
LET S MAKE A DEAL! ACTIVITY NAME: DATE: SCENARIO: Suppose you are on the game show Let s Make A Deal where Monty Hall (the host) gives you a choice of three doors. Behind one door is a valuable prize.
More informationProbability Theory, Part 4: Estimating Probabilities from Finite Universes
8 Resampling: The New Statistics CHAPTER 8 Probability Theory, Part 4: Estimating Probabilities from Finite Universes Introduction Some Building Block Programs Problems in Finite Universes Summary Introduction
More informationBasic Probability Concepts
page 1 Chapter 1 Basic Probability Concepts 1.1 Sample and Event Spaces 1.1.1 Sample Space A probabilistic (or statistical) experiment has the following characteristics: (a) the set of all possible outcomes
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An ndimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0534405967. Systems of Linear Equations Definition. An ndimensional vector is a row or a column
More informationWEAK DOMINANCE: A MYSTERY CRACKED
WEAK DOMINANCE: A MYSTERY CRACKED JOHN HILLAS AND DOV SAMET Abstract. What strategy profiles can be played when it is common knowledge that weakly dominated strategies are not played? A comparison to the
More informationLab 11. Simulations. The Concept
Lab 11 Simulations In this lab you ll learn how to create simulations to provide approximate answers to probability questions. We ll make use of a particular kind of structure, called a box model, that
More informationStatistical Machine Translation: IBM Models 1 and 2
Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation
More informationIntroduction to Flocking {Stochastic Matrices}
Supelec EECI Graduate School in Control Introduction to Flocking {Stochastic Matrices} A. S. Morse Yale University Gif sur  Yvette May 21, 2012 CRAIG REYNOLDS  1987 BOIDS The Lion King CRAIG REYNOLDS
More informationA FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING
A FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING Sumit Goswami 1 and Mayank Singh Shishodia 2 1 Indian Institute of TechnologyKharagpur, Kharagpur, India sumit_13@yahoo.com 2 School of Computer
More informationTerm Project: Roulette
Term Project: Roulette DCY Student January 13, 2006 1. Introduction The roulette is a popular gambling game found in all major casinos. In contrast to many other gambling games such as black jack, poker,
More informationUnderstanding. Probability and LongTerm Expectations. Chapter 16. Copyright 2005 Brooks/Cole, a division of Thomson Learning, Inc.
Understanding Chapter 16 Probability and LongTerm Expectations Copyright 2005 Brooks/Cole, a division of Thomson Learning, Inc. Thought Question 1: Two very different queries about probability: a. If
More informationA Learning Based Method for SuperResolution of Low Resolution Images
A Learning Based Method for SuperResolution of Low Resolution Images Emre Ugur June 1, 2004 emre.ugur@ceng.metu.edu.tr Abstract The main objective of this project is the study of a learning based method
More informationSOLVING LINEAR SYSTEMS
SOLVING LINEAR SYSTEMS Linear systems Ax = b occur widely in applied mathematics They occur as direct formulations of real world problems; but more often, they occur as a part of the numerical analysis
More informationM2L1. Random Events and Probability Concept
M2L1 Random Events and Probability Concept 1. Introduction In this lecture, discussion on various basic properties of random variables and definitions of different terms used in probability theory and
More information