Decision Theory. 36.1 Rational prospecting



Similar documents
ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

MA 1125 Lecture 14 - Expected Values. Friday, February 28, Objectives: Introduce expected values.

= = 106.

A New Interpretation of Information Rate

Betting with the Kelly Criterion

Random Fibonacci-type Sequences in Online Gambling

Economics 1011a: Intermediate Microeconomics

Economics 1011a: Intermediate Microeconomics

In the situations that we will encounter, we may generally calculate the probability of an event

Lecture 13. Understanding Probability and Long-Term Expectations

THE POKIES: BEFORE YOU PRESS THE BUTTON, KNOW THE FACTS.

Week 5: Expected value and Betting systems

The Mathematics of Gambling

1 Interest rates, and risk-free investments

Probability and Expected Value

Pascal is here expressing a kind of skepticism about the ability of human reason to deliver an answer to this question.

Lecture 11 Uncertainty

The overall size of these chance errors is measured by their RMS HALF THE NUMBER OF TOSSES NUMBER OF HEADS MINUS NUMBER OF TOSSES

We never talked directly about the next two questions, but THINK about them they are related to everything we ve talked about during the past week:

Chapter 2 An Introduction to Forwards and Options

SOME ASPECTS OF GAMBLING WITH THE KELLY CRITERION. School of Mathematical Sciences. Monash University, Clayton, Victoria, Australia 3168

How to Win the Stock Market Game

2 HOW DOES TRADING WORK?. 4 HOW THE SYSTEM WORKS. 6 WHAT IF IT GOES WRONG?. 13 A RECAP. 15 BANKROLL. 16 A SUCCESSFUL TRADE EXAMPLE.

Practical Probability:

Decision Making under Uncertainty

3 Introduction to Assessing Risk

Arbs2U Is it right for me? 17 October 2014

National Sun Yat-Sen University CSE Course: Information Theory. Gambling And Entropy

Problem sets for BUEC 333 Part 1: Probability and Statistics

SCRATCHING THE SURFACE OF PROBABILITY. Robert J. Russell, University of Paisley, UK

Mathematical Expectation

Elementary Statistics and Inference. Elementary Statistics and Inference. 16 The Law of Averages (cont.) 22S:025 or 7P:025.

You Are What You Bet: Eliciting Risk Attitudes from Horse Races

Expected Value and the Game of Craps

LOOKING FOR A GOOD TIME TO BET

Gambling and Data Compression

1. Probabilities lie between 0 and 1 inclusive. 2. Certainty implies unity probability. 3. Probabilities of exclusive events add. 4. Bayes s rule.

! Insurance and Gambling

V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE

AMS 5 CHANCE VARIABILITY

The Binomial Distribution

How to bet and win: and why not to trust a winner. Niall MacKay. Department of Mathematics

LIFE INSURANCE AND PENSIONS by Peter Tryfos York University

Decision & Risk Analysis Lecture 6. Risk and Utility

Intelligent Agent for Playing Casino Card Games

Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80)

Wald s Identity. by Jeffery Hein. Dartmouth College, Math 100

Statistics and Random Variables. Math 425 Introduction to Probability Lecture 14. Finite valued Random Variables. Expectation defined

AP Physics 1 and 2 Lab Investigations

Moral Hazard. Itay Goldstein. Wharton School, University of Pennsylvania

Financial Markets. Itay Goldstein. Wharton School, University of Pennsylvania

Risk Aversion. Expected value as a criterion for making decisions makes sense provided that C H A P T E R Risk Attitude

THE UNIVERSITY OF MELBOURNE MELBOURNE BUSINESS SCHOOL. MANAGERIAL ECONOMICS Term First Mid-Term Solutions DR.

How To Increase Your Odds Of Winning Scratch-Off Lottery Tickets!

Lecture 6: Portfolios with Stock Options Steven Skiena. skiena

1 Introduction to Option Pricing

Annuities and Sinking Funds

(SEE IF YOU KNOW THE TRUTH ABOUT GAMBLING)

Ch. 13.2: Mathematical Expectation

Week 7 - Game Theory and Industrial Organisation

Options Pricing. This is sometimes referred to as the intrinsic value of the option.

Example: Find the expected value of the random variable X. X P(X)

International Statistical Institute, 56th Session, 2007: Phil Everson

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

"TWO DOLLARS TO SHOW ON NUMBER

Minesweeper as a Constraint Satisfaction Problem

6 Hedging Using Futures

Unit 19: Probability Models

3.2 Roulette and Markov Chains

Lecture 14. Chapter 7: Probability. Rule 1: Rule 2: Rule 3: Nancy Pfenning Stats 1000

1 Uncertainty and Preferences

Finance 350: Problem Set 6 Alternative Solutions

6.042/18.062J Mathematics for Computer Science. Expected Value I

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

Chapter 4 Lecture Notes

Gambling with Information Theory

Chicago Booth BUSINESS STATISTICS Final Exam Fall 2011

The Kelly criterion for spread bets

ECE302 Spring 2006 HW4 Solutions February 6,

Choice Under Uncertainty

The New Mexico Lottery

Chapter 7: Proportional Play and the Kelly Betting System

FIT TO BURST. BY Lester Cavanagh

STRIKE FORCE ROULETTE

Bonus Maths 2: Variable Bet Sizing in the Simplest Possible Game of Poker (JB)

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

Chapter 4 DECISION ANALYSIS

ECE302 Spring 2006 HW3 Solutions February 2,

Time Value of Money Dallas Brozik, Marshall University

Statistics 100A Homework 3 Solutions

Standard 12: The student will explain and evaluate the financial impact and consequences of gambling.

Review Horse Race Gambling and Side Information Dependent horse races and the entropy rate. Gambling. Besma Smida. ES250: Lecture 9.

arxiv: v1 [math.pr] 5 Dec 2011

Why is Insurance Good? An Example Jon Bakija, Williams College (Revised October 2013)

Transcription:

36 Decision Theory Decision theory is trivial, apart from computational details (just like playing chess!). You have a choice of various actions, a. The world may be in one of many states x; which one occurs may be influenced by your action. The world s state has a probability distribution P (x a). Finally, there is a utility function U(x, a) which specifies the payoff you receive when the world is in state x and you chose action a. The task of decision theory is to select the action that maximizes the expected utility, E[U a] = d K x U(x, a)p (x a). (36.1) That s all. The computational problem is to maximize E[U a] over a. [Pessimists may prefer to define a loss function L instead of a utility function U and minimize the expected loss.] Is there anything more to be said about decision theory? Well, in a real problem, the choice of an appropriate utility function may be quite difficult. Furthermore, when a sequence of actions is to be taken, with each action providing information about x, we have to take into account the effect that this anticipated information may have on our subsequent actions. The resulting mixture of forward probability and inverse probability computations in a decision problem is distinctive. In a realistic problem such as playing a board game, the tree of possible cogitations and actions that must be considered becomes enormous, and doing the right thing is not simple because the expected utility of an action cannot be computed exactly (Russell and Wefald, 1991; Baum and Smith, 1993; Baum and Smith, 1997). Let s explore an example. 36.1 Rational prospecting Suppose you have the task of choosing the site for a Tanzanite mine. Your final action will be to select the site from a list of N sites. The nth site has a net value called the return x n which is initially unknown, and will be found out exactly only after site n has been chosen. [x n equals the revenue earned from selling the Tanzanite from that site, minus the costs of buying the site, paying the staff, and so forth.] At the outset, the return x n has a probability distribution P (x n ), based on the information already available. Before you take your final action you have the opportunity to do some prospecting. Prospecting at the nth site has a cost c n and yields data d n which reduce the uncertainty about x n. [We ll assume that the returns of 453

454 36 Decision Theory the N sites are unrelated to each other, and that prospecting at one site only yields information about that site and doesn t affect the return from that site.] Your decision problem is: given the initial probability distributions P (x 1 ), P (x 2 ),..., P (x N ), first, decide whether to prospect, and at which sites; then, in the light of your prospecting results, choose which site to mine. For simplicity, let s make everything in the problem Gaussian and focus on the question of whether to prospect once or not. We ll assume our utility function is linear in x n ; we wish to maximize our expected return. The utility function is U = x na, (36.2) if no prospecting is done, where n a is the chosen action site; and if prospecting is done the utility is U = c np + x na, (36.3) where n p is the site at which prospecting took place. The prior distribution of the return of site n is P (x n ) = Normal(x n ; µ n, σ 2 n). (36.4) If you prospect at site n, the datum d n is a noisy version of x n : P (d n x n ) = Normal(d n ; x n, σ 2 ). (36.5) The notation P (y) = Normal(y; µ, σ 2 ) indicates that y has Gaussian distribution with mean µ and variance σ 2. Exercise 36.1. [2 ] Given these assumptions, show that the prior probability distribution of d n is P (d n ) = Normal(d n ; µ n, σ 2 +σ 2 n) (36.6) (mnemonic: when independent variables add, variances add), and that the posterior distribution of x n given d n is ( ) P (x n d n ) = Normal x n ; µ n, σn 2 (36.7) where µ n = d n/σ 2 + µ n /σ 2 n 1/σ 2 + 1/σ 2 n and 1 σn 2 = 1 σ 2 + 1 σn 2 (mnemonic: when Gaussians multiply, precisions add). (36.8) To start with let s evaluate the expected utility if we do no prospecting (i.e., choose the site immediately); then we ll evaluate the expected utility if we first prospect at one site and then make our choice. From these two results we will be able to decide whether to prospect once or zero times, and, if we prospect once, at which site. So, first we consider the expected utility without any prospecting. Exercise 36.2. [2 ] Show that the optimal action, assuming no prospecting, is to select the site with biggest mean n a = argmax n and the expected utility of this action is µ n, (36.9) E[U optimal n] = max n µ n. (36.10) [If your intuition says surely the optimal decision should take into account the different uncertainties σ n too?, the answer to this question is reasonable if so, then the utility function should be nonlinear in x.]

36.2: Further reading 455 Now the exciting bit. Should we prospect? Once we have prospected at site n p, we will choose the site using the decision rule (36.9) with the value of mean µ np replaced by the updated value µ n given by (36.8). What makes the problem exciting is that we don t yet know the value of d n, so we don t know what our action n a will be; indeed the whole value of doing the prospecting comes from the fact that the outcome d n may alter the action from the one that we would have taken in the absence of the experimental information. From the expression for the new mean in terms of d n (36.8), and the known variance of d n (36.6) we can compute the probability distribution of the key quantity, µ n, and can work out the expected utility by integrating over all possible outcomes and their associated actions. Exercise 36.3. [2 ] Show that the probability distribution of the new mean µ n (36.8) is Gaussian with mean µ n and variance σ 2 n s 2 σn 2 σ 2 + σn 2. (36.11) Consider prospecting at site n. Let the biggest mean of the other sites be µ 1. When we obtain the new value of the mean, µ n, we will choose site n and get an expected return of µ n if µ n > µ 1, and we will choose site 1 and get an expected return of µ 1 if µ n < µ 1. So the expected utility of prospecting at site n, then picking the best site, is E[U prospect at n] = c n + P (µ n < µ 1) µ 1 + dµ n µ n Normal(µ n ; µ n, s 2 ). µ 1 (36.12) The difference in utility between prospecting and not prospecting is the quantity of interest, and it depends on what we would have done without prospecting; and that depends on whether µ 1 is bigger than µ n. { µ1 if µ E[U no prospecting] = 1 µ n (36.13) µ n if µ 1 µ n. So E[U prospect at n] E[U no prospecting] c n + dµ n (µ n µ 1 ) Normal(µ n; µ n, s 2 ) if µ 1 µ n = µ 1 µ1 (36.14) c n + dµ n (µ 1 µ n ) Normal(µ n ; µ n, s 2 ) if µ 1 µ n. We can plot the change in expected utility due to prospecting (omitting c n ) as a function of the difference (µ n µ 1 ) (horizontal axis) and the initial standard deviation σ n (vertical axis). In the figure the noise variance is σ 2 = 1. 0-6 -4-2 0 2 4 6 (µ n µ 1 ) σ n 3.5 3 2.5 2 1.5 1 0.5 Figure 36.1. Contour plot of the gain in expected utility due to prospecting. The contours are equally spaced from 0.1 to 1.2 in steps of 0.1. To decide whether it is worth prospecting at site n, find the contour equal to c n (the cost of prospecting); all points [(µ n µ 1 ), σ n ] above that contour are worthwhile. 36.2 Further reading If the world in which we act is a little more complicated than the prospecting problem for example, if multiple iterations of prospecting are possible, and the cost of prospecting is uncertain then finding the optimal balance between exploration and exploitation becomes a much harder computational problem. Reinforcement learning addresses approximate methods for this problem (Sutton and Barto, 1998).

456 36 Decision Theory 36.3 Further exercises Exercise 36.4. [2 ] The four doors problem. A new game show uses rules similar to those of the three doors (exercise 3.8 (p.57)), but there are four doors, and the host explains: First you will point to one of the doors, and then I will open one of the other doors, guaranteeing to choose a non-winner. Then you decide whether to stick with your original pick or switch to one of the remaining doors. Then I will open another non-winner (but not the current pick). You will then make your final decision by sticking with the door picked on the previous decision or by switching to the only other remaining door. What is the optimal strategy? Should you switch on the first opportunity? Should you switch on the second opportunity? Exercise 36.5. [3 ] One of the challenges of decision theory is figuring out exactly what the utility function is. The utility of money, for example, is notoriously nonlinear for most people. In fact, the behaviour of many people cannot be captured by a coherent utility function, as illustrated by the Allias paradox, which runs as follows. Which of these choices do you find most attractive? A. 1 million guaranteed. B. 89% chance of 1 million; 10% chance of 2.5 million; 1% chance of nothing. Now consider these choices: C. 89% chance of nothing; 11% chance of 1 million. D. 90% chance of nothing; 10% chance of 2.5 million. Many people prefer A to B, and, at the same time, D to C. Prove that these preferences are inconsistent with any utility function U(x) for money. Exercise 36.6. [4 ] Optimal stopping. A large queue of N potential partners is waiting at your door, all asking to marry you. They have arrived in random order. As you meet each partner, you have to decide on the spot, based on the information so far, whether to marry them or say no. Each potential partner has a desirability d n, which you find out if and when you meet them. You must marry one of them, but you are not allowed to go back to anyone you have said no to. There are several ways to define the precise problem. (a) Assuming your aim is to maximize the desirability d n, i.e., your utility function is dˆn, where ˆn is the partner selected, what strategy should you use? (b) Assuming you wish very much to marry the most desirable person (i.e., your utility function is 1 if you achieve that, and zero otherwise); what strategy should you use?

36.3: Further exercises 457 (c) Assuming you wish very much to marry the most desirable person, and that your strategy will be strategy M : Strategy M Meet the first M partners and say no to all of them. Memorize the maximum desirability d max among them. Then meet the others in sequence, waiting until a partner with d n > d max comes along, and marry them. If none more desirable comes along, marry the final Nth partner (and feel miserable). what is the optimal value of M? Exercise 36.7. [3 ] Regret as an objective function? The preceding exercise (parts b and c) involved a utility function based on regret. If one married the tenth most desirable candidate, the utility function asserts that one would feel regret for having not chosen the most desirable. Many people working in learning theory and decision theory use minimizing the maximal possible regret as an objective function, but does this make sense? Imagine that Fred has bought a lottery ticket, and offers to sell it to you before it s known whether the ticket is a winner. For simplicity say the probability that the ticket is a winner is 1/100, and if it is a winner, it is worth 10. Fred offers to sell you the ticket for 1. Do you buy it? The possible actions are buy and don t buy. The utilities of the four possible action outcome pairs are shown in table 36.2. I have assumed that the utility of small amounts of money for you is linear. If you don t buy the ticket then the utility is zero regardless of whether the ticket proves to be a winner. If you do buy the ticket you either end up losing one pound (with probability 99/100) or gaining nine (with probability 1/100). In the minimax regret community, actions are chosen to minimize the maximum possible regret. The four possible regret outcomes are shown in table 36.3. If you buy the ticket and it doesn t win, you have a regret of 1, because if you had not bought it you would have been 1 better off. If you do not buy the ticket and it wins, you have a regret of 9, because if you had bought it you would have been 9 better off. The action that minimizes the maximum possible regret is thus to buy the ticket. Discuss whether this use of regret to choose actions can be philosophically justified. The above problem can be turned into an investment portfolio decision problem by imagining that you have been given one pound to invest in two possible funds for one day: Fred s lottery fund, and the cash fund. If you put f 1 into Fred s lottery fund, Fred promises to return 9f 1 to you if the lottery ticket is a winner, and otherwise nothing. The remaining f 0 (with f 0 = 1 f 1 ) is kept as cash. What is the best investment? Show that the minimax regret community will invest f 1 = 9/10 of their money in the high risk, high return lottery fund, and only f 0 = 1/10 in cash. Can this investment method be justified? Outcome Action Buy Don t buy No win 1 0 Wins +9 0 Table 36.2. Utility in the lottery ticket problem. Outcome Action Buy Don t buy No win 1 0 Wins 0 9 Table 36.3. Regret in the lottery ticket problem. Exercise 36.8. [3 ] Gambling oddities (from Cover and Thomas (1991)). A horse race involving I horses occurs repeatedly, and you are obliged to bet all your money each time. Your bet at time t can be represented by

458 36 Decision Theory a normalized probability vector b multiplied by your money m(t). The odds offered by the bookies are such that if horse i wins then your return is m(t+1) = b i o i m(t). Assuming the bookies odds are fair, that is, i 1 o i = 1, (36.15) and assuming that the probability that horse i wins is p i, work out the optimal betting strategy if your aim is Cover s aim, namely, to maximize the expected value of log m(t ). Show that the optimal strategy sets b equal to p, independent of the bookies odds o. Show that when this strategy is used, the money is expected to grow exponentially as: where W = i p i log b i o i. 2 nw (b,p) (36.16) If you only bet once, is the optimal strategy any different? Do you think this optimal strategy makes sense? Do you think that it s optimal, in common language, to ignore the bookies odds? What can you conclude about Cover s aim? Exercise 36.9. [3 ] Two ordinary dice are thrown repeatedly; the outcome of each throw is the sum of the two numbers. Joe Shark, who says that 6 and 8 are his lucky numbers, bets even money that a 6 will be thrown before the first 7 is thrown. If you were a gambler, would you take the bet? What is your probability of winning? Joe then bets even money that an 8 will be thrown before the first 7 is thrown. Would you take the bet? Having gained your confidence, Joe suggests combining the two bets into a single bet: he bets a larger sum, still at even odds, that an 8 and a 6 will be thrown before two 7s have been thrown. Would you take the bet? What is your probability of winning?