1 Gambler s Ruin Problem



Similar documents
6.042/18.062J Mathematics for Computer Science December 12, 2006 Tom Leighton and Ronitt Rubinfeld. Random Walks

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

Stat 134 Fall 2011: Gambler s ruin

Pythagorean Triples and Rational Points on the Unit Circle

Binomial lattice model for stock prices

Price Elasticity of Demand MATH 104 and MATH 184 Mark Mac Lean (with assistance from Patrick Chan) 2011W

As we have seen, there is a close connection between Legendre symbols of the form

FREQUENCIES OF SUCCESSIVE PAIRS OF PRIME RESIDUES

SOME PROPERTIES OF EXTENSIONS OF SMALL DEGREE OVER Q. 1. Quadratic Extensions

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n

The Online Freeze-tag Problem

Stochastic Derivation of an Integral Equation for Probability Generating Functions

Section 1.3 P 1 = 1 2. = P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., =

1 Interest rates, and risk-free investments

POISSON PROCESSES. Chapter Introduction Arrival processes

PRIME NUMBERS AND THE RIEMANN HYPOTHESIS

Week 4: Gambler s ruin and bold play

MARTINGALES AND GAMBLING Louis H. Y. Chen Department of Mathematics National University of Singapore

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

LECTURE 4. Last time: Lecture outline

Wald s Identity. by Jeffery Hein. Dartmouth College, Math 100

Markov Chains. Chapter Stochastic Processes

Random access protocols for channel access. Markov chains and their stability. Laurent Massoulié.

Lecture 13: Martingales

Point Location. Preprocess a planar, polygonal subdivision for point location queries. p = (18, 11)

Basic Proof Techniques

Lectures on Stochastic Processes. William G. Faris

Math 115 Spring 2011 Written Homework 5 Solutions

Number Theory Naoki Sato

Sample Induction Proofs

5.1 Radical Notation and Rational Exponents

Risk in Revenue Management and Dynamic Pricing

arxiv: v1 [math.pr] 5 Dec 2011

Introduction to NP-Completeness Written and copyright c by Jie Wang 1

TRANSCENDENTAL NUMBERS

10.2 Series and Convergence

An important observation in supply chain management, known as the bullwhip effect,

Stochastic Processes and Advanced Mathematical Finance. Ruin Probabilities

Modeling and Performance Evaluation of Computer Systems Security Operation 1

1 Portfolio mean and variance

25 Integers: Addition and Subtraction

6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation

Random variables, probability distributions, binomial random variable

Index Numbers OPTIONAL - II Mathematics for Commerce, Economics and Business INDEX NUMBERS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Section 4.2: The Division Algorithm and Greatest Common Divisors

Multiperiod Portfolio Optimization with General Transaction Costs

Large firms and heterogeneity: the structure of trade and industry under oligopoly

Maths Workshop for Parents 2. Fractions and Algebra

Jena Research Papers in Business and Economics

WRITING PROOFS. Christopher Heil Georgia Institute of Technology

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

Solutions for Practice problems on proofs

M/M/1 and M/M/m Queueing Systems

Stochastic Processes and Advanced Mathematical Finance. Duration of the Gambler s Ruin

Goal Problems in Gambling and Game Theory. Bill Sudderth. School of Statistics University of Minnesota

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

United Arab Emirates University College of Sciences Department of Mathematical Sciences HOMEWORK 1 SOLUTION. Section 10.1 Vectors in the Plane

Frequentist vs. Bayesian Statistics

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

How to Gamble If You Must

F inding the optimal, or value-maximizing, capital

Properties of sequences Since a sequence is a special kind of function it has analogous properties to functions:

On Multicast Capacity and Delay in Cognitive Radio Mobile Ad-hoc Networks

26 Integers: Multiplication, Division, and Order

Essentials of Stochastic Processes

Mathematical Induction

Automatic Search for Correlated Alarms

A Little Set Theory (Never Hurt Anybody)

Fundamentele Informatica II

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

6.4 Normal Distribution

1 Error in Euler s Method

Cartesian Products and Relations

Continued Fractions and the Euclidean Algorithm

3.2 Roulette and Markov Chains

Factoring of Prime Ideals in Galois Extensions

Probability Generating Functions

Lecture 2. Marginal Functions, Average Functions, Elasticity, the Marginal Principle, and Constrained Optimization

Software Model Checking: Theory and Practice

Lies My Calculator and Computer Told Me

Master s Theory Exam Spring 2006

CONTINUED FRACTIONS AND PELL S EQUATION. Contents 1. Continued Fractions 1 2. Solution to Pell s Equation 9 References 12

8 Primes and Modular Arithmetic

Coin ToGa: A Coin-Tossing Game

DIVISIBILITY AND GREATEST COMMON DIVISORS

Monitoring Frequency of Change By Li Qin

X How to Schedule a Cascade in an Arbitrary Graph

Stationary random graphs on Z with prescribed iid degrees and finite mean connections

Accentuate the Negative: Homework Examples from ACE

Number Theory Naoki Sato

Let s explore the content and skills assessed by Heart of Algebra questions.

EECS 122: Introduction to Communication Networks Homework 3 Solutions

Math 120 Final Exam Practice Problems, Form: A

Transcription:

Coyright c 2009 by Karl Sigman 1 Gambler s Ruin Problem Let N 2 be an integer and let 1 i N 1. Consider a gambler who starts with an initial fortune of $i and then on each successive gamble either wins $1 or loses $1 indeendent of the ast with robabilities and q = 1 resectively. Let X n denote the total fortune after the n th gamble. The gambler s objective is to reach a total fortune of $N, without first getting ruined (running out of money). If the gambler succeeds, then the gambler is said to win the game. In any case, the gambler stos laying after winning or getting ruined, whichever haens first. {X n } yields a Markov chain (MC) on the state sace S = {0, 1,..., N}. The transition robabilities are given by P i,i+1 =, P i,i 1 = q, 0 < i < N, and both 0 and N are absorbing states, P 00 = P NN = 1. 1 For examle, when N = 4 the transition matrix is given by 1 0 0 0 0 q 0 0 0 P = 0 q 0 0. 0 0 q 0 0 0 0 0 1 While the game roceeds, this MC forms a simle random walk X n = i + 1 + + n, n 1, X 0 = i, where { n } forms an i.i.d. sequence of r.v.s. distributed as P ( = 1) =, P ( = 1) = q = 1, and reresents the earnings on the successive gambles. Since the game stos when either X n = 0 or X n = N, let τ i = min{n 0 : X n {0, N} X 0 = i}, denote the time at which the game stos when X 0 = i. If X τi = N, then the gambler wins, if X τi = 0, then the gambler is ruined. Let P i (N) = P (X τi = N) denote the robability that the gambler wins when X 0 = i. P i (N) denotes the robability that the gambler, starting initially with $i, reaches a total fortune of N before ruin; 1 P i (N) is thus the corresonding robably of ruin Clearly P 0 (N) = 0 and P N (N) = 1 by definition, and we next roceed to comute P i (N), 1 i N 1. Proosition 1.1 (Gambler s Ruin Problem) 1 ( q )i P i (N) = 1 ( q, if q; )N (1) i N, if = q = 0.5. 1 There are three communication classes: C 1 = {0}, C 2 = {1,..., N 1}, C 3 = {N}. C 1 and C 3 are recurrent whereas C 2 is transient. 1

Proof : For our derivation, we let P i = P i (N), that is, we suress the deendence on N for ease of notation. The key idea is to condition on the outcome of the first gamble, 1 = 1 or 1 = 1, yielding P i = P i+1 + qp i 1. (2) The derivation of this recursion is as follows: If 1 = 1, then the gambler s total fortune increases to X 1 = i+1 and so by the Markov roerty the gambler will now win with robability P i+1. Similarly, if 1 = 1, then the gambler s fortune decreases to X 1 = i 1 and so by the Markov roerty the gambler will now win with robability P i 1. The robabilities corresonding to the two outcomes are and q yielding (2). Since + q = 1, (2) can be re-written as P i + qp i = P i+1 + qp i 1, yielding P i+1 P i = q (P i P i 1 ). In articular, P 2 P 1 = (q/)(p 1 P 0 ) = (q/)p 1 (since P 0 = 0), so that P 3 P 2 = (q/)(p 2 P 1 ) = (q/) 2 P 1, and more generally Thus P i+1 P i = ( q )i P 1, 0 < i < N. P i+1 P 1 = = i (P k+1 P k ) k=1 i k=1 ( q )k P 1, yielding i P i+1 = P 1 + P 1 ( q i )k = P 1 ( q )k = k=1 k=0 1 ( q P )i+1 1 1 ( q ), if q; P 1 (i + 1), if = q = 0.5. (3) (Here we are using the geometric series equation i n=0 a i = 1 ai+1 1 a, for any number a and any integer i 1.) Choosing i = N 1 and using the fact that P N = 1 yields 1 ( q 1 = P N = P )N 1 1 ( q, if q; ) P 1 N, if = q = 0.5, from which we conclude that P 1 = 1 q 1 ( q )N, if q; 1 N, if = q = 0.5, 2

thus obtaining from (3) (after algebra) the solution 1 ( q )i P i = 1 ( q, if q; )N (4) i N, if = q = 0.5. 1.1 Becoming infinitely rich or getting ruined In the formula (1), it is of interest to see what haens as N ; denote this by P i ( ) = lim N P i (N). This limiting quantity denotes the robability that the gambler, if allowed to lay forever unless ruined, will in fact never get ruined and instead will obtain an infinitely large fortune. Proosition 1.2 Define P i ( ) = lim N P i (N). If > 0.5, then P i ( ) = 1 ( q )i > 0. (5) If 0.50, then P i ( ) = 0. (6) Thus, unless the gambles are strictly better than fair ( > 0.5), ruin is certain. Proof : If > 0.5, then q < 1; hence in the denominator of (1), ( q )N 0 yielding the result. If < 0.50, then q > 1; hence in the the denominator of (1), ( q )N yielding the result. Finally, if = 0.5, then i (N) = i/n 0. Examles 1. John starts with $2, and = 0.6: What is the robability that John obtains a fortune of N = 4 without going broke? i = 2, N = 4 and q = 1 = 0.4, so q/ = 2/3, and we want P 2 (4) = 1 (2/3)2 1 (2/3) 4 = 0.91 2. What is the robability that John will become infinitely rich? P 2 ( ) = 1 (2/3) 2 = 5/9 = 0.56 3. If John instead started with i = $1, what is the robability that he would go broke? The robability he becomes infinitely rich is P 1 ( ) = 1 (q/) = 1/3, so the robability of ruin is 1 P 1 ( ) = 2/3. 3

1.2 Alications Risk insurance business Consider an insurance comany that earns $1 er day (from interest), but on each day, indeendent of the ast, might suffer a claim against it for the amount $2 with robability q = 1. Whenever such a claim is suffered, $2 is removed from the reserve of money. Thus on the n th day, the net income for that day is exactly n as in the gamblers ruin roblem: 1 with robability, 1 with robability q. If the insurance comany starts off initially with a reserve of $i 1, then what is the robability it will eventually get ruined (run out of money)? The answer is given by (5) and (??): If > 0.5 then the robability is given by ( q )i > 0, whereas if 0.5 ruin will always ocurr. This makes intuitive sense because if > 0.5, then the average net income er day is E( ) = q > 0, whereas if 0.5, then the average net income er day is E( ) = q 0. So the comany can not exect to stay in business unless earning (on average) more than is taken away by claims. 1.3 Random walk hitting robabilities Let a > 0 and b > 0 be integers, and let R n = 1 + + n, n 1, R 0 = 0 denote a simle random walk initially at the origin. Let (a) = P ({R n } hits level a before hitting level b). By letting i = b, and N = a + b, we can equivalently imagine a gambler who starts with i = b and wishes to reach N = a + b before going broke. So we can comute (a) by casting the roblem into the framework of the gambler s ruin roblem: (a) = P i (N) where N = a + b, i = b. Thus Examles 1 ( q )b (a) = 1 ( q, if q; )a+b (7) b a+b, if = q = 0.5. 1. Ellen bought a share of stock for $10, and it is believed that the stock rice moves (day by day) as a simle random walk with = 0.55. What is the robability that Ellen s stock reaches the high value of $15 before the low value of $5? We want the robability that the stock goes u by 5 before going down by 5. This is equivalent to starting the random walk at 0 with a = 5 and b = 5, and comuting (a). (a) = 1 ( q )b 1 (0.82)5 1 ( q = = 0.73 )a+b 1 (0.82) 10 2. What is the robability that Ellen will become infinitely rich? 4

Here we equivalently want to know the robability that a gambler starting with i = 10 becomes infinitely rich before going broke. Just like Examle 2 on Page 3: 1 (q/) i = 1 (0.82) 10 1 0.14 = 0.86. 1.4 Maximums and minimums of the simle random walk Formula (7) can immediately be used for comuting the robability that the simle random walk {R n }, starting initially at R 0 = 0, will ever hit level a, for any given ositive integer a 1: Kee a fixed while taking the limit as b in (7). The result deends on wether < 0.50 or 0.50. A little thought reveals that we can state this roblem as comuting the tail P (M a), a 0, where M def = max{r n : n 0} is the all-time maximum of the random walk; a non-negative random variable, because {M a} = {R n = a, for some n 1}. Proosition 1.3 Let M def = max{r n : n 0} for the simle random walk starting initially at the origin (R 0 = 0). 1. When < 0.50, P (M a) = (/q) a, a 0; M has a geometric distribution with success robability 1 (/q): P (M = k) = (/q) k (1 (/q)), k 0. In this case, the random walk drifts down to, w1, but before doing so reaches the finite maximum M. 2. If 0.50, then P (M a) = 1, a 0: P (M = ) = 1; the random walk will, with robability 1, reach any ositive integer a no matter how large. Proof : Taking the limit in (7) as b yields the result by considering the two cases < 0.5 or 0.5: If < 0.5, then (q/) > 1 and so both (q/) b and (q/) a+b tend to as b. But before taking the limit, multily both numerator and denominator by (q/) b = (/q) b, yielding (a) = (/q)b 1 (/q) b (q/) a. Since (/q) b 0 as b, the result follows. If > 0.5, then (q/) < 1 and so both (q/) b and (q/) a+b tend to 0 as b yielding the limit in (7) as 1. If = 0.5, then (a) = b/(b + a) 1 as b. If < 0.5, then E( ) < 0, and if > 0.5, then E( ) > 0; so Proosition 1.3 is consistent with the fact that any random walk with E( ) < 0 (called the negative drift case) satisfies lim n R n =, w1, and any random walk with E( ) > 0 ( called the ositive drift case) satisfies lim n R n = +, w1. 2 But furthermore we learn that when < 0.5, although w1 the chain drifts off to, it first reaches a finite maximum M before doing so, and this rv M has a geometric distribution. 2 R From the strong law of large numbers, lim n n n + if E( ) > 0. = E( ), w1, so R n ne( ) if E( ) < 0 and 5

Finally Proosition 1.3 also offers us a roof that when = 0.5, the symmetric case, the random walk will w1 hit any ositive value, P (M a) = 1. By symmetry, we also obtain analogous results for the minimum, : Corollary 1.1 Let m def = min{r n : n 0} for the simle random walk starting initially at the origin (R 0 = 0). 1. If > 0.5, then P (m b) = (q/) b, b 0. In this case, the random walk drifts u to +, but before doing so dros down to a finite minimum m 0. Taking absolute values of m makes it non-negative and so we can exress this result as P ( m b) = (q/) b, b 0; m has a geometric distribution with success robability 1 (q/): P ( m = k) = (q/) k (1 (q/)), k 0. 2. If 0.50, then P (m b) = 1, b 0: P (m = ) = 1; the random walk will, with robability 1, reach any negative integer a no matter how small. Note that when < 0.5, P (M = 0) = 1 (/q) > 0. This is because it is ossible that the random walk will never enter the ositive axis before drifting off to ; with ositive robability R n 0, n 0. Similarly, if > 0.5, then P (m = 0) = 1 (q/) > 0; with ositive robability R n 0, n 0. Recurrence of the simle symmetric random walk Combining the results for both M and m in the revious section when = 0.5, we have Proosition 1.4 The simle symmetric ( = 0.50) random walk, starting at the origin, will w1 eventually hit any integer a, ositive or negative. In fact it will hit any given integer a infinitely often, always returning yet again after leaving; it is a recurrent Markov chain. Proof : The first statement follows directly from Proosition 1.3 and Corollary 1.1; P (M = ) = 1 = P (m = ) = 1. For the second statement we argue as follows: Using the first statement, we know that the simle symmetric random walk starting at 0 will hit 1 eventually. But when it does, we can use the same result to conclude that it will go back and hit 0 eventually after that, because that is stochastically equivalent to starting at R 0 = 0 and waiting for the chain to hit 1, which also will haen eventually. But then yet again it must hit 1 again and so on (back and forth infinitely often), all by the same logic. We conclude that the chain will, over and over again, return to state 0 w1; it will do so infinitely often; 0 is a recurrent state for the simle symmetric random walk. Thus (since the chain is irreducible) all states are recurrent. Let τ = min{n 1 : R n = 0 R 0 = 0}, the so-called return time to state 0. We just argued that τ is a roer random variable, that is, P (τ < ) = 1. This means that if the chain starts in state 0, then, if we wait long enough, we will (w1) see it return to state 0. What we will rove later is that E(τ) = ; meaning that on average our wait is infinite. This imlies that the simle symmetric random walk forms a null recurrent Markov chain. 6