On the mathematical theory of splitting and Russian roulette



Similar documents
Master s Theory Exam Spring 2006

Modeling and Performance Evaluation of Computer Systems Security Operation 1

A Model of Optimum Tariff in Vehicle Fleet Insurance

Random access protocols for channel access. Markov chains and their stability. Laurent Massoulié.

Some Research Problems in Uncertainty Theory

The Exponential Distribution

Monte Carlo Methods in Finance

Exam Introduction Mathematical Finance and Insurance

THE CENTRAL LIMIT THEOREM TORONTO

LECTURE 4. Last time: Lecture outline

BX in ( u, v) basis in two ways. On the one hand, AN = u+

Stochastic Processes and Queueing Theory used in Cloud Computer Performance Simulations

Introduction to Markov Chain Monte Carlo

1 Short Introduction to Time Series

Nonparametric adaptive age replacement with a one-cycle criterion

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

Pull versus Push Mechanism in Large Distributed Networks: Closed Form Results

Sums of Independent Random Variables

Microeconomic Theory: Basic Math Concepts

On a comparison result for Markov processes

Continued Fractions and the Euclidean Algorithm

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

**BEGINNING OF EXAMINATION** The annual number of claims for an insured has probability function: , 0 < q < 1.

Lecture 13: Martingales

General Theory of Differential Equations Sections 2.8, , 4.1

3. INNER PRODUCT SPACES

How to Gamble If You Must

Performance Analysis of a Telephone System with both Patient and Impatient Customers

Notes on Continuous Random Variables

The CUSUM algorithm a small review. Pierre Granjon

Math 526: Brownian Motion Notes

Lectures on Stochastic Processes. William G. Faris

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

The Monte Carlo Framework, Examples from Finance and Generating Correlated Random Variables

COMPLETE MARKETS DO NOT ALLOW FREE CASH FLOW STREAMS

Math 120 Final Exam Practice Problems, Form: A

Department of Economics

A characterization of trace zero symmetric nonnegative 5x5 matrices

Asian Option Pricing Formula for Uncertain Financial Market

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan , Fall 2010

Least Squares Estimation

Simple Markovian Queueing Systems

LOGNORMAL MODEL FOR STOCK PRICES

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS

Analysis of a Production/Inventory System with Multiple Retailers

THE DYING FIBONACCI TREE. 1. Introduction. Consider a tree with two types of nodes, say A and B, and the following properties:

Vilnius University. Faculty of Mathematics and Informatics. Gintautas Bareikis

RESULTANT AND DISCRIMINANT OF POLYNOMIALS

JUST THE MATHS UNIT NUMBER 1.8. ALGEBRA 8 (Polynomials) A.J.Hobson

Universal Algorithm for Trading in Stock Market Based on the Method of Calibration

QUANTIZED INTEREST RATE AT THE MONEY FOR AMERICAN OPTIONS

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Probability and statistics; Rehearsal for pattern recognition

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem. Lecture 12 04/08/2008. Sven Zenker

An Extension Model of Financially-balanced Bonus-Malus System

The Joint Distribution of Server State and Queue Length of M/M/1/1 Retrial Queue with Abandonment and Feedback

Large induced subgraphs with all degrees odd

Optimization of Supply Chain Networks

MATH PROBLEMS, WITH SOLUTIONS

Non-Separable Detachments of Graphs

Multi-state transition models with actuarial applications c

Supplement to Call Centers with Delay Information: Models and Insights

Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm.

Chapter 1. Introduction

Roots of Polynomials

Pricing and calibration in local volatility models via fast quantization

EXIT TIME PROBLEMS AND ESCAPE FROM A POTENTIAL WELL

9.2 Summation Notation

Lecture 10. Finite difference and finite element methods. Option pricing Sensitivity analysis Numerical examples

Statistical Machine Learning

A Portfolio Model of Insurance Demand. April Kalamazoo, MI East Lansing, MI 48824

ESTIMATION OF CLAIM NUMBERS IN AUTOMOBILE INSURANCE

VERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS

Important Probability Distributions OPRE 6301

M/M/1 and M/M/m Queueing Systems

MARTINGALES AND GAMBLING Louis H. Y. Chen Department of Mathematics National University of Singapore

Lectures 5-6: Taylor Series

Math 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 2 Solutions

A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS

Single item inventory control under periodic review and a minimum order quantity

GLMs: Gompertz s Law. GLMs in R. Gompertz s famous graduation formula is. or log µ x is linear in age, x,

How To Solve A Sequential Mca Problem

Wald s Identity. by Jeffery Hein. Dartmouth College, Math 100

Numerical Analysis Lecture Notes

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the school year.

A QUEUEING-INVENTORY SYSTEM WITH DEFECTIVE ITEMS AND POISSON DEMAND.

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

Transcription:

On the mathematical theory of splitting and Russian roulette techniques St.Petersburg State University, Russia

1. Introduction Splitting is an universal and potentially very powerful technique for increasing efficiency of simulation studies. The idea of this technique goes back to von Neumann (see [Kan and Harris,1951]). A theory of multilevel splitting is rather well developed (see e.g.[l Ecuryer et al.2006] ). However, in the most of papers strongly restrictive assumptions on transition probabilities are imposed. In [Melas, 1993, 1997, 2002] and [Ermakov and Melas, 1995] a more general theory is developed. It is based on the introduction of a probability measure governing the procedures of splitting and Russian roulette. In the present talk we will describe basic ideas and results of this theory.

2. Formulation of the problem Let random variables X 1, X 2,... are defined on probability space (X, A, P) and form a homogenous Markov chaim with transition function P(x, dy) = P{X 1 dy X 0 = x} measurable in the forst argument and such that X P(x, dy) = 1. We will consider Harris positively recurrent Markov chains (see, f.e., (Nummelin, 1984) for definitions). In fact, we need only a mild restrictions that secure existence of the stationary distribution (denote it by π(dx) ) and asymptotic unbiasedness of usual Monte Carlo estimators.

It is required to estimate one or several functionals of the form J i = h i (x) π(dx), i = 1, 2,..., l X with respect to simulation results. In case of chains with finite state space these functionals have the form m J i = h i (x) π x, i = 1, 2,..., l. x=0

3. Branching technique Let {X n } be a Markov chain of the general form described above. Let η n chains {X n } be simulated at the n th step. A procedure for regulating η n can be introduced in the following way. Definition. A measurable function β(x) on (X, A) with the following properties: 1 β(x) 0, x Ω, 2 β(x)π(dx) > 0, if π(b) > 0 B is called a branching density.

Introduce the random variable and r(x, y) = β(y)/β(x), with probability 1 β(y)/β(x) r(x, y) = β(y)/β(x) + 1, with probability β(y)/β(x), where a, with a = β(y)/β(x), means the integer part of a and a means the whole part of a. At step zero, set η 0 = 1. At each following step (n + 1 = 1, 2,...) if Xn γ = x, X γ n+1 = y, then when β(y)/β(x) < 1 the simulation of path γ discontinues with probability 1 β(y)/β(x). When β(y)/β(x) 1 we simulate r(x, y) 1 additional paths, beginning from point x, where k-th steps of these chains will be denoted by index n + k.

Let us simulate k loops: {X γ n (i); γ = 1, 2,..., η n (i)}, i = 1, 2,..., k. Let us determine estimators of the functionals J j (j = 1, 2,..., l) by the formula Ĵ k,β (j) = 1 k / k k Y β (i, j) Ỹ β (i, j), i=1 i=1 where Y β (i, j) = N(i) η n h j (Xn γ (i)) /β (Xn γ (i)), n=0 γ=1 Ỹ β (i, j) = Y β (i, j) with h i (x) 1. Note that when β(x) 1 the process reduces to direct simulation of the chain {X n }, and the estimators Ĵi,β become the known ratio estimators of the regeneration method (see Crane, Iglehart, 1974).

Denote D(β) = T (β) = lim k ( ) lim (Ĵk,β ) l kcov (i), Ĵk,β(j) k k T i,β /k. i=1 i,j=1 By virtue of Law of Large Numbers, T (β) is the expectation of the number of steps for all chains in one loop. Set ĥ j (x) = Ĵk,β(j), j = 1, 2,..., l, under the condition that all paths begin at x (X0 1 (i) = x, i = 1,..., k).,

4. Expression for covariances Theorem 1. When regularity conditions (see Melas(1993)) are valid for general Markov chains, estimators Ĵk,β(i) are asymptotically unbiased and D(β) = T (β) = ( ) l β 1 (x)π(dx) (d ij (x) + r β ), i,j=1 β(x) π(dx), where d ij (x) = Eĥi(x) ĥj(x) Eĥi(x) Eĥj(x), E is the expectation operator, and r β is a remainder term.

Theorem 1 permits us to introduce optimality criteria for the choice of the branching measure. To simplify the notation let us consider that π(dx) = π(x) dx. Set / τ(x) = π(x) β(x) β(x) π(x) dx (1) τ x = π x β x / x β x π x (finite chains). (2) Then T (τ) = 1 for any β. Dropping the remainder term, we obtain the matrix ( ) D(τ) = τ 1 (x) d ij (x) π 2 (x) dx, ( m ) D(τ) = τx 1 πx 2 d ij (x) (finite chains). x=0

As an estimation accuracy criterion let us take quantities trad(τ), (3) det D(τ) (4) where A is an arbitrary nonnegative-definite matrix. Note that when A = I (where I is identity matrix) and l = 1, trad(τ) is just the variance of the estimator Ĵ1. The probability measure τ(x) is called a design of the simulation experiment.

5. Optimal experimental designs Consider the minimization problem for criteria (3) and (4) in the class of probability measures τ induced by branching densities. The optimal design for criterion (3) can be found with the help of the Schwarz inequality. The direct application of this inequality brings the following result. Theorem 2. Let the hypothesis of Theorem 1 be satisfied. Then the experimental design minimizing the value of tr AD(τ) is given by formula (1), where β(x) = trad(τ).

Theorem 3. For finite regular Markov chains and functions h i (x) of the form h i (x) = δ ix, i = 1,..., m, there exists an experiment design minimizing det D(τ). This design is unique and satisfies the relation where τ(x) = ( tr B x B 1 (τ ) ) 1/2 πx / m, B x = (p xy δ yz p xy p xz ) m y,z=1, x = 0, 1,..., m, B(τ) = m ( ) π 2 x /τ x Bx. x=0 The design described in Theorem 3 will be called D-optimal.

In order to construct a D-optimal design the following iterative method can be used. Set τ 0 (y) = π y, y = 0,..., m, τ k+1 (y, α) = (1 α) τ k (y) + ατ (x) (y), k = 0, 1, 2,... where τ (x) (y) = δ xy x = arg max x π 2 x τ 2 k (x) tr B xb 1 (τ k ).

Set α k = arg min α [0,1] det B ({τ k+1(y, α)}), τ k+1 (y) = τ k+1 (y, α k ), y = 0, 1,..., m. Theorem 4. Under the hypothesis of Theorem 3 for k where det D(τ ) = min det D(τ). det D(τ k ) det D(τ ),

Consider a random walk process on a line S 1 = 0, S n+1 = S n + X n, k = 1, 2,..., where X n, k = 1, 2,... are independent random variables with common density function f (x), and connect it with the waiting process W 1 = 0, W n+1 = max(0, W n + X n ), n = 1, 2,... It is known (see Feller, 1970) that the quantities W n and M n = max{s 1,..., S n } have the same distribution. Under an additional condition on EX 1, M n M and W n W in distribution, where M and W are random variables.

Set π(0) = P{W = 0}, π(x) = W (x), W (x) = P{W < x}, π(x) = 0 for x < 0, X n = W n, S = {0}, π(dx) = af (x)dx, a = 1/(1 F (0)), and 1 F (0) = 0 f (t)dt, h 1 (x) = 0, x < v, h 1 (x) = 1, x v, l = 1. The problem is the estimation of the functional J = J 1 = h 1 (x)π(x)dx. Suppose that f (x) satisfies the condition e λ 0t f (t)dt = 1 for some positive λ 0 and EX 1 < 0. Set β(x) 1, β (t) = e λ 0t, 0 < t < v, β (t) = e λ 0v, t v.

Define the relative efficiency by the formula R = R v = T β DĴβ. T β DĴβ It is known (see Feller, 1970) that as v θ e λ0t 0. Theorem 5. Let the conditions formulated above be satisfied and for some ε > 0 e 2(λ 0+ε)t f (t) dt <. Then for v ( ) 1 R v = O θ ln 2, θ e λ0v. (1/θ)

6. Numerical example To compare the branching technique (BT) and the usual regeneration technique (RT) we simulated the corresponding random walk by both methods during 100,000 loops. The magnitude v was chosen so that θ = 10 2, 10 3,..., 10 6, a = 1/2. Denote by I the magnitude I = θ 2 /(ETD θ) where θ is the proper value of the estimated parameters, ET is the mean length of the simulated paths and D θ is the sample variance. We have obtained results presented in the following table: ln(1/θ) 2 3 4 5 6 I (RT ) 0, 385 0, 026 0, 000 0, 000 0, 000 I (BT ) 0, 686 0, 298 0, 185 0, 083 0, 041

7. Example of D-optimal design Consider a particular case of Markov chain imbedded into the random process, corresponding to the length of queue with one server and m places for waiting. We will consider the simplest case when the input stream is a simplest stream and the time of service is an exponentially distributed random value.

Let ρ be the load of the system. Then the matrix P is of the form 0 1 0 0... 1 0 0 0 1 0 P =,. 1 0 0 1 = ρ/(1 + ρ), π 1 = π 0, π i+1 = ρ i π 0, i = 1,..., m 1, π 0 = 1/(1 + + ρ +... + ρ m 1 ). We can calculate by induction that detb(τ) = ( m i=1 π 2 i τ i ) m (1 ) m.

From here we found that D-optimal design is τ = { 0, 1 m,..., 1 m and ( ) ( detb(π) 1/m m ) 1/m / I (τ) = = π i m detb(τ) i=1 with τ = τ. Note that I (τ) is a natural measure of efficiency of design τ. It means the ratio of the number of steps needed by immediate simulation for obtaining results with a given accurancy to the same number for the splitting and roulette approach. In table 1 values of the efficiency are given. m 5 5 5 10 10 10 ρ 1/2 1/3 1/4 1/2 1/3 1/4 I 2.0 4.0 6.8 6.0 31.6 109 }

8. Concluding remarks In fact, any immediate simulation method can be improved by including the branching technique A representation for the variance-covariance matrix derived here allows to find an optimal branching measure The full explanation of the theory can be found in the book (Ermakov, Melas, 1995) and in the papers (Melas, 1993, 1997, 2002)

References Crane, A. M., and D. L. Iglehart. 1974. Simulating stable stochastic systems. 2: Markov chains. Journal of Association of Computing Machinery 21 (1): 114-123. L Ecuyer, P., Demers, V., and B.Tuffin (2006). Splitting for rare-event simulation.in: Proceedings of the 2006 Winter Simulation Conference, 137 148,California Monterey. Ermakov, S. M., and V. B. Melas. 1995. Design and analysis of simulation experiments. Dordrecht, London: Kluwer Academic Publishers. Feller, W. 1970. An introduction to probability theory and its applications, vol. 2. 2nd ed. New York: Wiley. Kahn, H., and T. E. Harris. (1951). Estimation of particle transmission by random sampling. National Bureau os Standards Applied Mathematical Series 12:27 30.

Melas, V. B. 1993. Optimal simulation design by branching technique. Model Oriented Data Analysis, ed. W. G. Muller, H. P. Wynn, and A. A., Zhigljavsky. Heidelberg: Physica-Verlag, 113-127. Melas, V.B. 2002. On simulating finite Markov chains by the splitting and roulette approach. Information processes. vol 2, issue 2, pp. 228-230. Nummelin, E. 1984. General irreducible Markov chains and non-negative operators. Cambridge: Cambridge University Press. Sigman, K. 1988. Queues as Harris recurrent Markov chains. Queueing systems 3: 179-198.