Convergence of Feller Processes

Similar documents
Notes on metric spaces

Metric Spaces. Chapter Metrics

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

1 if 1 x 0 1 if 0 x 1

Chapter 5. Banach Spaces

Pull versus Push Mechanism in Large Distributed Networks: Closed Form Results

Fixed Point Theorems

1 Norms and Vector Spaces

Convex analysis and profit/cost/support functions

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

BANACH AND HILBERT SPACE REVIEW

THE CENTRAL LIMIT THEOREM TORONTO

NOTES ON LINEAR TRANSFORMATIONS

1. Prove that the empty set is a subset of every set.

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

1 The Brownian bridge construction

EXIT TIME PROBLEMS AND ESCAPE FROM A POTENTIAL WELL

LS.6 Solution Matrices

16.3 Fredholm Operators

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

SOLUTIONS TO EXERCISES FOR. MATHEMATICS 205A Part 3. Spaces with special properties

Fluid Approximation of Smart Grid Systems: Optimal Control of Energy Storage Units

Mathematics for Econometrics, Fourth Edition

Mathematical Methods of Engineering Analysis

Duality of linear conic problems

How To Prove The Dirichlet Unit Theorem

THE BANACH CONTRACTION PRINCIPLE. Contents

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

Stephane Crepey. Financial Modeling. A Backward Stochastic Differential Equations Perspective. 4y Springer

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

T ( a i x i ) = a i T (x i ).

I. Pointwise convergence

88 CHAPTER 2. VECTOR FUNCTIONS. . First, we need to compute T (s). a By definition, r (s) T (s) = 1 a sin s a. sin s a, cos s a

Finite dimensional topological vector spaces

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

Follow links for Class Use and other Permissions. For more information send to:

Continued Fractions and the Euclidean Algorithm

Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011

x a x 2 (1 + x 2 ) n.

Random graphs with a given degree sequence

Separation Properties for Locally Convex Cones

Adaptive Search with Stochastic Acceptance Probabilities for Global Optimization

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Lectures 5-6: Taylor Series

Finitely Additive Dynamic Programming and Stochastic Games. Bill Sudderth University of Minnesota

A domain of spacetime intervals in general relativity

Metric Spaces Joseph Muscat 2003 (Last revised May 2009)

Math 4310 Handout - Quotient Vector Spaces

Staffing and Control of Instant Messaging Contact Centers

0 <β 1 let u(x) u(y) kuk u := sup u(x) and [u] β := sup

Mathematics Course 111: Algebra I Part IV: Vector Spaces

The Ergodic Theorem and randomness

Some stability results of parameter identification in a jump diffusion model

ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING. 0. Introduction

4 Lyapunov Stability Theory

MA651 Topology. Lecture 6. Separation Axioms.

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 7, JULY G. George Yin, Fellow, IEEE, and Vikram Krishnamurthy, Fellow, IEEE

Let H and J be as in the above lemma. The result of the lemma shows that the integral

Nonparametric adaptive age replacement with a one-cycle criterion

Probability and Random Variables. Generation of random variables (r.v.)

The Exponential Distribution

Hydrodynamic Limits of Randomized Load Balancing Networks

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

GENERIC COMPUTABILITY, TURING DEGREES, AND ASYMPTOTIC DENSITY

Random access protocols for channel access. Markov chains and their stability. Laurent Massoulié.

Gambling and Data Compression

On a comparison result for Markov processes

Gambling Systems and Multiplication-Invariant Measures

Estimating the Degree of Activity of jumps in High Frequency Financial Data. joint with Yacine Aït-Sahalia

Introduction to Topology

2. Length and distance in hyperbolic geometry

Applying Mean-field Approximation to Continuous Time Markov Chains

ON COMPLETELY CONTINUOUS INTEGRATION OPERATORS OF A VECTOR MEASURE. 1. Introduction

9 More on differentiation

Degrees that are not degrees of categoricity

A class of infinite dimensional stochastic processes

Probability Generating Functions

Adaptive Online Gradient Descent

Non-Arbitrage and the Fundamental Theorem of Asset Pricing: Summary of Main Results

Sensitivity analysis of European options in jump-diffusion models via the Malliavin calculus on the Wiener space

How To Find Out How To Calculate A Premeasure On A Set Of Two-Dimensional Algebra

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

Notes from Week 1: Algorithms for sequential prediction

ALMOST COMMON PRIORS 1. INTRODUCTION

MATH1231 Algebra, 2015 Chapter 7: Linear maps

Bisimulation and Logical Preservation for Continuous-Time Markov Decision Processes

Cycles and clique-minors in expanders

4: SINGLE-PERIOD MARKET MODELS

Monte Carlo Methods in Finance

3. INNER PRODUCT SPACES

Columbia University in the City of New York New York, N.Y

and s n (x) f(x) for all x and s.t. s n is measurable if f is. REAL ANALYSIS Measures. A (positive) measure on a measurable space

E3: PROBABILITY AND STATISTICS lecture notes

1. Let X and Y be normed spaces and let T B(X, Y ).

Lecture 2: Universality

Mathematical Finance

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e.

Transcription:

Chapter 15 Convergence of Feller Processes This chapter looks at the convergence of sequences of Feller processes to a iting process. Section 15.1 lays some ground work concerning weak convergence of processes with cadlag sample paths. Section 15.2 states and proves the central theorem about the convergence of sequences of Feller processes. Section 15.3 examines a particularly important special case, the approximation of ordinary differential equations by pure-jump Markov processes. 15.1 Weak Convergence of Processes with Cadlag Paths (The Skorokhod Topology) Recall that a sequence of random variables X 1, X 2,... converges in distribution d on X, or weakly converges on X, X n X, if and only if E [f(xn )] E [f(x)], for all bounded, continuous functions f. This is still true when X n are random functions, i.e., stochastic processes, only now the relevant functions f are functionals of the sample paths. Definition 196 (Convergence in Finite-Dimensional Distribution) Random processes X n on T converge in finite-dimensional distribution on X, X n fd X, when, J Fin(T ), X n (J) d X(J). Lemma 197 (Finite and Infinite Dimensional Distributional Convergence) Convergence in finite-dimensional distribution is necessary but not sufficient for convergence in distribution. 107

CHAPTER 15. CONVERGENCE OF FELLER PROCESSES 108 Proof: Necessity is obvious: the coordinate projections π t are continuous functionals of the sample path, so they must converge if the distributions converge. Insufficiency stems from the problem that, even if a sequence of X n all have sample paths in some set U, the iting process might not: recall our example (79) of the version of the Wiener process with unmeasurable suprema. Definition 198 (The Space D) By D(T, Ξ) we denote the space of all cadlag functions from T to Ξ. By default, D will mean D(R +, Ξ). D admits of multiple topologies. For most purposes, the most convenient one is the Skorokhod topology, a.k.a. the J 1 topology or the Skorokhod J 1 topology, which makes D(Ξ) a complete separable metric space when Ξ is itself complete and separable. (See Appendix A2 of Kallenberg.) For our purposes, we need only the following notion and propositions. Definition 199 (Modified Modulus of Continuity) The modified modulus of continuity of a function x D(T, Ξ) at time t T and scale h > 0 is given by w(x, t, h) inf max sup ρ(x(s), x(r)) (15.1) (I k ) k r,s I k where the infimum is over partitions of [0, t) into half-open intervals whose length is at least h (except possibly for the last one). Because x is cadlag, for fixed x and t, w(x, t, h) 0 as h 0. Proposition 200 (Weak Convergence in D(R +, Ξ)) Let Ξ be a complete, separable metric space. Then a sequence of random functions X 1, X 2,... D(R +, Ξ) converges in distribution to X D if and only if i The set T c = {t T : X(t) = X(t )} has a countable dense subset T 0, and the finite-dimensional distributions of the X n converge on those of X on T 0. ii For every t, h 0 sup E [w(x n, t, h) 1] = 0 (15.2) Proof: See Kallenberg, Theorem 16.10, pp. 313 314. Proposition 201 (Sufficient Condition for Weak Convergence) The following three conditions are all equivalent, and all imply condition (ii) in Proposition 200. 1. For any sequence of a.s.-finite F Xn -optional times τ n and positive constants h n 0, ρ(x n (τ n ), X n (τ n + h n )) P 0 (15.3)

CHAPTER 15. CONVERGENCE OF FELLER PROCESSES 109 2. For all t > 0, for all h 0 sup sup E [ρ(x n (σ), X n (τ)) 1] = 0 (15.4) σ,τ where σ and τ are F Xn -optional times σ, τ t, with σ τ τ + h. 3. For all t > 0, δ 0 sup sup sup τ t 0 h δ E [ρ(x n (τ), X n (τ + h)) 1] = 0 (15.5) where the supremum in τ runs over all F Xn -optional times t. Proof: See Kallenberg, Theorem 16.11, pp. 314 315. 15.2 Convergence of Feller Processes We need some technical notions about generators. Definition 202 (Closed and Closable Generators, Closures) A linear operator A on a Banach space B is closed if its graph { f, g B 2 : f Dom(A), g = Af } is a closed set. An operator is closable if the closure of its graph is a function (and not just a relation). The closure of a closable operator is that function. Notice, by the way, that because A is linear, it is closable iff f n 0 and Af n g implies g = 0. Definition 203 (Core of an Operator) Let A be a closed linear operator on a Banach space B. A linear subspace D Dom(A) is a core of A if the closure of A restricted to D is, again A. The idea of a core is that we can get away with knowing how the operator works on a linear subspace, which is often much easier to deal with, rather than controlling how it acts on its whole domain. Lemma 204 (Feller Generators Are Closed) The generator of every Feller semigroup is closed. Proof: We need to show that the graph of G contains all of its it points, that is, if f n Dom(G) converges (in L ) on f, and Gf n g, then f Dom(G) and Gf = g. First we show that f Dom(G). (I G)f n = f n Gf n (15.6) n n = f g (15.7)

CHAPTER 15. CONVERGENCE OF FELLER PROCESSES 110 But (I G) 1 = R 1. Since this is a bounded linear operator, we can exchange applying the inverse and taking the it, i.e., R 1 n (I G)f n = R 1 (f g) (15.8) n R 1 (I G)f n = R 1 (f g) (15.9) n f n = R 1 (f g) (15.10) f = R 1 (f g) (15.11) Since the range of the resolvents is contained in the domain of the generator, f Dom(G). We can therefore say that f g = (I G)f, which implies that Gf = g. Hence, the graph of G contains all its it points, and G is closed. Theorem 205 (Convergence of Feller Processes) Let X n be a sequence of Feller processes with semigroups K n,t and generators G n, and X be another Feller process with semigroup K t and a generator G containing a core D. Then the following are equivalent. 1. If f D, there exists a sequence of f n Dom(G n ) such that f n f 0 and A n f n Af 0. 2. For every t > 0, K n,t f K t f for every f C 0 3. K n,t f K t f 0 for each f C 0, uniformly in t for bounded positive t 4. If X n (0) d X(0) in Ξ, then X n d X in D. Proof: See Kallenberg, Theorem 19.25, p. 385. Remark: The important versions of the property above are the second convergence of the semigroups and the fourth converge in distribution of the processes. The other two are there to simplify the proof. The way the proof works is to first show that conditions (1) (3) are all equvialent, i.e., that convergence of the operators in the semi-group implies the apparently-stronger conditions about uniform convergence as well as convergence on the core. Then one establishes the equivalence between the second condition and the fourth. To go from convergence in distribution to convergence of conditional expectations is fairly straightforward; to go the other way involves using both the first and third condition, and the first part of Proposition 201. This last step uses the Feller proporties to bound, in expectation, the amount by which the process can move in a small amount of time. Corollary 206 (Convergence of Discret-Time Markov Processes on Feller Processes) Let X be a Feller process with semigroup K t, generator G and core D as in Theorem 205. Let h n be a sequence of positive real constants converging (not necessarily monotonically) to zero. Let Y n be a sequence of discrete-time Markov processes with evolution operators H n. Finally, let

CHAPTER 15. CONVERGENCE OF FELLER PROCESSES 111 X n (t) Y n ( t/h n ), with corresponding semigroup K n,t = Hn t/hn, and generator A n = (1/h n )(H n I). Even though X n is not in general a homogeneous Markov process, the conclusions of Theorem 205 remain valid. Proof: Kalleneberg, Theorem 19.28, pp. 387 388. Remark 1: The basic idea is to show that the process X n is close (in the Skorokhod-topology sense) to a Feller process X n, whose generator is A n. One then shows that X n converges on X, using Theorem 205, and that X n d Xn. Remark 2: Even though K n,t in the corollary above is not, strictly speaking, the time-evolution operator of X n, because X n is not a Markov process, it is a conditional expectation operator. Much more general theorems can be proved on when non-markov processes converge on Markov processes, using the idea that K n,t K t. See Kurtz (1975). 15.3 Approximation of Ordinary Differential Equations by Markov Processes The following result, due to Kurtz (1970, 1971), is essentially an application of Theorem 205. First, recall that continuous-time, discrete-state Markov processes work essentially like a combination of a Poisson process (giving the time of transitions) with a Markov chain (giving the state moved to on transitions). This can be generalized to continuous-time, continuous-state processes, of what are called pure jump type. Definition 207 (Pure Jump Markov Process) A continuous-parameter Markov process is a pure jump process when its sample paths are piece-wise constant. For each state, there is an exponential distribution of times spent in that state, whose parameter is denoted λ(x), and a transition probability kernel or exit distribution µ(x, B). Observe that pure-jump Markov processes always have cadlag sample paths. Also observe that the average amount of time the process spends in state x, once it jumps there, is 1/λ(x). So the time-average velocity, i.e., rate of change, starting from x, λ(x) (y x)µ(x, dy) Ξ Proposition 208 (Pure-Jump Markov Processea and ODEs) Let X n be a sequence of pure-jump Markov processes with state spaces Ξ n, holding time parameters λ n and transition probabilities µ n. Suppose that, for all n Ξ n is a Borel-measurable subset of R k for some k. Let Ξ be another measurable subset of R k, on which there exists a function F (x) such that F (x) F (y) M x y for some constant M. Suppose all of the following conditions holds.

CHAPTER 15. CONVERGENCE OF FELLER PROCESSES 112 1. The time-averaged rate of change is always finite: sup sup λ n (x) y x µ n (x, dy) < (15.12) n x Ξ n Ξ Ξ n 2. There exists a positive sequence ɛ n 0 such that sup λ n (x) y x µ n (x, dy) = 0 (15.13) x Ξ n Ξ y x >ɛ 3. The worst-case difference between F (x) and the time-averaged rates of change goes to zero: sup F (x) λ n(x) (y x)µ n (x, dy) = 0 (15.14) x Ξ n Ξ Let X(s, x 0 ) be the solution to the initial-value problem where the differential is given by F, i.e., for each 0 s t, s X(s, x 0) = F (X(s, x 0 )) (15.15) X(0, x 0 ) = x 0 (15.16) and suppose there exists an η > 0 such that, for all n, { } Ξ n y R k : inf y X(s, x 0) η 0 s t Ξ (15.17) Then X n (0) = x 0 implies that, for every δ > 0, ( P sup X n (s) X(s, x 0 ) > δ 0 s t ) = 0 (15.18) The first conditions on the X n basically make sure that they are Feller processes. The subsequent ones make sure that the mean time-averaged rate of change of the jump processes converges on the instantaneous rate of change of the differential equation, and that, if we re sufficiently close to the solution of the differential equation in R k, we re not in some weird way outside the relevant domains of definition. Even though Theorem 205 is about weak convergence, converging in distribution on a non-random object is the same as converging in probability, which is how we get uniform-in-time convergence in probability for a conclusion. There are, broadly speaking, two kinds of uses for this result. One kind is practical, and has to do with justifying convenient approximations. If n is large, we can get away with using an ODE instead of the noisy stochastic scheme, or alternately we can use stochastic simulation to approximate the solutions of ugly ODEs. The other kind is theoretical, about showing that the large-population it behaves deterministically, even when the individual behavior is stochastic and strongly dependent over time.

CHAPTER 15. CONVERGENCE OF FELLER PROCESSES 113 15.4 Exercises Exercise 15.1 (Poisson Counting Process) Show that the Poisson counting process is a pure jump Markov process. Exercise 15.2 (Exponential Holding Times in Pure-Jump Processes) Prove that pure-jump Markov processes have exponentially-distributed holding times, i.e., that if X is a Markov process with piecewise-constant sample paths, and ɛ x = inf t > 0X(t) x, that ɛ x X(0) = x is exponentially distributed. Exercise 15.3 (Solutions of ODEs are Feller Processes) Let F : R d R d be a sufficiently smooth vector field that the ordinary differential equation dx/dt = F (x) as a unique solution for every initial condition x 0. Prove that the set of solutions forms a Feller process.