The Strong Markov Property

Similar documents
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

BANACH AND HILBERT SPACE REVIEW

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

1 if 1 x 0 1 if 0 x 1

Mathematics Course 111: Algebra I Part IV: Vector Spaces

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform

Moreover, under the risk neutral measure, it must be the case that (5) r t = µ t.

1 Sets and Set Notation.

Math 4310 Handout - Quotient Vector Spaces

3. Reaction Diffusion Equations Consider the following ODE model for population growth

Lecture 13 - Basic Number Theory.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Mathematical Finance

26 Integers: Multiplication, Division, and Order

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

Similarity and Diagonalization. Similar Matrices

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

Scalar Valued Functions of Several Variables; the Gradient Vector

36 CHAPTER 1. LIMITS AND CONTINUITY. Figure 1.17: At which points is f not continuous?

LEARNING OBJECTIVES FOR THIS CHAPTER

Continued Fractions and the Euclidean Algorithm

So let us begin our quest to find the holy grail of real analysis.

1 The Brownian bridge construction

NOTES ON LINEAR TRANSFORMATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

4.5 Linear Dependence and Linear Independence

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

Notes on metric spaces

Metric Spaces Joseph Muscat 2003 (Last revised May 2009)

LS.6 Solution Matrices

Lectures 5-6: Taylor Series

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

LINEAR ALGEBRA W W L CHEN

Solving Linear Systems, Continued and The Inverse of a Matrix

Matrix Representations of Linear Transformations and Changes of Coordinates

Adaptive Online Gradient Descent

CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e.

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Let H and J be as in the above lemma. The result of the lemma shows that the integral

9.4. The Scalar Product. Introduction. Prerequisites. Learning Style. Learning Outcomes

10.2 Series and Convergence

Unified Lecture # 4 Vectors

MEASURE AND INTEGRATION. Dietmar A. Salamon ETH Zürich

The degree of a polynomial function is equal to the highest exponent found on the independent variables.

Probability Theory. Florian Herzog. A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T..

Lecture 13: Martingales

1 Error in Euler s Method

11 Ideals Revisiting Z

Student Outcomes. Lesson Notes. Classwork. Discussion (10 minutes)

MATH APPLIED MATRIX THEORY

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

1 Norms and Vector Spaces

Rate of growth of D-frequently hypercyclic functions

160 CHAPTER 4. VECTOR SPACES

Section 1.3 P 1 = 1 2. = P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., =

3. Mathematical Induction

5.1 Radical Notation and Rational Exponents

1. Prove that the empty set is a subset of every set.

Some stability results of parameter identification in a jump diffusion model

A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University

CHAPTER IV - BROWNIAN MOTION

This asserts two sets are equal iff they have the same elements, that is, a set is determined by its elements.

Elements of probability theory

Normal distribution. ) 2 /2σ. 2π σ

Linear Algebra Notes for Marsden and Tromba Vector Calculus

ARBITRAGE-FREE OPTION PRICING MODELS. Denis Bell. University of North Florida

1.7 Graphs of Functions

THE DIMENSION OF A VECTOR SPACE

Markov random fields and Gibbs measures

Microeconomic Theory: Basic Math Concepts

Chapter 5. Banach Spaces

An Introduction to Partial Differential Equations in the Undergraduate Curriculum

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

9.2 Summation Notation

Notes on Factoring. MA 206 Kurt Bryan

Inverse Functions and Logarithms

Linear Algebra. A vector space (over R) is an ordered quadruple. such that V is a set; 0 V ; and the following eight axioms hold:

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

CONTINUED FRACTIONS AND FACTORING. Niels Lauritzen

Lecture L3 - Vectors, Matrices and Coordinate Transformations

A positive exponent means repeated multiplication. A negative exponent means the opposite of repeated multiplication, which is repeated

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

Representation of functions as power series

Review of Fundamental Mathematics

Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

How To Know If A Domain Is Unique In An Octempo (Euclidean) Or Not (Ecl)

Systems of Linear Equations

Lecture 7: Finding Lyapunov Functions 1

Cartesian Products and Relations

Vieta s Formulas and the Identity Theorem

Chapter 4 Lecture Notes

Notes on Determinant

CONSTANT-SIGN SOLUTIONS FOR A NONLINEAR NEUMANN PROBLEM INVOLVING THE DISCRETE p-laplacian. Pasquale Candito and Giuseppina D Aguí

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

1 Short Introduction to Time Series

God created the integers and the rest is the work of man. (Leopold Kronecker, in an after-dinner speech at a conference, Berlin, 1886)

Row Ideals and Fibers of Morphisms

0 <β 1 let u(x) u(y) kuk u := sup u(x) and [u] β := sup

Transcription:

The Strong Markov Property Throughout, X := {X } denotes a Lévy process on with triple (σ), and exponent Ψ. And from now on, we let { } denote the natural filtration of X, all the time remembering that, in accord with our earlier convention, { } satisfies the usual conditions. Transition measures and the Markov property Definition 1. The transition measures of X are the probability measures P (A) := P { + X A} defined for all,, and A ( ). In other words, each P ( ) is the law of X started at. We single out the case =by setting µ (A) := P ( A); thus, µ is the distribution of X for all >. Note, in particular, that µ = δ is the point mass at. Proposition 2. For all, and measurable : +, E[(X + ) ]= () P (X d) a.s. 53

54 9. The Strong Markov Property Consequently, for all, < 1 < 2 < <, and measurable 1 : +, E ( + X ) (1) =1 = P 1 ( d 1 ) P 2 1 ( 1 d 2 ) P 1 ( 1 d ) ( ) Property (1) is called the Chapman Kolmogorov equation. That property has the following ready consequence: Transition measures determine the finite-dimensional distributions of X uniquely. Definition 3. Any stochastic process {X } that satisfies the Chapman- Kolmogorov equation is called a Markov process. This definition continues to make sense if we replace ( ( )) by any measurable space on which we can construct infinite families of random variables. Thus, Lévy processes are cadlag Markov processes that have special addition properties. In particular, as Exercise below 1 shows, Lévy processes have the important property that the finite-dimensional distributions of X are described not only by {P ( )} but by the much-smaller family {µ ( )}. Note, in particular, that if : + is measurable,, and, then E( + X )= () P (d) = ( + ) µ (d) Therefore, if we define µ (A) := µ ( A) for all and A ( ), where A := { : A}, then we have the following convolution formula, valid for all measurable : +,, and : E( + X )=( µ )() And, more generally, for all measurable : +,, and [Why is this more general?] E[(X + ) ]=( µ )(X ) Proposition 4. The family {µ } of Borel probability measure on is a convolution semigroup in the sense that µ µ = µ + for all. Moreover, ˆµ (ξ) = exp( Ψ(ξ)) for all and ξ. Similarly, { µ } is a convolution semigroup with ˆ µ (ξ) = exp( Ψ( ξ)). a.s. =1

The strong Markov property 55 Proof. The assertion about µ follows from the assertion about µ [or you can repeat the following with µ in place of µ]. Since µ is the distribution of X, the characteristic function of X is described by ˆµ (ξ) = exp( Ψ(ξ)). The proposition follows immediately from this, because ˆµ (ξ) ˆµ (ξ) = exp( ( + )Ψ(ξ)) = ˆµ + (ξ). The strong Markov property Theorem 5 (The strong Markov property). Let T be a finite stopping time. Then, the process X T := {X T}, defined by X T := X T+ X T is a Lévy process with exponent Ψ and independent of T. Proof. X T is manifestly cadlag [because X is]. In addition, one checks that whenever < 1 < < and A 1 A ( ), P XT+ X T A T = P X A a.s.; =1 see Exercise 2 on page 32. This readily implies that the finite-dimensional distributions of X T are the same as the finite-dimensional distributions of X, and the result follows. Theorem 5 has a number of deep consequences. The following shows that Lévy processes have the following variation of strong Markov property. The following is attractive, in part because it can be used to study processes that do not have good additivity properties. Corollary 6. For all finite stopping times T, every, and all measurable functions : +. =1 E (X T+ ) T = ()P (X T d) a.s. Let T be a finite stopping time, and then define (T) = { (T) } to be the natural filtration of the Lévy process X T. The following is a useful corollary of the strong Markov property. Corollary 7 (Blumenthal s zero-one law; Blumenthal, 1957). Let T be a finite stopping time. Then (T) is trivial; i.e., P(A) {1} for all A (T).

56 9. The Strong Markov Property The following are nontrivial examples of elements of (T) : X T+ X T Υ 1 := lim inf 1/α = where α> is fixed; X T+ X T Υ 2 := lim sup =1 ; or 2 ln ln(1/) Υ 3 := such that X T+ X T > for all 1 in dimension one, etc. Proof of Blumenthal s zero-one law. The strong Markov property [Corollary 6] reduces the problem to T. And of course we do not need to write () since () is the same object as. For all 1 define to be the completion of the sigma-algebra generated by the collection {X +2 X 2 } [2 ]. By the Markov property, 1 2 are independent sigma-algebras. Their tail sigma-algebra is the smallest sigma-algebra that contains =N for all N 1. Clearly is complete, and Kolmogorov s zero-one law tells us that is trivial. Because =N contains the sigma-algebra generated by all increments of the form X + X where [2 2 +1 ] for some N, and since X as, it follows that contains, where denotes the sigma-algebra generated by {X } []. Since is complete, this implies [in fact, = ] as well, and hence is trivial because is. Thus, for example, take the set Υ 1 introduced earlier. We can apply the Blumenthal zero-one to the Υ 1 and deduce the following: X T+ X T For every α>, P lim inf 1/α = =or 1 You should construct a few more examples of this type. Feller semigroups and resolvents Define a collection {P } of linear operators by (P )() := E( + X )= ()P (d) =( µ )() for. [Since X =, P = δ is point mass at zero.] The preceding is well defined for various measurable functions :. For instance, everything is fine if is nonnegative, and also if (P )() < for all and [in that case, we can write P = P + P ]. The Markov property of X [see, in particular, Proposition 4] tells us that (P + )() =(P (P ))(). In other words, P + = P P = P P for all (2)

Feller semigroups and resolvents 57 where P P is shorthand for P (P ) etc. Since P and P commute, in the preceding sense, there is no ambiguity in dropping the parentheses. Definition 8. The family {P } is the semigroup associated with the Lévy process X. The resolvent { λ } λ> of the process X is the family of linear operators defined by ( λ )() := e λ (P )() d = E e λ ( + X ) d (λ >) This can make sense also for λ =, and we write in place of. Finally, λ is called the λ-potential of when λ>; when λ =, we call it the potential of instead. emark 9. It might be good to note that we can cast the strong Markov property in terms of the semigroup {P } as follows: For all, finite stopping times T, and : + measurable, E[(X T+ ) T ]=(P )(X T ) almost surely. Formally speaking, λ = e λ P d (λ ) defines the Laplace transform of the [infinite-dimensional] function P. Once again, λ is defined for all Borel measurable :, if either ; or if λ is well defined. ecall that C ( ) denotes the collection of all continuous : that vanish at infinity [() as ]; C ( ) is a Banach space in norm := sup (). The following are easy to verify: (1) Each P is a contraction [more precisely nonexpansive] on C ( ). That is, P for all ; (2) {P } is a Feller semigroup. That is, each P maps C ( ) to itself and lim P =; (3) If λ>, then λ λ is a contraction [nonexpansive] on C ( ); (4) If λ>, then λ λ maps C ( ) to itself. The preceding describe the smoothness behavior of P and λ for fixed and λ. It is also not hard to describe the smoothness properties of them as functions of and λ. For instance, Proposition 1. For all C ( ), lim sup P + P = and lim λ λ λ =

58 9. The Strong Markov Property Proof. We observe that P = sup E( + X ) () sup E ( + X ) () Now every C ( ) is uniformly continuous and bounded on all of. Since X is right continuous and is bounded, it follows from the bounded convergence theorem that lim P =. But the semigroup property implies that P + P = P (P ) P, since P is a contraction on C ( ). This proves the first assertion. The second follows from the first, since λ λ = e P /λ d by a change of variables. Proposition 11. If C ( ) L ( ) for some [1 ), then P L ( ) L ( ) for all and λ λ L ( ) L ( ) for all λ>. In words, the preceding states that P L ( ) for every [1 ) and λ >. and λ λ are contractions on Proof. If C ( ) L ( ), then for all, (P )() d = E( + X ) d E ( + X ) d = () d This proves the assertion about P ; the one about λ is proved similarly. The Hille Yosida theorem One checks directly that for all µ λ, λ µ =(µ λ) λ µ (3) This is called the resolvent equation, and has many consequences. For instance, the resolvent equation implies readily the commutation property µ λ = λ µ. For another consequence of the resolvent eqution, suppose = µ for some C ( ) and µ>. Then, C ( ) and by the resolvent equation, λ =(µ λ) λ. Consequently, = λ, where := +(λ µ) λ C ( ). In other words, µ (C ( )) = λ (C ( )), whence Dom[L] := µ : C ( ) does not depend on µ>. And Dom[L] is dense in C ( ) [Proposition 1]. For yet another application of the resolvent equation, let us suppose that λ for some λ> and C ( ). Then the resolvent equation implies that µ for all µ. Therefore, = lim µ µ µ =. This implies

The Hille Yosida theorem 59 that every λ is a one-to-one and onto map from C ( ) to Dom[L]; i.e., it is invertible! Definition 12. The [infinitesimal] generator of X is the linear operator L : Dom[L] C ( ) that is defined uniquely by L := λi λ 1 where I := defines the identity operator I on C ( ). The space Dom[L] is the domain of L. The following is perhaps a better way to think about L; roughly speaking, it asserts that P L for small, and λ λ λ 1 L for λ large. Theorem 13 (Hille XXX, Yosida XXX). If Dom[L], then lim sup λ( λ )() () λ (L)() 1/λ = lim sup (P )() () (L)() = Because = P, the Hille Yosida theorem implies, among other things, that ( / )P = = L, where the partial derivative is really a right derivative. See Exercise 4 for a consequence in partial integro-differential equations. Proof. Thanks to Proposition 1 and the definition of the generator, L = λ 1 for all Dom[L], whence λ λ λ L = λ λ L in C ( ) as λ 1/λ This proves half of the theorem. For the other half recall that Dom[L] is the collection of all functions of the form = λ, where C ( ) and λ>. By the semigroup property, for such λ and we have P λ = e λ P + d = e λ e λ P d = e λ λ e λ P d Consequently, for all = λ Dom[L], P e λ 1 = λ eλ λ λ e λ P d in C ( ) as But λ λ = λ 1 λ = L.

6 9. The Strong Markov Property The form of the generator Let denote the collection of all rapidly-decreasing test functions :. That is, if and only if C ( ), and and all of its partial derivatives vanish faster than any polynomial. In other words, if D is a differential operator [of finite order] and 1, then sup (1 + ) (D)() <. It is easy to see that L 1 ( ) C ( ) and is dense in C ( ). And it is well known that if, then ˆ as well, and vice versa. It is possible to see that if ˆ L 1 ( ), then for all and λ>, P (ξ) =e Ψ( ξ)ˆ(ξ) λ (ξ) = ˆ(ξ) λ + Ψ( ξ) for all ξ (4) Therefore, it follows fairly readily that when Dom[L] L 1 ( ), L L 1 ( ), and ˆ L 1 ( ), then we have L(ξ) = Ψ( ξ)ˆ(ξ) for every ξ (5) It follows immediately from these calculations that: (i) Every P and λ map to ; and (ii) Therefore, is dense in Dom[L]. Therefore, we can try to understand L better by trying to compute L not for all Dom[L], but rather for all in the dense subcollection. But the formula for the Fourier transform of L [together with the estimate Ψ(ξ) = O(ξ 2 )] shows that L : and (L)() = 1 (2π) e ξ Ψ( ξ)ˆ(ξ) dξ for all and Consider the simplest case that the process X satisfies X = for some ; i.e., Ψ(ξ) = ( ξ). In that case, we have (L)() = 1 (2π) ( ξ)e ξ ˆ(ξ) dξ = 1 (2π) e ξ ( (ξ)) dξ = ( )() thanks to the inversion formula. The very same computation works in the more general setting, and yields Theorem 14. If, then L = C + J, where (C)() = ( )()+ 1 (σ 2 σ) () 2 and (J)() := 1 ( + ) () ( )()1l [1) () (d) for all Moreover, J is the generator of the pure-jump part; and C = + 1 2 σ σ is the generator of the continuous/gaussian part.

Problems for Lecture 9 61 Here are some examples: If X is Brownian motion on, then L = 1 2 is one-half of the Laplace operator [on ]; If X is the Poisson process on with intensity λ ( ), then (L)() =λ[( + 1) ()] for [might be easier to check Fourier transforms]; If X is the isotropic stable process with index α ( 2), then for all, ( + ) () ( )()1l[1] () (L)() =const d +α Since L(ξ) ˆ(ξ) ξ α, L is called the fractional Laplacian with fractional power α/2. It is sometimes written as L = ( ) α/2 ; the notation is justified [and explained] by the symbolic calculus of pseudo-differential operators. Problems for Lecture 9 1. Prove that P (A)=P (A ) for all,, and A ( ), where A := { : A}. Conclude that the Chapman Kolmogorov equation is equivalent to the following formula for E =1 ( + X ): P 1 (d 1 ) P 2 1 (d 2 ) P 1 (d ) ( + + ) using the same notation as Proposition 2. 2. Suppose Y L 1 (P) is measurable with respect to σ({x } ) for a fixed nonrandom. Prove that E(Y )=E(Y X ) a.s. 3. Verify that X := { X } is a Lévy process; compute its transition measures P (d) and verify the following duality relationship: For all measurable : + and, () d () P (d) = () d () P (d) 4. Prove that () := (P )() solves [weakly] the partial integro-differential equation ()=(L)() for all > and subject to ( )= (). 5. Derive the resolvent equation (3). 6. Verify (4) and (5). =1

62 9. The Strong Markov Property 7. First, improve Proposition 11 in the case =2as follows: Prove that there exists a unique continuous extension of P to all of L 2 ( ). Denote that by P still. Next, define Dom 2 [L] := L 2 ( ): Ψ(ξ) 2 ˆ(ξ) 2 dξ< Then prove that lim 1 (P ) exists, as a limit in L 2 ( ), for all Dom 2 [L]. Identify the limit when C ( ).