Backward Equation, duality Jonathan Goodman, Fall, Backward equation and generator

Similar documents
Lecture 2 Matrix Operations

Introduction to Matrix Algebra

Similar matrices and Jordan form

Data Mining: Algorithms and Applications Matrix Math Review

F Matrix Calculus F 1

1 Introduction to Matrices

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

A Primer on Index Notation

Lecture. S t = S t δ[s t ].

The Method of Partial Fractions Math 121 Calculus II Spring 2015

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

ARBITRAGE-FREE OPTION PRICING MODELS. Denis Bell. University of North Florida

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

Linear Algebra Notes for Marsden and Tromba Vector Calculus

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Notes on Symmetric Matrices

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8

CS229 Lecture notes. Andrew Ng

Linear Algebra Review and Reference

The Characteristic Polynomial

Similarity and Diagonalization. Similar Matrices

x = + x 2 + x

Linear and quadratic Taylor polynomials for functions of several variables.

Factorization Theorems

Inner Product Spaces

The Black-Scholes-Merton Approach to Pricing Options

Linear Algebra Review. Vectors

Moreover, under the risk neutral measure, it must be the case that (5) r t = µ t.

Vector and Matrix Norms

LS.6 Solution Matrices

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

3.1 Least squares in matrix form

MATH APPLIED MATRIX THEORY

Continued Fractions and the Euclidean Algorithm

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Solving Linear Systems, Continued and The Inverse of a Matrix

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances

Simulating Stochastic Differential Equations

Eigenvalues and Eigenvectors

Diusion processes. Olivier Scaillet. University of Geneva and Swiss Finance Institute

Does Black-Scholes framework for Option Pricing use Constant Volatilities and Interest Rates? New Solution for a New Problem

Mathematical Finance

Multiple regression - Matrices

5.3 The Cross Product in R 3

Math 312 Homework 1 Solutions

(a) We have x = 3 + 2t, y = 2 t, z = 6 so solving for t we get the symmetric equations. x 3 2. = 2 y, z = 6. t 2 2t + 1 = 0,

Orthogonal Diagonalization of Symmetric Matrices

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Pricing Formula for 3-Period Discrete Barrier Options

Lecture 5 Principal Minors and the Hessian

8 Square matrices continued: Determinants

n 2 + 4n + 3. The answer in decimal form (for the Blitz): 0, 75. Solution. (n + 1)(n + 3) = n lim m 2 1

Lecture L3 - Vectors, Matrices and Coordinate Transformations

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Matrix Differentiation

Notes on Determinant

How To Price A Call Option

On Market-Making and Delta-Hedging

1 Short Introduction to Time Series

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

DERIVATIVES AS MATRICES; CHAIN RULE

5 Numerical Differentiation

Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication).

CURVE FITTING LEAST SQUARES APPROXIMATION

A note on companion matrices

Recall that two vectors in are perpendicular or orthogonal provided that their dot

1 Lecture: Integration of rational functions by decomposition

ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE

RESULTANT AND DISCRIMINANT OF POLYNOMIALS

1.7. Partial Fractions Rational Functions and Partial Fractions. A rational function is a quotient of two polynomials: R(x) = P (x) Q(x).

expression is written horizontally. The Last terms ((2)( 4)) because they are the last terms of the two polynomials. This is called the FOIL method.

Partial Fractions Examples

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Brief Introduction to Vectors and Matrices

Høgskolen i Narvik Sivilingeniørutdanningen STE6237 ELEMENTMETODER. Oppgaver

Nonlinear Programming Methods.S2 Quadratic Programming

by the matrix A results in a vector which is a reflection of the given

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

Derivative Approximation by Finite Differences

7. DYNAMIC LIGHT SCATTERING 7.1 First order temporal autocorrelation function.

Integrals of Rational Functions

DIVERSIFICATION, REBALANCING, AND THE GEOMETRIC MEAN FRONTIER. William J. Bernstein. David Wilkinson

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Direct Methods for Solving Linear Systems. Matrix Factorization

7 Gaussian Elimination and LU Factorization

Section Inner Products and Norms

How the Greeks would have hedged correlation risk of foreign exchange options

Numerical Analysis Lecture Notes

Factoring Polynomials and Solving Quadratic Equations

Examination paper for TMA4115 Matematikk 3

Cointegration Pairs Trading Strategy On Derivatives 1

Solution to Homework 2

The Behavior of Bonds and Interest Rates. An Impossible Bond Pricing Model. 780 w Interest Rate Models

CBE 6333, R. Levicky 1. Tensor Notation.

CONTROLLABILITY. Chapter Reachable Set and Controllability. Suppose we have a linear system described by the state equation

1 IEOR 4700: Introduction to stochastic integration

A characterization of trace zero symmetric nonnegative 5x5 matrices

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA

MAT 242 Test 2 SOLUTIONS, FORM T

Transcription:

Stochastic Calculus, Courant Institute, Fall 0 http://www.math.nyu.edu/faculty/goodman/teaching/stochcalc0/index.html 1 Introduction Stochastic Calculus Backward Equation, duality Jonathan Goodman, Fall, 0 Backward equation and generator Suppose X t is a diffusion process described by the SDE dx t = a(x t ) dt + b(x t ) dw t. (1) Consider this to be a multi-component process with d components. This means that the SDE may be written in component form as dx j,t = a j (X t ) dt + b jk (X t ) dw k,t, for j = 1,, d. () k=1 The W k,t are independent standard Brownian motions. The coefficient b jk is the influence of Brownian motion W k on component X j of the process. These form the entries of the noise matrix b in the vector form (1). The number m is the number of sources of noise. We have seen that a non-diagonal b can be used to model correlations in the components dx j,t. In fact, a model might start with an idea of single component infinitesimal variance and cross-component correlation: σ j,t dt = var(dx j,t ), ρ jk,t = corr(dx j,t, dx k,t ). The noise coefficient matrix b then is constructed to match these modeling specifications. A modeling example with correlations It often happens that m = d, so b is a square matrix. If b is non-singular, we say the diffusion is non-degenerate. Degenerate diffusion models with m < d also are common. For example, suppose r t is the short rate, which is the interest rate paid for loans that are repaid very soon (the next day, often) and have no risk of not being repaid. A simple model of the fluctuations in the short rate might be dr t = γ(r r t ) dt + σdw t. (3) 1

This is an equilibrium model (because there is a steady state probability distribution) built on the idea that there is a natural short rate r that the fluctuating rate returns to with rate γ, while being taken away from r by a constant infinitesimal variance term σdw t. People used to criticize this model because it is possible to have r t < 0. In the mean time, several European countries have at some time experienced periods of a negative short rate. A money market account is an amount of money whose value changes only because of interest at the short rate. This is modeled by dm t = r t M t dt. (4) Together, the components r t and M t form a two component degenerate diffusion process that may be written in matrix/vector form as ( ) ( ) ( rt γ(r rt ) σ d = dt + dw M t r t 0) t. The matrix b is 1, with a single source of noise for a two component model. This model could be enriched by adding a risky asset S t that satisfies ds t = µs t dt + σs t dw t. (5) Of course, the σ in the r t equation (short rate volatility) is different from the σ in the risky asset equation (stock volatility). Moreover, we need to model correlations between interest rate and stock price changes. A natural to do this adapting the one factor market model that Markowitz used to model prices of different stocks. In that model, there is a market factor, which will be called dz 0,t, and there are idiosyncratic factors for each stochastic process, which will be called dz r,t and dz S,t. These three factors are taken to be independent. Correlations between market ds and dr are created by weighting factors differently. A market factor is a source of noise that effects all the financial instruments in a market in some way. Here, there is just one market factor (a one factor model), dz 0,t. The individual instrument prices have weightings β to the market, so that the σdw t in the r equation has a contribution β r dz 0,t. The σdw in the S equation has a contribution βsdz 0,t. The β factors are different but the noise dz 0,t is the same. This term is responsible for correlations between dr and ds. The instruments also have independent idiosyncratic factors, σ r and σ S. These are sources of uncertainty that affect one instrument and not the other. Idiosyncratic factors are responsible for the less than 100% correlation between the noisy instrument values. The SDEs that incorporate these modeling ideas are dr t = γ(r r t ) dt + β r dz 0,t + σ r dz r,t, (6) and ds t = µs t dt + S t ( β S dz 0,t + σ S dz S,t ). (7)

Using the independence of the three noise terms dz 0.t, dz r,t, and dz S,t, we find var(dr t ) = ( β r + σ r) dt (8) var(ds t ) = S t ( β S + σ S) dt (9) β r β S corr(dr t, ds t ) = β r + σr β S + σs (10) d r t M t S t γ(r r t ) β r σ r 0 = r t dt + 0 0 0 µs t S t β S 0 S t σ S dz 0,t dz r,t dz S,t. (11) The first row of this matrix equation is equivalent to the r equation (6). The second row is the noise-free M equation (4). The third row is the risky asset equation (7). This is in the form of the general SDE (1) with the identifications X t = r t M t S t a(x t ) = a(r t, M t, S t ) = γ(r r t) r t µs t dw t = dz 0,t dz r,t dz S,t β r σ r 0 b(r t, M t, S t ) = 0 0 0. (1) S t β S 0 S t σ S The d = 3 component diffusion process (11) seems to have m = 3 sources of noise, Z 0, Z r, and Z S. But this is misleading, because the noise matrix (1) does not have full rank. The process is a degenerate diffusion. In fact, the same model can be re-written using just two independent noise sources. Infinitesimal generator The stochastic process X t has the infinitesimal generator (usually just generator) Lg(x) = 1 µ jk (x) xj xk g + a j (x) xj g. (13) j=1 k=1 Let f(x, t) be a value function of the form Then f satisfies a backward equation j=1 f(x, t) = E[ V (X T ) X t = x]. t f + Lf = 0. (14) 3

This has been stated before, but here is a derivation. Before the derivation, some explanation of the notation (13) used to define the generator. The numbers µ jk form a d d matrix µ(x). This is related to the noise matrix b by µ(x) = b(x) b T (x). (15) If you want to work in component form, the entries of µ are given by µ jk = b jl b kl. (16) l=1 The formulas (15) or (16) are just covariance formulas. The double sum in the generator formula (13) may be written in matrix form in several ways. These use notations for the matrix of second partial derivatives of g. This is the hessian matrix of g, which may be denoted H, with entries H jk = xj xk. The hessian matrix is also written as H = D g. The double sum in (13) is the trace of the matrix product µ and D g j=1 k=1 µ jk (x) xj xk g = Tr(µD g). (17) If A is a square matrix, then the trace is the sum of the diagonal entries Tr(A) = A jj. It may seem surprising that this seemingly arbitrary sum is so useful. One explanation is the fact that the trace commutes even if the matrices do not. If A is d n and B is n d, then AB is d d and BA is n n. If n d, then AB BA, just because they are different size. Even in that case, the trace of the n n matrix is equal to the trace of the d d matrix Tr(AB) = Tr(BA). In particular, if A is diagonalizable with RAR 1 = Λ (here, Λ is the diagonal matrix with eigenvalues on the diagonal), then Tr(Λ) = Tr(RAR 1 ) = Tr [ R ( AR 1) ] = Tr [ ( AR 1) R ] = Tr [ A ( R 1 R ) ] Tr(Λ) = Tr(A). 4

This shows that the trace of A is the trace of Λ, which is the sum of the eigenvalues of A. If A and B are compatible for multiplication, then (AB) jl = l A jk B kl. Thus, Tr(AB) = j ( ) A jk B kj. If B is symmetric, we can replace B kj with B jk. This justifies the trace formula (17). Another notation for the double sum in (17) is µ::d g. You get the inner product of two vectors by multiplying corresponding components and adding the results. You can think of doing the same thing with two matrices. This is written A::B = A jk B jk. jk These are related by k A::B = Tr(AB T ). This is the same as AB if B is symmetric. Thus, the generator may be written in matrix/vector form as Lg(x) = 1 µ(x)::d g + a(x) g(x). Here is a derivation of the generator formula (13) from the general definition of generator E[ g(x t ) X 0 = x ] g(x) Lg(x) = lim. t 0 t For this, we use the Euler approximation X j, t = x j + a j (x) t + t b jk (x) ξ k, [ ξ k N (0, 1), i.i.d. ]. k=1 For the upcoming calculations, we write this in the form X j, t = x j + X j, X j = a j (x) t + t b jk (x) ξ k, (18) We expand g(x t ) in a Taylor series around x. We simplify by leaving out the argument x, writing g for g(x), j g for j g(x), etc. The result is g(x + X) = g + j g X j + 1 j=1 j=1 k=1 k=1 xj xk g X j X k + O( X 3 ). 5

The calculation is done to second order because X is on the order of t so X (the quadratic terms in the sum) are on the order of t. We do not calculate the X 3 terms because they are of order t 3 and vanish in the limit t 0 even when divided by t. We now find the expectation values needed to evaluate the generator. We start with the simpler terms: E[ X j ] = a j (x) t + O( t 3 ). This is because the terms ξ k have mean value zero. Thus E[ xj g X j ] = xj g a j (x) t + O( t 3 ). Next, we look at the quadratic terms. For these, use the fact that X j X k has a contribution of order t that involves products ξ i ξ l, but that all other terms in X j X k have powers t 3 or t, which do not contribute as t after dividing by t. Therefore, we calculate keeping only terms of order t: E[ X j X k ] = t i=1 l=1 b ji b kl E[ξ i ξ l ] + O( t 3 ). Now, use the fact that ξ i and ξ l are independent and mean zero if i l, while E[ξ l ξ l ] = 1. This leads to (using the relation µ = bb T ) E[ X j X k ] = t b ji b kl + O( t 3 ) = t µjk + O( t 3 ). l=1 This evaluates the hard part of the expectation E 1 xj xk g X j X k = 1 xj xk g µ jk t + O( t 3 ). jk This is the derivation of the generator formula (13) from the definition of the generator. It is one of the fundamental calculations of the subject of Stochastic Calculus. Backward equation, tower property The backward equation can be derived in generator form but here is a slightly different presentation that emphasizes the tower property. The tower property is that the conditional expectation of the conditional expectation is the conditional expectation. Specifically, suppose (U, V, W ) is a three component random variable. Let F (u, v, w) be some function of these variables and consider the the conditional expectations G(u, v) = E[ F (u, v, W ) U = u, V = v ] jk H(u) = E[ F (u, V, W ) U = u ]. 6

Then H(u) is the conditional expectation of F, conditioned on U = u. It is also the conditional expectation of G, which is the conditional expectation of F, conditioned on U = u and V = v. In formulas, H(u) = E[ G(u, V ) U = u ]. We apply this generic tower property to the three variables X t, X t+s, and X T, and to the function F = V (X T ). We assume t < t + s < T. We are interested in the conditional expectation of a function that depends on one of the three variables f9x) = E [V (X T ) X t = x ]. 7