General Bivariate Normal. Lecture 22: Bivariate Normal Distribution. General Bivariate Normal - Cov/Corr. General Bivariate Normal - Marginals

Similar documents
Sections 2.11 and 5.8

Data Mining: Algorithms and Applications Matrix Math Review

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Multivariate Normal Distribution Rebecca Jennings, Mary Wakeman-Linn, Xin Zhao November 11, 2010

6. Cholesky factorization

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Some probability and statistics

Multivariate Normal Distribution

CS229 Lecture notes. Andrew Ng

1 Introduction to Matrices

Introduction to General and Generalized Linear Models

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

The Monte Carlo Framework, Examples from Finance and Generating Correlated Random Variables

Slides for Risk Management VaR and Expected Shortfall

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Chapter 4: Vector Autoregressive Models

Quadratic forms Cochran s theorem, degrees of freedom, and all that

Slides for Risk Management

Lecture Notes 1. Brief Review of Basic Probability

Chapter 2 Portfolio Management and the Capital Asset Pricing Model

Math 312 Homework 1 Solutions

y t by left multiplication with 1 (L) as y t = 1 (L) t =ª(L) t 2.5 Variance decomposition and innovation accounting Consider the VAR(p) model where

Understanding and Applying Kalman Filtering

Introduction to Matrix Algebra

Lecture 21. The Multivariate Normal Distribution

General Sampling Methods

Monte Carlo Simulation

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

( ) is proportional to ( 10 + x)!2. Calculate the

Gaussian Conjugate Prior Cheat Sheet

Stat 704 Data Analysis I Probability Review

Introduction to Multivariate Analysis

Lecture 9: Introduction to Pattern Analysis

Orthogonal Projections

1 Determinants and the Solvability of Linear Systems

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = i.

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Inner product. Definition of inner product

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Systematic risk modelisation in credit risk insurance

SYSTEMS OF REGRESSION EQUATIONS

The Bivariate Normal Distribution

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Lecture 2 of 4-part series. Spring School on Risk Management, Insurance and Finance European University at St. Petersburg, Russia.

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix

Name: Math 29 Probability. Practice Second Midterm Exam Show all work. You may receive partial credit for partially completed problems.

Least Squares Estimation

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices

Lecture 14. Chapter 7: Probability. Rule 1: Rule 2: Rule 3: Nancy Pfenning Stats 1000

Chapter 17. Orthogonal Matrices and Symmetries of Space

A.II. Kernel Estimation of Densities

State Space Time Series Analysis

[1] Diagonal factorization

8. Linear least-squares

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

5. Orthogonal matrices

Solving Systems of Linear Equations

LS.6 Solution Matrices

Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression

Normal Distribution. Definition A continuous random variable has a normal distribution if its probability density. f ( y ) = 1.

Lecture 4: Joint probability distributions; covariance; correlation

Direct Methods for Solving Linear Systems. Matrix Factorization

Pricing of a worst of option using a Copula method M AXIME MALGRAT

Variance Reduction. Pricing American Options. Monte Carlo Option Pricing. Delta and Common Random Numbers

Matrix Differentiation

Data Mining: Exploring Data. Lecture Notes for Chapter 3. Slides by Tan, Steinbach, Kumar adapted by Michael Hahsler

NOTES ON LINEAR TRANSFORMATIONS

The Characteristic Polynomial

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Name: Section Registered In:

Lecture 8: More Continuous Random Variables

So, using the new notation, P X,Y (0,1) =.08 This is the value which the joint probability function for X and Y takes when X=0 and Y=1.

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Hedging Illiquid FX Options: An Empirical Analysis of Alternative Hedging Strategies

Lecture 8: Signal Detection and Noise Assumption

DETERMINANTS TERRY A. LORING

Similarity and Diagonalization. Similar Matrices

A Tutorial on Probability Theory

Lecture notes on linear algebra

STATISTICS AND DATA ANALYSIS IN GEOLOGY, 3rd ed. Clarificationof zonationprocedure described onpp

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

The Analysis of Data. Volume 1. Probability. Guy Lebanon

IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS

Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

A =

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

University of Lille I PC first year list of exercises n 7. Review

HURDLE AND SELECTION MODELS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics July 2009

Introduction to Matrices for Engineers

CONDITIONAL, PARTIAL AND RANK CORRELATION FOR THE ELLIPTICAL COPULA; DEPENDENCE MODELLING IN UNCERTAINTY ANALYSIS

STAT 830 Convergence in Distribution

BNG 202 Biomechanics Lab. Descriptive statistics and probability distributions I

Copula Simulation in Portfolio Allocation Decisions

Linear Threshold Units

Master s Theory Exam Spring 2006

CS3220 Lecture Notes: QR factorization and orthogonal transformations

Component Ordering in Independent Component Analysis Based on Data Power

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

Transcription:

General Bivariate Normal Lecture : Bivariate Normal Distribution Statistics 04 Colin Rundel April, 0 Let Z, Z N 0,, which we will use to build a general bivariate normal distribution. f z, z π exp z + z We want to transform these unit normal distributions to have the follow arbitrary parameters: µ, µ Y,, σ Y, ρ Z + µ Y σ Y ρz + ρ Z + µ Y Statistics 04 Colin Rundel Lecture April, 0 / General Bivariate Normal - Marginals General Bivariate Normal - Cov/Corr First, lets examine the marginal distributions of and Y, Second, we can find Cov, Y and ρ, Y Z + µ N 0, + µ N µ, σ Y σ Y ρz + ρ Z + µ Y σ Y ρn 0, + ρ N 0, + µ Y σ Y N 0, ρ + N 0, ρ + µ Y σ Y N 0, + µ Y N µ Y, σy Cov, Y E E Y EY E Z + µ µ σ Y ρz + ρ Z + µ Y µ Y E Z σ Y ρz + ρ Z σ Y E ρz + ρ Z Z σ Y ρez σ Y ρ ρ, Y Cov, Y σ Y ρ Statistics 04 Colin Rundel Lecture April, 0 / Statistics 04 Colin Rundel Lecture April, 0 3 /

General Bivariate Normal - RNG Multivariate Change of Variables Consequently, if we want to generate a Bivariate Normal random variable with N µ, σ and Y N µ Y, σy where the correlation of and Y is ρ we can generate two independent unit normals Z and Z and use the transformation: Z + µ Y σ Y ρz + ρ Z + µ Y We can also use this result to find the joint density of the Bivariate Normal using a d change of variables. Let,..., n have a continuous joint distribution with pdf f defined of S. We can define n new random variables Y,..., Y n as follows: Y r,..., n Y n r n,..., n If we assume that the n functions r,..., r n define a one-to-one differentiable transformation from S to T then let the inverse of this transformation be x s y,..., y n x n s n y,..., y n Then the joint pdf g of Y,..., Y n is { f s,..., s n J for y,..., y n T gy,..., y n 0 otherwise Where s y J det....... s n y s y n s n y n Statistics 04 Colin Rundel Lecture April, 0 4 / Statistics 04 Colin Rundel Lecture April, 0 5 / General Bivariate Normal - Density General Bivariate Normal - Density The first thing we need to find are the inverses of the transformation. If x r z, z and y r z, z we need to find functions h and h such that Z s, Y and Z s, Y. Next we calculate the Jacobian, J det s x s x s y s y det 0 ρ ρ σ Y ρ σ Y ρ Therefore, Y µ Y σ Y Z + µ Z µ Y σ Y ρz + ρ Z + µ Y ρ µ + ρ Z Y µy Z ρ σ Y ρ µ s x, y x µ y µy s x, y ρ σ Y ρ x µ The joint density of and Y is then given by f x, y f z, z J π exp z + z J π σ Y ρ exp π σ Y ρ exp z + z x µ + y µy ρ σ Y x µ π σ Y ρ exp / ρ σ + y µ Y σ Y ρ x µ ρ x µ y µ Y σ Y Statistics 04 Colin Rundel Lecture April, 0 6 / Statistics 04 Colin Rundel Lecture April, 0 7 /

General Bivariate Normal - Density Matrix Notation Obviously, the density for the Bivariate Normal is ugly, and it only gets worse when we consider higher dimensional joint densities of normals. We can write the density in a more compact form using matrix notation, x µ σ x µ Σ ρ σ Y y µ Y ρ σ Y f x π det Σ / exp x µt Σ x µ We can confirm our results by checking the value of det Σ / and x µ T Σ x µ for the bivariate case. det Σ / σ σ Y ρ σ / σ Y σ Y ρ / σ Y General Bivariate Normal - Density Matrix Notation Recall for a matrix, Then, a b A c d A det A d b c a x µ T Σ x µ T x µx σ σ Y σ Y ρ y µ y ρ σ Y σ σ Y ρ ad bc ρ σ Y σ σ Y x µ ρ σ Y y µ Y ρ σ Y x µ + σ y µ Y d b c a x µx y µ y T x µx y µ y σ σ σ Y ρ Y x µ ρ σ Y x µ y µ Y + σ y µ Y x µ ρ σ ρ x µ y µ Y + y µ Y σ Y σy Statistics 04 Colin Rundel Lecture April, 0 8 / Statistics 04 Colin Rundel Lecture April, 0 9 / N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, ρ 0 ρ 0 ρ 0 ρ 0.5 ρ 0.5 ρ 0.75 Statistics 04 Colin Rundel Lecture April, 0 0 / Statistics 04 Colin Rundel Lecture April, 0 /

N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, N 0,, Y N 0, ρ 0.5 ρ 0.5 ρ 0.75 ρ 0.75 ρ 0.75 ρ 0.75 Statistics 04 Colin Rundel Lecture April, 0 / Statistics 04 Colin Rundel Lecture April, 0 3 / Multivariate Normal Distribution Multivariate Normal Distribution - Cholesky Matrix notation allows us to easily express the density of the multivariate normal distribution for an arbitrary number of dimensions. We express the k-dimensional multivariate normal distribution as follows, N k µ, Σ where µ is the k column vector of means and Σ is the k k covariance matrix where {Σ} i,j Cov i, j. The density of the distribution is f x π k/ det Σ / exp x µt Σ x µ In the bivariate case, we had a nice transformation such that we could generate two independent unit normal values and transform them into a sample from an arbitrary bivariate normal distribution. There is a similar method for the multivariate normal distribution that takes advantage of the Cholesky decomposition of the covariance matrix. The Cholesky decomposition is defined for a symmetric, positive definite matrix as L Chol where L is a lower triangular matrix such that LL T. Statistics 04 Colin Rundel Lecture April, 0 4 / Statistics 04 Colin Rundel Lecture April, 0 5 /

Multivariate Normal Distribution - RNG Let Z,..., Z k N 0, and Z Z,..., Z k T then µ + CholΣZ N k µ, Σ this is offered without proof in the general k-dimensional case but we can check that this results in the same transformation we started with in the bivariate case and should justify how we knew to use that particular transformation. Cholesky and the Bivariate Transformation We need to find the Cholesky decomposition of Σ for the general bivariate case where σ Σ ρ σ Y ρ σ Y σy We need to solve the following for a, b, c a 0 b c a b 0 c a ab ab b + c σ ρ σ Y ρ σ Y σy This gives us three unique equations and three unknowns to solve for, a σ ab ρ σ Y b + c σ Y a b ρ σ Y /a ρσ Y c σy b σ Y ρ / Statistics 04 Colin Rundel Lecture April, 0 6 / Statistics 04 Colin Rundel Lecture April, 0 7 / Cholesky and the Bivariate Transformation Conditional Expectation of the Bivariate Normal Let Z, Z N 0, then Y µ + CholΣZ µ µ Y µ µ Y σ 0 + + µ + Z ρσ Y σ Y ρ / Z Z ρσ Y Z + σ Y ρ / Z Y µ Y + σ Y ρz + ρ / Z Z Using µ + Z and Y µ Y + σ Y ρz + ρ / Z where Z, Z N 0, we can find EY. EY x E µ Y + σ Y ρz + ρ / Z x E µ Y + σ Y ρ x µ + ρ / Z x µ Y + σ Y ρ x µ + ρ / EZ x x µ µ Y + σ Y ρ By symmetry, y µy E Y y µ + ρ σ Y Statistics 04 Colin Rundel Lecture April, 0 8 / Statistics 04 Colin Rundel Lecture April, 0 9 /

Conditional Variance of the Bivariate Normal Using µ + Z and Y µ Y + σ Y ρz + ρ / Z where Z, Z N 0, we can find VarY. VarY x Var µ Y + σ Y ρz + ρ / Z x Var µ Y + σ Y ρ x µ + ρ / Z x Varσ Y ρ Z x σ Y ρ Example - Husbands and Wives Example 5.0.6, degroot Suppose that the heights of married couples can be explained by a bivariate normal distribution. If the wives have a mean heigh of 66.8 inches and a standard deviation of inches while the heights of the husbands have a mean of 70 inches and a standard deviation of inches. The correlation between the heights is 0.68. What is the probability that for a randomly selected couple the wife is taller than her husband? By symmetry, Var Y y σ ρ Statistics 04 Colin Rundel Lecture April, 0 0 / Statistics 04 Colin Rundel Lecture April, 0 / Example - Conditionals Example - Conditionals, cont. Suppose that and have a bivariate normal distribution where E 3.7 0.5, E 0.4 0.6, and Var 3.64. Find E, Var, E, Var, and ρ,. Statistics 04 Colin Rundel Lecture April, 0 / Statistics 04 Colin Rundel Lecture April, 0 3 /