Transformations and Expectations of random variables



Similar documents
Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Random variables, probability distributions, binomial random variable

( ) is proportional to ( 10 + x)!2. Calculate the

Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

THE CENTRAL LIMIT THEOREM TORONTO

Manual for SOA Exam MLC.

Math 431 An Introduction to Probability. Final Exam Solutions

M2S1 Lecture Notes. G. A. Young ayoung

1.1 Introduction, and Review of Probability Theory Random Variable, Range, Types of Random Variables CDF, PDF, Quantiles...

RAJALAKSHMI ENGINEERING COLLEGE MA 2161 UNIT I - ORDINARY DIFFERENTIAL EQUATIONS PART A

Introduction to Probability

Math 526: Brownian Motion Notes

Lecture 8: More Continuous Random Variables

ST 371 (IV): Discrete Random Variables

CITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION

BNG 202 Biomechanics Lab. Descriptive statistics and probability distributions I

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

Probability and Random Variables. Generation of random variables (r.v.)

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

STA 256: Statistics and Probability I

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Aggregate Loss Models

Lecture 6: Discrete & Continuous Probability and Random Variables

THE SINE PRODUCT FORMULA AND THE GAMMA FUNCTION

5.3 Improper Integrals Involving Rational and Exponential Functions

University of California, Los Angeles Department of Statistics. Random variables

Section 5.1 Continuous Random Variables: Introduction

Review of Random Variables

Approximation of Aggregate Losses Using Simulation

Order Statistics. Lecture 15: Order Statistics. Notation Detour. Order Statistics, cont.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS

1 if 1 x 0 1 if 0 x 1

f x a 0 n 1 a 0 a 1 cos x a 2 cos 2x a 3 cos 3x b 1 sin x b 2 sin 2x b 3 sin 3x a n cos nx b n sin nx n 1 f x dx y

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Probability Generating Functions

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Calculus. Contents. Paul Sutcliffe. Office: CM212a.

Manual for SOA Exam MLC.

About the Gamma Function

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Lectures 5-6: Taylor Series

Financial Time Series Analysis (FTSA) Lecture 1: Introduction

Probability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X

1.5 / 1 -- Communication Networks II (Görg) Transforms

Applications to Data Smoothing and Image Processing I

Systems with Persistent Memory: the Observation Inequality Problems and Solutions

A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Center: Finding the Median. Median. Spread: Home on the Range. Center: Finding the Median (cont.)

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

SOLUTIONS. f x = 6x 2 6xy 24x, f y = 3x 2 6y. To find the critical points, we solve

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference

Math 151. Rumbos Spring Solutions to Assignment #22

Math 370, Actuarial Problemsolving Spring 2008 A.J. Hildebrand. Practice Test, 1/28/2008 (with solutions)

z 0 and y even had the form

5. Continuous Random Variables

IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS

Algebra I Vocabulary Cards

Microeconomic Theory: Basic Math Concepts

Nonparametric adaptive age replacement with a one-cycle criterion

ARBITRAGE-FREE OPTION PRICING MODELS. Denis Bell. University of North Florida

Chapter 4 Lecture Notes

Paper II ( CALCULUS ) Shahada. College, Navapur. College, Shahada. Nandurbar

m i: is the mass of each particle

Properties of moments of random variables

Math 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 3 Solutions

TOPIC 4: DERIVATIVES

Permanents, Order Statistics, Outliers, and Robustness

Electrostatic Fields: Coulomb s Law & the Electric Field Intensity

Joint Exam 1/P Sample Exam 1

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Pricing Formula for 3-Period Discrete Barrier Options

MBA 611 STATISTICS AND QUANTITATIVE METHODS

Lecture 7: Finding Lyapunov Functions 1

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

e.g. arrival of a customer to a service station or breakdown of a component in some system.

Example 4.1 (nonlinear pendulum dynamics with friction) Figure 4.1: Pendulum. asin. k, a, and b. We study stability of the origin x

Lecture 18: The Time-Bandwidth Product

Some probability and statistics

Trend and Seasonal Components

Solutions for Review Problems

ECE302 Spring 2006 HW3 Solutions February 2,

An Introduction to Partial Differential Equations

Random Variate Generation (Part 3)

TRANSACTIONS JUNE, 1969 AN UPPER BOUND ON THE STOP-LOSS NET PREMIUM--ACTUARIAL NOTE NEWTON L. BOWERS, JR. ABSTRACT

Probability and statistics; Rehearsal for pattern recognition

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

4. Complex integration: Cauchy integral theorem and Cauchy integral formulas. Definite integral of a complex-valued function of a real variable

Chapter 5 Discrete Probability Distribution. Learning objectives

6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:30-9:30 PM. SOLUTIONS

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan , Fall 2010

Lectures on Stochastic Processes. William G. Faris

Probability Theory. Florian Herzog. A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T..

LOGNORMAL MODEL FOR STOCK PRICES

EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines

Example. A casino offers the following bets (the fairest bets in the casino!) 1 You get $0 (i.e., you can walk away)

Chapter 8 - Power Density Spectrum

Lecture 6 Black-Scholes PDE

Transcription:

Transformations and Epectations of random variables X F X (): a random variable X distributed with CDF F X. Any function Y = g(x) is also a random variable. If both X, and Y are continuous random variables, can we find a simple way to characterize F Y and f Y (the CDF and PDF of Y ), based on the CDF and PDF of X? For the CDF: PDF: f Y (y) = F y(y) F Y (y) = P Y (Y y) = P Y (g(x) y) = P X ( X : g(x) y) (X is sample space for X) = f X (s)ds. { X :g(x) y} Caution: need to consider support of y. Consider several eamples:. X U[, ] and y = ep() That is: f X () = { 2 if [, ] 0 otherwise F X () = 2 +, for [, ]. 2 F Y (y) = P rob(ep(x) y) = P rob(x log y) Be careful about the bounds of the support! = F X (log y) = 2 + 2 log y, for y [ e, e]. f Y (y) = y F Y (y) = f X (log y) y = 2y, for y [ e, e].

2. X U[, ] and Y = X 2 F Y (y) = P rob(x 2 y) = P rob( y X y) = F X ( y) F X ( y) = 2F X ( y), by symmetry: F X ( y) = F X ( y). f Y (y) = y F Y (y) = 2f X ( y) 2 y = 2, for y [0, ]. y As the first eample above showed, it s easy to derive the CDF and PDF of Y when g( ) is a strictly monotonic function: Theorems 2..3, 2..5: When g( ) is a strictly increasing function, then F Y (y) = g (y) f X ()d = F X (g (y)) f Y (y) = f X (g (y)) y g (y) using chain rule. Note: by the inverse function theorem, y g (y) = / [g ()] =g (y). When g( ) is a strictly decreasing function, then F Y (y) = g (y) f X ()d = F X (g (y)) f Y (y) = f X (g (y)) y g (y) using chain rule. These are the change of variables formulas for transformations of univariate random variables. Thm 2..8 generalizes this to piecewise monotonic transformations. 2

Here is a special case of a transformation: Thm 2..0: Let X have a continuous CDF F X ( ) and define the random variable Y = F X (X). Then Y U[0, ], i.e., F Y (y) = y, for y [0, ]. Note: all that is required is that the CDF F X is continuous, not that it must be strictly increasing. The result also goes through when F X is continuous but has flat parts (cf. discussion in CB, pg. 34). Epected value (Definition 2.2.): The epected value, or mean, of a random variable g(x) is { Eg(X) = g()f X()d if X continuous X g()p (X = ) if X discrete provided that the integral or the sum eists The epectation is a linear operator (just like integration): so that [ ] n n E α g i (X) + b = α Eg i (X) + b. i= Note: Epectation is a population average, i.e., you average values of the random variable g(x) weighting by the population density f X (). A statistical eperiment yields sample observations X, X 2,..., X n F X. From these sample observations, we can calculate sample avg. Xn n i X i. In general: Xn EX. But under some conditions, as n, then X n EX in some sense (which we discuss later). Epected value is commonly used measure of central tendency of a random variable X. Eample: But mean may not eist: Cauchy random variable with density f() = π(+ 2 ) for (, ). Note that π( + 2 ) d = 0 = 0 lim a π( + 2 ) d + 0 a i= d + lim π( + 2 ) b 0 π( + 2 ) d b 0 b = lim 0 a [log( 2 )] 0 a + lim = + undefined π( + 2 ) d [log( 2 )] b 0 3

Other measures:. Median: med(x) = m such that F X () = 0.5. Robust to outliers, and has nice invariance property: for Y = g(x) and g( ) monotonic increasing, then med(y ) = g(med(x)). 2. Mode: Mode(X) = ma f X (). Moments: important class of epectations For each integer n, the n-th (uncentred) moment of X F X ( ) is µ n EX n. The n-th centred moment is µ n E(X µ) n = E(X EX) n. (It is centred around the mean EX.) For n = 2: µ 2 = E(X EX) 2 is the Variance of X. µ 2 is the standard deviation. Important formulas: V ar(ax + b) = a 2 V arx (variance is not a linear operation) V arx = E(X 2 ) (EX) 2 : alternative formula for the variance The moments of a random variable are summarized in the moment generating function. Definition: the moment-generating function of X is M X (t) E ep(tx), provided that the epectation eists in some neighborhood t [ h, h] of zero. This is also called the Laplace transform. Specifically: { M (t) = et f X ()d for X continuous X et P (X = ) for X discrete. The uncentred moments of X are generated from this function by: EX n = M (n) dn X (0) dt M X(t) n, t=0 4

which is the n-th derivative of the MGF, evaluated at t = 0. Eample: standard normal distribution: ) M X (t) = ep (t 2 d 2 ( = ep ) 2 (( t)2 t 2 ) d = ep( 2 t2 ) ep ( 2 ) ( t)2 d = ep( 2 t2 ) where last term on RHS is integral over density function of N(t, ), which integrates to one. First moment: EX = M X (0) = t ep( 2 t2 ) t=0 = 0. Second moment: EX 2 = M 2 X (0) = ep( 2 t2 ) + t 2 ep( 2 t2 ) =. In many cases, the MGF can characterize a distribution. But problem is that it may not eist (eg. Cauchy distribution) For a RV X, is its distribution uniquely determined by its moment generating function? Thm 2.3.: For X F X and Y F Y, if M X and M Y eist, and M X (t) = M Y (t) for all t in some neighborhood of zero, then F X (u) = F Y (u) for all u. Note that if the MGF eists, then it characterizes a random variable with an infinite number of moments (because the MGF is infinitely differentiable). Converse not necessarily true. (e. log-normal random variable: X N(0, ), Y = ep(x)) Characteristic function: The characteristic function of a random variable g(), defined as where f() is the density for. φ g() (t) = E ep(itg()) = This is also called the Fourier transform. Features of characteristic function: 5 ep(itg())f()d

The CF always eists. This follows from the equality e it = cos(t) + i sin(t), and both the real and comple parts of the integrand are bounded functions. Consider a symmetric density function, with f( ) = f() (symmetric around zero). Then resulting φ(t) is real-valued, and symmetric around zero. The CF completely determines the distribution of X (every cdf has a unique characteristic function). Let X have characteristic function φ X (t). Then Y = ax +b has characteristic function φ Y (t) = e ibt φ X (at). X and Y, independent, with characteristic functions φ X (t) and φ Y (t). Then φ X+Y (t) = φ X (t)φ Y (t) φ(0) =. For a given characteristic function φ X (t) such that φ X(t) dt <, the corresponding density f X () is given by the inverse Fourier transform, which is f X () = φ X (t) ep( it)dt. Eample: N(0, ) distribution, with density f() = ep( 2 /2). Take as given that the characteristic function of N(0, ) is φ N(0,) (t) = ep ( it 2 /2) ) d = ep( t 2 /2). () Hence the inversion formula yields f() = Now making substitution z = t, we get ep ( iz z 2 /2 ) dz ep( t 2 /2) ep( it)dt. = φ N(0,) () = ep( 2 /2) = f N(0,) (). (Use Eq. ()) Here denotes the modulus of a comple number. For + iy, we have + iy = 2 + y 2. 6

Characteristic function also summarizes the moments of a random variable. Specifically, note that the h-th derivative of φ(t) is φ h (t) = i h g() h ep(itg())f()d. (2) Hence, assuming the h-th moment, denoted µ h g() E[g()]h eists, it is equal to µ h g() = φ h (0)/i h. Hence, assuming that the required moments eist, we can epand the characteristic function around t = 0 to get: φ(t) = + it µ g() + (it)2 2 µ2 g() +... + (it)k µ k g() + o(t k ). k! Cauchy distribution, cont d: The characteristic function for the Cauchy distribution is φ(t) = ep( t ). This is not differentiable at t = 0, which by Eq. (2) is saying that its mean does not eist. Hence, the epansion of the characteristic function in this case is invalid. 7