Lecture 8: More Continuous Random Variables

Similar documents
Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Math 431 An Introduction to Probability. Final Exam Solutions

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Lecture 8. Confidence intervals and the central limit theorem

Section 5.1 Continuous Random Variables: Introduction

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

CHI-SQUARE: TESTING FOR GOODNESS OF FIT

5. Continuous Random Variables

4. Continuous Random Variables, the Pareto and Normal Distributions

1 Sufficient statistics

Random variables, probability distributions, binomial random variable

Chapter 7 Notes - Inference for Single Samples. You know already for a large sample, you can invoke the CLT so:

Normal Distribution. Definition A continuous random variable has a normal distribution if its probability density. f ( y ) = 1.

Normal distribution. ) 2 /2σ. 2π σ

Math/Stats 342: Solutions to Homework

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

An Introduction to Basic Statistics and Probability

Maximum Likelihood Estimation

EMPIRICAL FREQUENCY DISTRIBUTION

Questions and Answers


STAT 830 Convergence in Distribution

Determining distribution parameters from quantiles

Lecture 7: Continuous Random Variables

Notes on the Negative Binomial Distribution

Confidence Intervals for Cp

6.4 Normal Distribution

Gaussian Conjugate Prior Cheat Sheet

Review of Random Variables

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

The Standard Normal distribution

Approximation of Aggregate Losses Using Simulation

Lecture 5 : The Poisson Distribution

Generating Random Variables and Stochastic Processes

Chapter 3 RANDOM VARIATE GENERATION

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

Lecture 6: Discrete & Continuous Probability and Random Variables

Algebra. Exponents. Absolute Value. Simplify each of the following as much as possible. 2x y x + y y. xxx 3. x x x xx x. 1. Evaluate 5 and 123

Normal Distribution as an Approximation to the Binomial Distribution

( ) is proportional to ( 10 + x)!2. Calculate the

The Normal distribution

Statistics 100A Homework 7 Solutions

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Generating Random Variables and Stochastic Processes

2 Binomial, Poisson, Normal Distribution

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page

A logistic approximation to the cumulative normal distribution

1.1 Introduction, and Review of Probability Theory Random Variable, Range, Types of Random Variables CDF, PDF, Quantiles...

Exact Confidence Intervals

The normal approximation to the binomial

Math 461 Fall 2006 Test 2 Solutions

Practice Problems for Homework #6. Normal distribution and Central Limit Theorem.

Transformations and Expectations of random variables

Stats on the TI 83 and TI 84 Calculator

CHAPTER 5 Round-off errors

8. THE NORMAL DISTRIBUTION

Notes on Continuous Random Variables

Dongfeng Li. Autumn 2010

Conductance, the Normalized Laplacian, and Cheeger s Inequality

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

Lesson 20. Probability and Cumulative Distribution Functions

MATH 10: Elementary Statistics and Probability Chapter 5: Continuous Random Variables

Week 3&4: Z tables and the Sampling Distribution of X

Normal Approximation. Contents. 1 Normal Approximation. 1.1 Introduction. Anthony Tanbakuchi Department of Mathematics Pima Community College

A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University

Experimental Design. Power and Sample Size Determination. Proportions. Proportions. Confidence Interval for p. The Binomial Test

Joint Exam 1/P Sample Exam 1

Descriptive Statistics

Quadratic forms Cochran s theorem, degrees of freedom, and all that

1 The Brownian bridge construction

Statistics 100A Homework 4 Solutions

4.3 Areas under a Normal Curve

1.5 Oneway Analysis of Variance

Aggregate Loss Models

About the Gamma Function

2.6. Probability. In general the probability density of a random variable satisfies two conditions:

Math 151. Rumbos Spring Solutions to Assignment #22

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

WHERE DOES THE 10% CONDITION COME FROM?

Chapter 9 Monté Carlo Simulation

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

Chapter 4. Polynomial and Rational Functions. 4.1 Polynomial Functions and Their Graphs

Hypothesis Testing for Beginners

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Package SHELF. February 5, 2016

Parametric Survival Models

Pr(X = x) = f(x) = λe λx

INSURANCE RISK THEORY (Problems)

THE SINE PRODUCT FORMULA AND THE GAMMA FUNCTION

Lesson 4 Measures of Central Tendency

The Normal Distribution. Alan T. Arnholt Department of Mathematical Sciences Appalachian State University

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation

The Exponential Distribution

Confidence Intervals for the Difference Between Two Means

Stat 5102 Notes: Nonparametric Tests and. confidence interval

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

Homework 2 Solutions

Transcription:

Lecture 8: More Continuous Random Variables 26 September 2005 Last time: the eponential. Going from saying the density e λ, to f() λe λ, to the CDF F () e λ. Pictures of the pdf and CDF. Today: the Gaussian or normal distribution. (I prefer Gaussian.) Starting point: Make the density proportional to e z2 /2. This falls off much faster than the ordinary eponential. Also, we can have it work on both sides of the origin. (Why /2? It turns out to make life nicer for us later.) (Why z? Tradition.) To get the density, we need the normalizing factor, which is e z2 /2 dz. This is. There are several proofs, but none are really elementary. (There s a nice one involving contour integrals in the comple plane, for instance.) Still, trust me on this one. We now have a probability density, f(z) e z2 /2 which looks like this: This is the standard Gaussian or standard normal density. (I like Gaussian better.) It s the famous bell curve. The CDF of the standard Gaussian is an important function, so much so it has its own notation, Φ(z) (instead of the usual F (z)). It s also called the error function. Unfortunately, it doesn t have any nice analytical form, but there are well-known numerical approimations which are very accurate, at least as accurate at the numerical approimations for sine, cosine, logarithm, etc. From the symmetry of the pdf, though, we know that Φ(0) 0.5. Also, from symmetry, for z > 0, Pr (Z z) Pr (Z z) Φ(z) Φ( z) The CDF of the standard Gaussian is an important function, so much so it has its own notation, Φ(z) (instead of the usual F (z)). It s also called the error function. Unfortunately, it doesn t have any nice analytical form, but there are well-known numerical approimations which are very accurate, at least as accurate at the numerical approimations for sine, cosine, logarithm, etc.

standard Gaussian density 0.0 0. 0.2 0.3 0.4 4 2 0 2 4 Figure : Probability density of the standard Gaussian 2

Standard Gaussian distribution Cummulative probability 0.0 0.2 0.4 0.6 0.8.0 4 2 0 2 4 Figure 2: CDF Φ of the standard Gaussian 3

At this point we can already find the mean, using the fact that e z2 /2 is symmetric around the origin. E [Z] 0 0 0 z e z2 /2 dz z e z2 /2 dz + z e z2 /2 dz 0 0 z e z2 dz z e z2 /2 dz As for the variance, Var (Z) E [ Z 2] (E [Z]) 2 E [ Z 2]. So let s see that. E [ Z 2] z 2 e z2 /2 dz This doesn t cancel out, because z 2 ( z) 2. In fact, one can show that this integral works out to be eactly. So the standard Gaussian has mean zero and variance. What about non-standard Gaussians? Well, consider the density of X Z + µ. If a X b, then a µ Z b µ. Pr (a µ Z b µ) f X () zb µ za µ b a b a e z2 /2 dz e ( µ)2 /2 d f X ()d e ( µ)2 2 Now consider X σz, σ > 0. Pr (a X B) ( a Pr σ Z b ) σ f X () z b σ z a σ b a b a e z2 /2 dz e z2 /2σ 2 d σ f X ()d 2 e 2σ 2 σ 2 4

Putting these together, X σz + µ has the density f X () ( µ)2 e 2σ 2 σ 2 From the rules for epectation, E [X] σe [Z] + µ µ. From the rules for variance, Var (X) σ 2 Var (Z) σ 2. We say that the equation above gives the distribution of a Gaussian with mean µ and variance σ 2, N (µ, σ 2 ). An arbitrary Gaussian can be transformed into standard form by Z (X µ)/σ. By standardizing a Gaussian random variable, we keep ourselves from having to think about any distribution function other than Φ. Two numbers µ and σ completely determine the distribution, and completely specify the transformation back to standard form; they re the parameters. The sum of two Gaussian variables is always another Gaussian, whether or not they are independent. A Gaussian times a constant is Gaussian. A Gaussian plus a constant is Gaussian. Here are some important facts: Pr ( Z ) Φ() Φ( ) 0.6826895 Pr ( Z 2) Φ(2) Φ( 2) 0.9544997 Pr ( Z 3) Φ(3) Φ( 3) 0.9973002 By applying the standardization above, we can see that a Gaussian falls within one standard deviation of its mean about two-thirds of the time, within two standard deviations about ninety-five percent of the time, and within three standard deviations about 99.7% percent of the time. Remember the Chebyshev inequality from before: If X is Gaussian, then Pr ( X µ kσ) Var (X) σ 2 Pr ( X µ kσ) (Φ(k) Φ( k)) Let s compare the probability of large deviations from the mean given by Chebyshev to that obtained by the Gaussian. Clearly, the Gaussian distribution is much more tightly concentrated around its mean than arbitrary distributions. This will be very useful later. We will see starting net week that the sum of independent random variables tends towards a Gaussian; this is why it is so important and so ubiquitous. In particular, let s look at a binomial distribution with contant p and growing n. I claim that X Bin(n, p) approaches in distribution N (np, np( p)). Let s look at some plots, keeping p 0.5 for simplicity. There are a number of other distributions which are closely related to the Gaussian. 5

Prob( X mu > k sigma) 0.0 0.2 0.4 0.6 0.8.0 0 2 3 4 k Prob( X mu > ksigma) e 5 e e 07 e 03 0 2 4 6 8 k Figure 3: Comparison of the probability of large deviations for the Gaussian distribution (solid line) against the Chebyshev bounds, which apply to all distributions (dashed line). The lower figure shows the probability on a log scale. 6

p() 0.05 0.0 0.5 0.20 0.25 0.30 0.35 0 2 3 4 p() 0.00 0.05 0.0 0.5 0.20 0.25 0 2 4 6 8 0 p() 0.00 0.05 0.0 0.5 0 5 0 5 20 25 30 Figure 4: Binomial distribution (solid line) versus a Gaussian density with the same mean and variance (dotted line), with p 0.5 and n 4, 0 and 30, respectively. 7

χ 2 distribution: if Z is a standard Gaussian, then the distribution of Z 2 is said to be χ 2 with one degree of freedom : f χ 2,() /2 e /2 E [X] Var (X) 2 This comes up a lot in statistical testing, especially in the following variant form. If Z, Z 2... Z n are all independent standard normals, then n i Z2 i has a χ 2 distribution with n degrees of freedom: f χ2,n n/2 e /2 Γ(n/2)2 n/2 E [X] n Var (X) 2n where Γ() 0 t e tdt, the gamma function. If is an integer, then Γ() ( )!. For any, Γ() ( )Γ( ). Finally, Γ(/2) π. The χ 2 distribution is a special case of a broader family of distributions called the γ distributions, which have the pdf f γ(α,θ) () α e /θ Γ(α)θ α They re defined over non-negative. If α >, then for small this looks like α, and rises, but for large it always looks like an eponential. In fact, the eponential is the space case where α. For more, see the book. If Y log X is Gaussian, we say that X is log-gaussian or lognormal. Then X 0, and the distribution is specified by the mean and variance of Y, i.e. of log X, call them M and S 2. In eplicit form, (log M) 2 f lognormal () e 2S 2 S 2 S2 M+ E [X] e 2 ( ) Var (X) e 2M+S2 e S2 The lognormal is the theoretical distribution that best fits the observed distribution of income and wealth. The mean and variance vary from country to country, but the general form doesn t. One consequence of this is that rich people are really a lot richer than ordinary people, and that the mean isn t so representative as the median. The median income is after all e M, but the mean is much higher! For instance, median household income in the United States is $42,228 (as of 200), and the standard deviation of the log income is about 8

probability density 0.00 0.05 0.0 0.5 0.20 0.25 0 20 40 60 80 cummulative probability 0.0 0.2 0.4 0.6 0.8.0 0 20 40 60 80 Figure 5: PDF and CDF of a gamma distribution with α 3 and θ. 9

0.879. This gives a mean income of about $62 thousand, which is very close to the true value of $58 thousand. The main problem with the log-normal model of income distribution is that there are slightly more really poor people than it allows for, and the rich people are much, much richer. 0

probability density 0.0e+00 5.0e 06.0e 05.5e 05 0e+00 2e+04 4e+04 6e+04 8e+04 e+05 Income ($) Cummulative probability 0.0 0.2 0.4 0.6 0.8.0 0 50000 50000 250000 Income ($) Figure 6: Density and CDF of a lognormal distribution with M 42, 228 and S 0.879, based on US income data.