Section 6.1 Joint Distribution Functions

Similar documents
Feb 28 Homework Solutions Math 151, Winter Chapter 6 Problems (pages )

Math 432 HW 2.5 Solutions

Section 5.1 Continuous Random Variables: Introduction

Practice with Proofs

Statistics 100A Homework 7 Solutions

Maximum Likelihood Estimation

MULTIVARIATE PROBABILITY DISTRIBUTIONS

M2S1 Lecture Notes. G. A. Young ayoung

Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

5. Continuous Random Variables

Lecture 7: Continuous Random Variables

STA 256: Statistics and Probability I

Lecture 6: Discrete & Continuous Probability and Random Variables

INSURANCE RISK THEORY (Problems)

MAS108 Probability I

Stats on the TI 83 and TI 84 Calculator

DERIVATIVES AS MATRICES; CHAIN RULE

Microeconomic Theory: Basic Math Concepts

6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:30-9:30 PM. SOLUTIONS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Notes on Continuous Random Variables

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

ECE302 Spring 2006 HW5 Solutions February 21,

E3: PROBABILITY AND STATISTICS lecture notes

e.g. arrival of a customer to a service station or breakdown of a component in some system.

Calculus 1: Sample Questions, Final Exam, Solutions

Statistics 100A Homework 8 Solutions

. (3.3) n Note that supremum (3.2) must occur at one of the observed values x i or to the left of x i.

1 Sufficient statistics

Chapter 4 - Lecture 1 Probability Density Functions and Cumul. Distribution Functions

Practice problems for Homework 11 - Point Estimation

Chapter 17. Review. 1. Vector Fields (Section 17.1)

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

2WB05 Simulation Lecture 8: Generating random variables

Student Outcomes. Lesson Notes. Classwork. Discussion (10 minutes)

Probability Calculator

Probability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X

Lecture Notes 1. Brief Review of Basic Probability

The Mean Value Theorem

Properties of Future Lifetime Distributions and Estimation

Vector and Matrix Norms

TOPIC 4: DERIVATIVES

Lecture 5 Principal Minors and the Hessian

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL

Scalar Valued Functions of Several Variables; the Gradient Vector

Probability and Random Variables. Generation of random variables (r.v.)

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference

Chapter 3 RANDOM VARIATE GENERATION

Chapter 4 Lecture Notes

Lecture Notes on Elasticity of Substitution

4. Continuous Random Variables, the Pareto and Normal Distributions

1. Prove that the empty set is a subset of every set.

The normal approximation to the binomial

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

Statistics 100A Homework 4 Solutions

CITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION

Normal distribution. ) 2 /2σ. 2π σ

RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS

Let H and J be as in the above lemma. The result of the lemma shows that the integral

CHAPTER FIVE. Solutions for Section 5.1. Skill Refresher. Exercises

Principle of Data Reduction

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

Final Mathematics 5010, Section 1, Fall 2004 Instructor: D.A. Levin

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Math 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 2 Solutions

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

3 Contour integrals and Cauchy s Theorem

Department of Civil Engineering-I.I.T. Delhi CEL 899: Environmental Risk Assessment Statistics and Probability Example Part 1

Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Ex. 2.1 (Davide Basilio Bartolini)

36 CHAPTER 1. LIMITS AND CONTINUITY. Figure 1.17: At which points is f not continuous?

STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE

Lecture 13: Martingales

Manual for SOA Exam MLC.

Introduction to Probability

Inequalities of Analysis. Andrejs Treibergs. Fall 2014

Class Meeting # 1: Introduction to PDEs

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Second Order Linear Partial Differential Equations. Part I

Partial Fractions Examples

Systems of Linear Equations

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

More Quadratic Equations

1. A survey of a group s viewing habits over the last year revealed the following

BANACH AND HILBERT SPACE REVIEW

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

To give it a definition, an implicit function of x and y is simply any relationship that takes the form:

Confidence Intervals for Exponential Reliability

Exponential Distribution

Joint Exam 1/P Sample Exam 1

2.2 Derivative as a Function

Some probability and statistics

Chapter 20. Vector Spaces and Bases

Probability: Terminology and Examples Class 2, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Math 3C Homework 3 Solutions

Math 431 An Introduction to Probability. Final Exam Solutions

Transcription:

Section 6.1 Joint Distribution Functions We often care about more than one random variable at a time. DEFINITION: For any two random variables X and Y the joint cumulative probability distribution function of X and Y is F(a, b) = P {X a, Y b}, < a, b < One can get the distribution of X alone from the joint distribution function { } F X (a) = P {X a} = P {X a, Y < } = P lim {X a, Y t} t Similarly, = lim t F(a, t) F(a, ) F Y (b) = lim t F(t, b) F(, b) We can also show that for a 1 < a 2 and b 1 < b 2 we have P {a 1 < X a 2, b 1 < Y b 2 } = F(a 2, b 2 ) + F(a 1, b 1 ) F(a 1, b 2 ) F(a 2, b 1 ) For discrete random variables we define the joint probability mass function for X and Y by p(x, y) = P {X = x, Y = y} Note that and p X (x) = P {X = x} = P {X = x, Y R} = p(x, y) p Y (y) = p(x, y) x:p(x,y)> y:p(x,y)> EXAMPLE: We flip a fair coin twice. Let X be 1 if head on first flip, if tail on first. Let Y be number of heads. Find p(x, y) and p X, p Y. 1

EXAMPLE: We flip a fair coin twice. Let X be 1 if head on first flip, if tail on first. Let Y be number of heads. Find p(x, y) and p X, p Y. Solution: The ranges for X and Y are {, 1}, {, 1, 2}, respectively. We have p(, ) = P(X =, Y = ) = P(X = )P(Y = X = ) = (1/2)(1/2) = 1/4 p(, 1) = P(X =, Y = 1) = P(X = )P(Y = 1 X = ) = (1/2)(1/2) = 1/4 p(, 2) = P(X =, Y = 2) = p(1, ) = P(X = 1, Y = ) = p(1, 1) = P(X = 1, Y = 1) = P(X = 1)P(Y = 1 X = 1) = (1/2)(1/2) = 1/4 p(1, 2) = P(X = 1, Y = 2) = P(X = 1)P(Y = 2 X = 1) = (1/2)(1/2) = 1/4 x... y 1 2 Row Sum = P {X = x} 1/4 1/4 1/2 1 1/4 1/4 1/2 Column Sum = P {Y = y} 1/4 1/2 1/4 Further, p X () = p X (1) = p Y () = p Y (1) = p Y (2) = p(, y) = p(, ) + p(, 1) + p(, 2) = 1/2 y= p(1, y) = p(1, ) + p(1, 1) + p(1, 2) = 1/2 y= p(x, ) = p(, ) + p(1, ) = 1/4 x= p(x, 1) = p(, 1) + p(1, 1) = 1/2 x= p(x, 2) = p(, 2) + p(1, 2) = 1/4 x= DEFINITION: We say that X and Y are jointly continuous if there exists a function f(x, y) defined for all x, y such that for any C R 2 we have P {(X, Y ) C} = f(x, y)dxdy (x,y) C where f(x, y) is called the joint probability density function of X and Y. Thus, P {X A, Y B} = f(x, y)dxdy B A 2

Also, 2 2 F(a, b) = a b a b b a f(x, y)dxdy = f(a, b) Similar to before, we view the joint density as a measure of the likelihood that the random vector (X, Y ) will be in the vicinity of (a, b). As before, P {a < X < a + da, b < Y < b + db} = b+db a+da f(x, y)dxdy f(a, b)dadb b a where da and db should be thought of as very small. We can find the marginal probability density functions as follows P {X A} = P {X A, Y (, )} = f(x, y)dydx = f X (x)dx A A where f X (x) = f(x, y)dy is therefore the density function for X. Similarly, f Y (y) = f(x, y)dx EXAMPLE: Suppose the joint density is f(x, y) = 2e x e 2y if x, y > and zero otherwise. Find the marginal of X, f X (x). Compute P {X > 1, Y < 1} and P {X < Y }. 3

EXAMPLE: Suppose the joint density is f(x, y) = 2e x e 2y if x, y > and zero otherwise. Find the marginal of X, f X (x). Compute P {X > 1, Y < 1} and P {X < Y }. Solution: For x > we have f X (x) = f(x, y)dy = 2e x e 2y dy = e x Next, 1 1 P {X > 1, Y < 1} = 2e x e 2y dxdy = 2e 2y e 1 dy = e 1 (1 e 2 ) Finally, 1 P {X < Y } = y f(x, y)dxdy = 2e x e 2y dxdy = 2e 2y (1 e y )dy (x,y):x<y = 2e 2y dy 2e 3y dy = 1 2 3 = 1 3 Everything extends to more than two random variables in the obvious way. For example, for n random variables we have F(a 1, a 2,...,a n ) = P {X 1 a 1, X 2 a 2,..., X n a n } Also, joint probability mass functions and joint densities are defined in the obvious manner. Section 6.2 Independent Random Variables DEFINITION: Two random variables are said to be independent if for any two sets of real numbers A and B we have P {X A, X B} = P {X A}P {X B} (1) That is, the events E A = {X A} and F B = {Y B} are independent. One can show that the above equation holds if and only if for all a and b we have P {X a, Y b} = P {X a}p {Y b} Therefore, two random variables X and Y are independent if and only if the joint distribution function is the product of the marginal distribution functions: F(a, b) = F X (a)f Y (b) 4

It s easy to see that for discrete random variables the condition (1) is satisfied if and only if p(x, y) = p X (x)p Y (y) (2) for all x, y where p(x, y) = P {X = x, Y = y} is the joint probability density function. This is actually easy to show. Proof: Note that (2) follows immediately from condition (1). However, for the other direction, for any A and B we have P {X A, Y B} = p(x, y) = p X (x)p Y (y) y B x A y B x A = y B p Y (y) x A p X (x) = P {Y B}P {X A} Similarly, in the jointly continuous case independence is equivalent to f(x, y) = f X (x)f Y (y) It should be clear that independence can be defined in the obvious manner for more than two random variables. That is, X 1, X 2,..., X n are independent if for all sets of real numbers A 1, A 2,..., A n we have P {X 1 A 1, X 2 A 2,..., X n A n } = n P {X i A i } i=1 The obvious conditions for the joint probability mass functions (in the discrete case) and joint densities (in the continuous case) hold. That is, in both cases independence is equivalent the join probability mass function or density being equal to the respective product of the marginals. EXAMPLES: 1. Let the joint density of X and Y be given by f(x, y) = 6e 2x e 3y for x, y > and zero otherwise. Are these random variables independent? 5

1. Let the joint density of X and Y be given by f(x, y) = 6e 2x e 3y for x, y > and zero otherwise. Are these random variables independent? Solution: We compute the marginals. For x > we have f X (x) = 6e 2x e 3y dy = 2e 2x For Y we have f Y (y) = 6e 2x e 3y dx = 3e 3y Therefore, f(x, y) = f X (x)f Y (y) for all x, y R (note that the relation also holds if one, or both, of x and y are negative). Also, X is Exp(2), and Y is Exp(3). 2. Let the joint density of X and Y be given by f(x, y) = 24xy for < x < 1, < y < 1, < x + y < 1 and zero otherwise. Are X and Y independent? 6

2. Let the joint density of X and Y be given by f(x, y) = 24xy for < x < 1, < y < 1, < x + y < 1 and zero otherwise. Are X and Y independent? Solution (see the book for another way to solve): Note that the domain in which f is nonzero depends upon an interplay between x and y. Therefore, you expect that these are not independent. Let us compute the marginals. For x (, 1) we have f(x, y) = unless < y < 1 x. Therefore, f X (x) = f(x, y)dy = 1 x 24xydy = 12x(1 x) 2 Similarly, f Y (y) = f(x, y)dx = 1 y 24xydx = 12y(1 y) 2 Therefore, we do not have f(x, y) = f X (x)f Y (y) that for all x, y and hence X and Y are not independent. 3. For i = 1,...,n, let X i Exp(λ i ) be independent exponential random variables with parameters λ i >. What is the distribution of the minimum of the X i? Solution: Let Z = min{x i }. Clearly, the range of Z is t >. Therefore P {Z > t} = P {min{x i } > t} = P {X 1 > t, X 2 > t,...,x n > t} = P {X 1 > t}p {X 2 > t}...p {X n > t} (by independence) { ( n ) } = e λ1t e λ2t...e λnt = exp λ i t It follows that for t we have { ( n } F Z (t) = 1 P {Z > t} = 1 exp λ i )t therefore Z Exp(λ 1 +... + λ n ). One natural way to interpret this result is the following: if n alarm clocks are set to go off after an exponentially distributed amount of time (each with a potentially different rate λ i ), then the time that the first alarm rings is also exponentially distributed with the parameter which equals to the sum of the parameters of all the clocks, that is λ i. i i=1 i=1 7