10 BIVARIATE DISTRIBUTIONS

Similar documents
6 PROBABILITY GENERATING FUNCTIONS

Solutions to Homework 10

Section 12.6: Directional Derivatives and the Gradient Vector

THE CENTRAL LIMIT THEOREM TORONTO

Chapter 9. Systems of Linear Equations

Feb 28 Homework Solutions Math 151, Winter Chapter 6 Problems (pages )

Notes on Continuous Random Variables

x 2 + y 2 = 1 y 1 = x 2 + 2x y = x 2 + 2x + 1

Practice Final Math 122 Spring 12 Instructor: Jeff Lang

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Lecture 8 : Coordinate Geometry. The coordinate plane The points on a line can be referenced if we choose an origin and a unit of 20

Joint Exam 1/P Sample Exam 1

3 Contour integrals and Cauchy s Theorem

L 2 : x = s + 1, y = s, z = 4s Suppose that C has coordinates (x, y, z). Then from the vector equality AC = BD, one has

Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points

Exam 1 Sample Question SOLUTIONS. y = 2x

Understanding Basic Calculus

APPLIED MATHEMATICS ADVANCED LEVEL

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Section 6.1 Joint Distribution Functions

Statistics 100A Homework 7 Solutions

Chapter 3 RANDOM VARIATE GENERATION

SOLUTIONS. f x = 6x 2 6xy 24x, f y = 3x 2 6y. To find the critical points, we solve

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

FURTHER VECTORS (MEI)

GRAPHING IN POLAR COORDINATES SYMMETRY

Review Sheet for Test 1

Metric Spaces. Chapter Metrics

Section 5.1 Continuous Random Variables: Introduction

Parametric Equations and the Parabola (Extension 1)

4. Continuous Random Variables, the Pareto and Normal Distributions

Sums of Independent Random Variables

(a) We have x = 3 + 2t, y = 2 t, z = 6 so solving for t we get the symmetric equations. x 3 2. = 2 y, z = 6. t 2 2t + 1 = 0,

Lecture 8. Generating a non-uniform probability distribution

Math 2443, Section 16.3

TOPIC 4: DERIVATIVES

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

Homework #2 Solutions

THE COMPLEX EXPONENTIAL FUNCTION

11.1. Objectives. Component Form of a Vector. Component Form of a Vector. Component Form of a Vector. Vectors and the Geometry of Space

Analysis of Stresses and Strains

Vector Math Computer Graphics Scott D. Anderson

Mathematics Course 111: Algebra I Part IV: Vector Spaces

1.2 GRAPHS OF EQUATIONS. Copyright Cengage Learning. All rights reserved.

Copyrighted Material. Chapter 1 DEGREE OF A CURVE

Some probability and statistics

n 2 + 4n + 3. The answer in decimal form (for the Blitz): 0, 75. Solution. (n + 1)(n + 3) = n lim m 2 1

Section 1.1. Introduction to R n

Copyright 2011 Casa Software Ltd. Centre of Mass

Factoring Patterns in the Gaussian Plane

AB2.5: Surfaces and Surface Integrals. Divergence Theorem of Gauss

ECE302 Spring 2006 HW5 Solutions February 21,

27.3. Introduction. Prerequisites. Learning Outcomes

Lecture 6: Discrete & Continuous Probability and Random Variables

Solutions to Practice Problems for Test 4

PYTHAGOREAN TRIPLES KEITH CONRAD

4. How many integers between 2004 and 4002 are perfect squares?

Planar Curve Intersection

Estimating the Average Value of a Function

( 1) = 9 = 3 We would like to make the length 6. The only vectors in the same direction as v are those

Integrals of Rational Functions

5.3 Improper Integrals Involving Rational and Exponential Functions

Algebra Academic Content Standards Grade Eight and Grade Nine Ohio. Grade Eight. Number, Number Sense and Operations Standard

Solving Quadratic Equations

D Alembert s principle and applications

Covariance and Correlation

Math 1B, lecture 5: area and volume

Review of Fundamental Mathematics

Line and surface integrals: Solutions

Solutions - Homework sections

1. First-order Ordinary Differential Equations

Math 241, Exam 1 Information.

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference

Math 370, Actuarial Problemsolving Spring 2008 A.J. Hildebrand. Practice Test, 1/28/2008 (with solutions)

Solutions for Review Problems

Elliptical copulae. Dorota Kurowicka, Jolanta Misiewicz, Roger Cooke

Math 461 Fall 2006 Test 2 Solutions

Module 3: Correlation and Covariance

7 CONTINUOUS PROBABILITY DISTRIBUTIONS

MULTIPLE INTEGRALS. h 2 (y) are continuous functions on [c, d] and let f(x, y) be a function defined on R. Then

Normal distribution. ) 2 /2σ. 2π σ

This makes sense. t /t 2 dt = 1. t t 2 + 1dt = 2 du = 1 3 u3/2 u=5

E3: PROBABILITY AND STATISTICS lecture notes

1. Introduction sine, cosine, tangent, cotangent, secant, and cosecant periodic

1 3 4 = 8i + 20j 13k. x + w. y + w

Geometry of Vectors. 1 Cartesian Coordinates. Carlo Tomasi

Probability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X

Consumer Theory. The consumer s problem

Session 7 Bivariate Data and Analysis

SECOND-ORDER LINEAR DIFFERENTIAL EQUATIONS

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

LINEAR INEQUALITIES. Mathematics is the art of saying many things in many different ways. MAXWELL

Numerical Analysis Lecture Notes

CHAPTER FIVE. 5. Equations of Lines in R 3

Math 120 Final Exam Practice Problems, Form: A

Econ 132 C. Health Insurance: U.S., Risk Pooling, Risk Aversion, Moral Hazard, Rand Study 7

MATH Fundamental Mathematics IV

3. INNER PRODUCT SPACES

THREE DIMENSIONAL GEOMETRY

Transcription:

BIVARIATE DISTRIBUTIONS After some discussion of the Normal distribution, consideration is given to handling two continuous random variables. The Normal Distribution The probability density function f(x) associated with the general Normal distribution is: f(x) = (x µ) e σ (.) πσ The range of the Normal distribution is to + and it will be shown that the total area under the curve is. It will also be shown that µ is the mean and that σ is the variance. A graphical representation of the Normal distribution is: X f(x) x µ It is immediately clear from (.) that f(x) is symmetrical about x = µ. This value of x is marked. When µ = the curve is symmetrical about the vertical axis. The value of σ governs the width of the hump and it turns out that the inflexion points (one each side of the peak) are at x = µ ± σ where σ = σ is the standard deviation. When integrating expressions which incorporate the probability density function, it is essential to know the following standard result (a derivation is presented on page.): e x dx = π (.) This result makes it easy to check that the total area under the curve is : Let t = x µ σ Total Area = so dx = σ dt. Then: πσ e (x µ) σ dx Total Area = πσ e t σ dt = π e t dt = π. π =.

From the symmetry of the probability density function (.) it is clear that the expectation E(X) = µ but to give a further illustration of the use of the standard result (.) the expectation is here evaluated by integration: Let t = x µ σ E(X) = πσ x.e (x µ) σ so x = σ t + µ and dx = σ dt. Then: dx E(X) = πσ ( σ t + µ).e t σ dt = σ π t.e t dt + µ π e t dt σ = [ ] + π e t σ = [ ] + µ π = µ + µ π. π The variance, for once, is conveniently evaluated from the formula V(X) = E ( (X µ) ) : Let t = x µ σ V(X) = E ( (X µ) ) = so x µ = σ t πσ and dx = σ dt. Then: (x µ).e (x µ) σ dx V(X) = πσ σ t.e t σ dt = σ π t d dt( e t ) dt The integrand has been put into a form ready for integration by parts: [ ] + V(X) = σ t.e t + σ e t dt π π = σ π [ ] + σ π. π = σ.

Standard Form of the Normal Distribution The general Normal distribution is described as: Normal(µ,σ ) or simply N(µ,σ ) The smaller the variance σ the narrower and taller the hump of the probability density function. A particularly unfortunate difficulty with the Normal distribution is that integrating the probability density function between arbitrary limits is intractable. Thus, for arbitrary a and b, it is impossible to evaluate: P(a X < b) = πσ b a e (x µ) σ dx The traditional way round this problem is to use tables though these days appropriate facilities are built into any decent spreadsheet application. Unsurprisingly, tables are not available for a huge range of possible (µ,σ ) pairs but the special case when the mean is zero and the variance is one, the distribution Normal(,), is well documented. The probability distribution function for Normal(,) is often referred to as Φ(x): Φ(x) = x e t dt π Tables of Φ(x) are readily available so, given a random variable X distributed Normal(,), it is easy to determine the probability that X lies in the range a to b. This is: P(a X < b) = Φ(b) Φ(a) It is now straightforward to see how to deal with the general case of a random variable X distributed Normal(µ,σ ). To determine the probability that X lies in the range a to b first reconsider: Let t = x µ σ P(a X < b) = so dx = σ dt. Then: πσ b a e (x µ) σ dx P(a X < b) = πσ b µ σ a µ σ e t σ dt = b µ σ π a µ σ e t dt The integration is thereby transformed into that for the distribution Normal(,) and is said to be in standard form. Noting the new limits: ( ) ( ) b µ a µ P(a X < b) = Φ Φ σ σ.3

The Central Limit Theorem No course on probability or statistics is complete without at least a passing reference to the Central Limit Theorem. This is an extraordinarily powerful theorem but only the most token of introductory remarks will be made about it here. Suppose X,X,...X n are n independent and identically distributed random variables each with mean µ and variance σ. Consider two derived random variables Y and Z which are respectively the sum and mean of X,X,...X n : Y = X + X + + X n and Z = X + X + + X n n From the rules for Expectation and Variance discussed on pages 4.6 and 4.8 it is simple to determine the Expectation and Variance of Y and Z: E(Y ) = nµ, V(Y ) = nσ and E(Z) = µ, V(Z) = σ n The Central Limit Theorem addresses the problem of how the derived random variables Y and Z are distributed and asserts that, as n increases indefinitely, both Y and Z tend to a Normal distribution whatever the distribution of the individual X i. More specifically: ) Y tends to N(nµ,nσ ) and Z tends to N (µ, σ n In many cases Y and Z are approximately Normal for remarkably small values of n. Bivariate Distributions Reference Discrete Example It has been noted that many formulae for continuous distributions parallel equivalent formulae for discrete distributions. In introducing examples of two continuous random variables it is useful to employ a reference example of two discrete random variables. Consider two discrete random variables X and Y whose values are r and s respectively and suppose that the probability of the event {X = r} {Y = s} is given by: r + s, if r,s 3 P(X = r,y = s) =, otherwise The probabilities may be tabulated thus: X r 3 3 Y s 3 6 3 4 P(Y = s) 3 4 5 4.4 3 4 5 6 8 6 P(X = r) 4 8

The following is a graphical representation of the same function: Y 3 X 3 The probabilities exist for only integer values of r and s so the effect is of a kind of pegboard. Different lengths of peg correspond to different probabilities. Axiom II requires the sum of all the entries in the table (which is the total length of the pegs) to be. The table also includes the marginal sums which separately tabulate the probabilities P(X = r) and P(Y = s). Bivariate Distributions Continuous Random Variables When there are two continuous random variables, the equivalent of the two-dimensional array is a region of the x y (cartesian) plane. Above the plane, over the region of interest, is a surface which represents the probability density function associated with a bivariate distribution. Suppose X and Y are two continuous random variables and that their values, x and y respectively, are constrained to lie within some region R of the cartesian plane. The associated probability density function has the general form f XY (x,y) and, regarding this function as defining a surface over region R, axiom II requires: f XY (x,y) dxdy = R This is equivalent to requiring that the volume under the surface is and corresponds to the requirement that the total length of the pegs in the pegboard is..5

As an illustration, consider a continuous analogy of the reference example of two discrete random variables. Suppose X and Y are two continuous random variables and that their values x and y are constrained to lie in the unit square x,y <. Suppose further that the associated bivariate probability density function is: { x + y, if x,y < f XY (x,y) =, otherwise This probability density function can be regarded as defining a surface over the unit square. Continuous functions cannot satisfactorily be tabulated but it is not difficult to depict a graphical representation of f XY (x,y): Y... X. Note that when x = y = the value of f XY (x,y) is, an impossible value for a probability but a perfectly possible value for a probability density function. Probabilities correspond to volumes now and it is easy to check that the total volume under the surface is. Given that region R represents the specified unit square: [ x ] f XY (x,y) dxdy = (x + y) dxdy = + yx dy R = ( + y ) dy = [ y ] + y = In the general case where the probability density function is f XY (x,y), the probability that (x,y) lies in a particular sub-region R s of R is given by: P(X,Y lies in R s ) = f XY (x,y) dxdy R s.6

In the example case, suppose that the sub-region R s is defined by the half-unit square x,y < : P(X,Y lies in R s ) = = (x + y) dxdy = ( 8 + y ) dy = [ y 8 + y 4 [ x ] + yx ] = 8 dy In the reference discrete example (where the values of both random variables are confined to the range to 3) this result loosely corresponds to determining P(X,Y < ): P(X,Y < ) = r= P(X = r,y = s) = + + + = 4 = s= Why is the answer different? The Equivalent of Marginal Sums With two discrete random variables, the marginal sums P(X = r) and P(Y = s) are given by the relationships: P(X = r) = s P(X = r,y = s) and P(Y = s) = r P(X = r,y = s) A pair of continuous random variables X and Y governed by a bivariate distribution function f XY (x,y) will, separately, have associated probability density functions f X (x) and f Y (y). By analogy with the discrete case, these functions are given by the relationships: f X (x) = ymax y min f XY (x,y) dy and f Y (y) = xmax x min f XY (x,y) dx The limits of integration merit a moment s thought. The region, R, over which f XY (x,y) is defined is not, in general, a square. Accordingly, the valid range of y depends on x and, in consequence, the limits y min and y max will depend on x. Likewise, the limits x min and x max will depend on y. In the example case, where R is the unit square, there is no difficulty about the limits: f X (x) = (x + y) dy = ] [xy + y = x + and: f Y (y) = [ x ] (x + y) dx = + yx = y +.7

The Equivalent of Marginal Sums and Independence With two discrete random variables, the two sets of marginal sums each sum to. With continuous random variables, integrating each of the functions f X (x) and f Y (y) must likewise yield. It is easy to verify that this requirement is satisfied in the present case: ( x + ) [ x dx = + x ] = and ( y + ) [ y dy = + y ] = With two discrete random variables, marginal sums are used to test for independence. Two variables variables X and Y are said to be independent if: P(X = r,y = s) = P(X = r).p(y = s) for all r,s By analogy, two continuous random variables X and Y are said to be independent if: In the present case: f XY (x,y) = f X (x).f Y (y) f XY (x,y) = x + y and f X (x).f Y (y) = ( x + )( y + ) = xy + x + y Clearly two continuous random variables X and Y whose probability density function is x + y are not independent but the function just derived can be dressed up as bivariate probability density function whose associated random variables are independent: { (4xy + x + y + )/4 if x,y < f XY (x,y) =, otherwise Illustration The Uniform Distribution Suppose X and Y are independent and that both are distributed Uniform(,). Their associated probability density functions are: + 4 {, if x < f X (x) =, otherwise {, if y < and f Y (y) =, otherwise Given this trivial case, the product is clearly the bivariate distribution function: {, if x,y < f XY (x,y) =, otherwise The surface defined by f XY (x,y) is a unit cube so it is readily seen that: f XY (x,y) dxdy = R.8

Illustration The Normal Distribution Suppose X and Y are independent and that both are distributed Normal(,). Their associated probability density functions are: f X (x) = π e x and f Y (y) = π e y In this case the product leads to the probability density function: f XY (x,y) = e x. e y = π π π e (x +y ) (.3) The surface approximates that obtained by placing a large ball in the centre of a table and then draping a tablecloth over the ball. Glossary The following technical terms have been introduced: standard form bivariate distribution Exercises X. Find the points of inflexion of the probability density function associated with a random variable which is distributed Normal(µ,σ ). Hence find the points at which the tangents at the points of inflection intersect the x-axis.. The Cauchy distribution has probability density function: f(x) = c + x where < x < + Evaluate c. Find the probability distribution function F(x). Calculate P( x < ). 3. Two continuous random variables X and Y have the following bivariate probability function which is defined over the unit square: { (9 6x 6y + 4xy)/4 if x,y < f XY (x,y) =, otherwise (a) Given that R is the unit square, verify that R f XY(x,y) dxdy =. (b) Determine f X (x) and f Y (y). (c) Hence state whether or not the two random variables are independent..9

ADDENDUM AN IMPORTANT INTEGRATION For arbitrary limits a and b it is impossible to evaluate b a e x dx but evaluation is possible if a = and b = + and an informal analysis is presented below. As a preliminary observation note that: e x dx = e x dx e y dy = e (x +y ) dxdy The item under the square root sign can be integrated by the substitution of two new variables r and θ using the transformation functions: x = r cos θ and y = r sin θ Further discussion of integration by the substitution of two new variables will be given on pages. and.. In the present case, the integration is transformed thus: e (x +y ) dxdy = π e r rdr dθ The first pair of limits reflects integration over a square constituting the quadrant of the cartesian plane in which x and y are both positive. The second pair of limits reflects integration over a quarter circle in the same quadrant. Since the function being integrated is strongly convergent as x and y or r increase, it is valid to equate the two integrations. The right-hand integration is straightforward: Accordingly: π e r rdr dθ = π [ e r e x dx = ] dθ = π 4 = π dθ = [ θ This leads, by symmetry, to the standard result quoted as (.): e x dx = π π ] π = π 4.