Notes on Continuous Random Variables



Similar documents
5. Continuous Random Variables

Section 5.1 Continuous Random Variables: Introduction

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

Lecture 7: Continuous Random Variables

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

An Introduction to Basic Statistics and Probability

Lecture 6: Discrete & Continuous Probability and Random Variables

Math 461 Fall 2006 Test 2 Solutions

Lecture 5 : The Poisson Distribution

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

Normal distribution. ) 2 /2σ. 2π σ

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

ECE302 Spring 2006 HW5 Solutions February 21,

4. Continuous Random Variables, the Pareto and Normal Distributions

MATH 10: Elementary Statistics and Probability Chapter 5: Continuous Random Variables

Statistics 100A Homework 7 Solutions

Chapter 4 Lecture Notes

6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:30-9:30 PM. SOLUTIONS

Important Probability Distributions OPRE 6301

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables

Probability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X

Continuous Random Variables

The Normal Distribution. Alan T. Arnholt Department of Mathematical Sciences Appalachian State University

Lecture 2: Discrete Distributions, Normal Distributions. Chapter 1

Practice problems for Homework 11 - Point Estimation

Math 425 (Fall 08) Solutions Midterm 2 November 6, 2008

1.1 Introduction, and Review of Probability Theory Random Variable, Range, Types of Random Variables CDF, PDF, Quantiles...

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Exponential Distribution

2WB05 Simulation Lecture 8: Generating random variables

Joint Exam 1/P Sample Exam 1

Stats on the TI 83 and TI 84 Calculator

The Exponential Distribution

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Math 151. Rumbos Spring Solutions to Assignment #22

Math 431 An Introduction to Probability. Final Exam Solutions

Lesson 20. Probability and Cumulative Distribution Functions

Estimating the Average Value of a Function

Chapter 3 RANDOM VARIATE GENERATION

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Chapter 4 - Lecture 1 Probability Density Functions and Cumul. Distribution Functions

16. THE NORMAL APPROXIMATION TO THE BINOMIAL DISTRIBUTION

Week 2: Exponential Functions

Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

1 Sufficient statistics

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL

2.6. Probability. In general the probability density of a random variable satisfies two conditions:

The normal approximation to the binomial

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

Exploratory Data Analysis

WEEK #22: PDFs and CDFs, Measures of Center and Spread

Pr(X = x) = f(x) = λe λx

Probability Calculator

Math 120 Final Exam Practice Problems, Form: A

Statistics 100A Homework 4 Solutions

Hypothesis Testing for Beginners

Probability Theory. Florian Herzog. A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T..

Lecture 8. Generating a non-uniform probability distribution

CHAPTER 7 SECTION 5: RANDOM VARIABLES AND DISCRETE PROBABILITY DISTRIBUTIONS

e.g. arrival of a customer to a service station or breakdown of a component in some system.

Probability Distribution for Discrete Random Variables

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

ECE302 Spring 2006 HW4 Solutions February 6,

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

3.4 The Normal Distribution

Stat 515 Midterm Examination II April 6, 2010 (9:30 a.m. - 10:45 a.m.)

VISUALIZATION OF DENSITY FUNCTIONS WITH GEOGEBRA

Chapter 5. Random variables

The Normal Distribution

Chapter 5 Discrete Probability Distribution. Learning objectives

Tenth Problem Assignment

Random Variables. Chapter 2. Random Variables 1

Practice Problems for Homework #6. Normal distribution and Central Limit Theorem.

6.4 Normal Distribution

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

The normal approximation to the binomial

Probability Distributions

Lecture 8. Confidence intervals and the central limit theorem

Practice Problems #4

correct-choice plot f(x) and draw an approximate tangent line at x = a and use geometry to estimate its slope comment The choices were:

Homework 4 - KEY. Jeff Brenion. June 16, Note: Many problems can be solved in more than one way; we present only a single solution here.

Sums of Independent Random Variables

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Some special discrete probability distributions

LECTURE 16. Readings: Section 5.1. Lecture outline. Random processes Definition of the Bernoulli process Basic properties of the Bernoulli process

Notes on the Negative Binomial Distribution

10.2 Series and Convergence

Normal Distribution as an Approximation to the Binomial Distribution

Feb 28 Homework Solutions Math 151, Winter Chapter 6 Problems (pages )

Probability Distributions

AP CALCULUS AB 2008 SCORING GUIDELINES

2 Binomial, Poisson, Normal Distribution

Section 6.4: Work. We illustrate with an example.

LINEAR EQUATIONS IN TWO VARIABLES

Master s Theory Exam Spring 2006

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

ST 371 (IV): Discrete Random Variables

4 The M/M/1 queue. 4.1 Time-dependent behaviour

Transcription:

Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes them from discrete random variables, which can take on only a sequence of values, usually integers. Typically random variables that represent, for example, time or distance will be continuous rather than discrete. Just as we describe the probability distribution of a discrete random variable by specifying the probability that the random variable takes on each possible value, we describe the probability distribution of a continuous random variable by giving its density function. A density function is a function f which satisfies the following two properties: 1. f(x) for all x. 2. f(x)dx = 1. The first condition says that the density function is always nonnegative, so the graph of the density function always lies on or above the x-axis. The second condition ensures that the area under the density curve is 1. We think of a continuous random variable with density function f as being a random variable that can be obtained by picking a point at random from under the density curve and then reading off the x-coordinate of that point. Because the total area under the density curve is 1, the probability that the random variable takes on a value between a and b is the area under the curve between a and b. More precisely, if X is a random variable with density function f and a < b, then P(a X b) = b a f(x)dx. Example 1: Suppose the income (in tens of thousands of dollars) of people in a community can be approximated by a continuous distribution with density { 2x 2 if x 2 if x < 2 a) Find the probability that a randomly chosen person has an income between $, and $5,. b) Find the probability that a randomly chosen person has an income of at least $,. c) Find the probability that a randomly chosen person has an income of at most $4,. Solution: Let X be the income of a randomly chosen person. The probability that a randomly chosen person has an income between $, and $5, is 5 5 x=5 P( X 5) = f(x)dx = 2x 2 dx = 2x 1 = 2 ( 5 2 ) = 2 2 5 = 4 15. The probability that a randomly chosen person has an income of at least $, is P(X ) = f(x)dx = = lim n 2x 1 x=n x= x= n 2x 2 dx = lim 2x 2 dx n ( 2 n + 2 ) = 1. = lim n 1

Finally, for part c), we get P(X 4) = 4 f(x)dx = 2 dx+ 4 2 x=4 2x 2 dx = 2x 1 = 2 4 + 2 2 = 1 2. Remark 1: Note that part b) required evaluating an improper integral, in which the upper endpoint was infinity. This integral could be evaluated by integrating the density f(x) from to n and taking a limit as n. Evaluation of improper integrals is often required when working with unbounded random variables that can take on any positive number as a value. Remark 2: When the density function of a continuous random variable is defined in two pieces, it is important to be careful about the limits of integration. In part c), we needed to integrate the density from to 4. However, since the density is zero to the left of 2, we only integrated 2x 1 from 2 to 4. It is often useful to draw pictures of the density to avoid mistakenly integrating over the wrong interval. Remark : If we were interested in finding the probability that the random variable X in the Example 1 were exactly equal to, then we would be integrating from to, and we would get zero. This is a general fact about continuous random variables that helps to distinguish them from discrete random variables. A discrete random variable takes on certain values with positive probability. However, if X is a continuous random variable with density f, then P(X = y) = for all y. This may seem counterintuitive at first, since after all X will end up taking some value, but the point is that since X can take on a continuum of values, the probability that it takes on any one particular value is zero. Expected value and standard deviation The procedure for finding expected values and standard deviations for continuous random variables of continuous random variables is similar to the procedure used to calculate expected values and standard deviations for discrete random variables. The differences are that sums in the formula for discrete random variables get replaced by integrals (which are the continuous analogs of sums), while probabilities in the formula for discrete random variables get replaced by densities. More precisely, if X is a random variable with density f(x), then the expected value of X is given by while the variance is given by Note that µ = E[X] = Var(X) = E[(X µ) 2 ] = xf(x)dx, (x µ) 2 f(x)dx. Var(X) = E[(X µ) 2 ] = E[X 2 2µX +µ 2 ] = E[X 2 ] 2µE[X]+µ 2 = E[X 2 ] 2µ 2 +µ 2 = E[X 2 ] µ 2. x=2 2

This formula for the variance (which is valid for all random variables) is often easier to use in computations, so we may calculate ( ) Var(X) = E[X 2 ] µ 2 = x 2 f(x)dx µ 2. As in the case of discrete random variables, the standard deviation of X is the square root of the variance of X. Example 2: Suppose a train arrives shortly after 1: PM each day, and that the number of minutes after 1: that the train arrives can be modeled as a continuous random variable with density { 2(1 x) if x 1 Find the mean and standard deviation of the number of minutes after 1: that the train arrives. Solution: Let X be the number of minutes after 1: that the train arrives. The mean (or, equivalently, the expected value) of X is given by µ = E[X] = Also, we have E[X 2 ] = Therefore, xf(x)dx = x 2 f(x)dx = x 2(1 x)dx = x 2 2(1 x) = Var(X) = 1 ( 1 and the standard deviation is 1/18.24. 2x 2x 2 dx = ) (x 2 2x x=1 = 1 x=. ( 2x 2x 2 2x ) dx = 2x4 x=1 = 2 4 2 4 = 1. ) 2 = 1 1 9 = 1 18, Special Continuous Distributions As was the case with discrete random variables, when we gave special attention to the geometric, binomial, and Poisson distributions, some continuous distributions occur repeatedly in applications. Probably the three most important continuous distributions are the uniform distribution, the exponential distribution, and the normal distribution. Uniform Distribution: If a < b, then we say a random variable X has the uniform distribution on [a,b] if { 1 b a if a x b Note that the density function is zero outside [a,b], so a random variable having the uniform distribution on [a,b] always falls between a and b. Because the density is flat between a and b, we can think of the uniform distribution as representing a number chosen uniformly at random from the interval [a, b]. x=

Example : If X has the uniform distribution on [2,5], calculate P(X > 4). Solution: The density of X is given by Therefore, P(X 4) = 4 f(x)dx = { 1 if 2 x 5 5 4 1 dx = x x=5 x=4 = 5 4 = 1. Alternatively, we could observe that the area under the density function between x = 4 and x = 5 is a rectangle of length 1 and height 1/, which has area 1/. Exponential Distribution: If λ >, then we say a random variable X has the exponential distribution with rate λ if its density is given by { λe λx if x The exponential distribution is a model of the amount of time we have to wait for an event that happens at rate λ per unit time. The exponential distribution should also describe the time we have to wait for a customer to arrive in a store, a goal to be scored in a hockey game, a radioactive particle to be emitted, an accident to occur on a roadway, or an earthquake to occur in California. The exponential distribution arises in the same circumstances under which the Poisson distribution arises. However, whereas the number of events that occur in a certain time interval has the Poisson distribution, the amount of time to wait for the next event has the exponential distribution. Because the Poisson distribution is integer-valued and the exponential distribution is a continuous distribution, one should be able to keep the two distributions straight. The most important property of the exponential distribution is known as the memoryless property, whichsays thatif thetimetowait foranevent tooccurhastheexponential distribution, then the probability that we have to wait an additional time t is the same no matter how long we have already waited. More formally, if X has an exponential distribution with rate λ, then P(X s+t X s) = P(X t). That is, the probability that we have to wait for an additional time t (and therefore a total time of s+t) given that we have already waited for time s is the same as the probability at the start that we would have had to wait for time t. This is true for all s, that is, no matter how long we have already waited. To see why the memoryless property holds, note that for all t, we have It follows that P(X t) = P(X s+t X s) = t λe λx dx = e λx = e λt. P(X s+t and X s) P(X s) = e λ(s+t) e λs = e λs e λt e λs 4 t = P(X s+t) P(X s) = e λt = P(X t),

which verifies the memoryless property. If X has an exponential distribution with parameter λ, then E[X] = 1/λ. Thus, the expected time to wait for an event to occur is inversely proportional to the rate at which the event occurs, as would be expected. Example 4: If customers arrive in a store at the rate of 4 per hour, what is the probability that we will have to wait between 15 and minutes for the next customer to arrive. Solution: Let X be the time we have to wait for the next customer to arrive. Then X has the exponential distribution with parameter λ = 4, so X has density { 4e 4x if x The probability that we have to wait between 15 and minutes is P(1/4 X 1/2) = /2 1/4 x=1/2 4e 4x dx = e 4x = e 4/2 ( e 4/4 ) = e 1 e 2.2. x=1/4 Normal Distribution: A random variable has a normal distribution with mean µ and standard deviation σ if its density is given by 1 σ /2σ2 e (x µ)2. 2π The normal distribution is the most important distribution in statistics, arising in numerous applications because of a famous result known as the Central Limit Theorem. Because the normal distribution is discussed in our textbook, we do not pursue it further here. Problems 1: Suppose X is a random variable with density function { 2x if < x < 1. a) Find P(X 1/2). b) Find P(X /4). c) Find P(X 2). d) Find E[X]. e) Find the standard deviation of X. 2: Suppose X is a random variable with density function { cx 2 if < x < 2 for some positive number c. What is c? 5

: Suppose the number of years that a television set lasts has density { 18x if x. a) Find the probability that the television set lasts between 5 and years. b) Find the probability that the television set lasts at least 4 years. c) Find the probability that the television set lasts less than 2.5 years. d) Find the probability that the television set lasts exactly.7 years. e) Find the expected value of the number of years that the television set lasts. 4: Suppose that the number of hours that it takes for a student to finish an exam has density { 2 5 (x+1) if 1 < x < 2. a) Find the probability that the student finishes the exam in less than 1.5 hours. b) Find the mean and standard deviation of the number of hours it takes to finish the exam. 5: Suppose that if you arrive at a bus stop at 8:, the amount of time that you will have to wait for the next bus (in minutes) is uniformly distributed on [,]. a) Find the probability that you will have to wait at least 1 minute for the bus. b) Find the probability that you will have to wait between seconds and 2 minutes. c) Find the expected value and standard deviation of the amount of time that you have to wait for the next bus. : Suppose the random variable X has the uniform distribution on [a,b]. Find expressions involving a and b for the expected value, variance, and standard deviation of X. Check that your expressions when a = and b = agree with what you got in part c) of problem 5. 7: Suppose that the amount of time (in months) that a light bulb lasts before it burns out has an exponential distribution with parameter λ = 1/4. a) What is the probability that it lasts at least three months? b) If it has already lasted two months, what is the probability that it will last at least three more months? c) On average, how many months will the light bulb last? 8: Suppose major earthquakes in a region occur independently of one another at the rate of one every ten years. Find the probability that the next major earthquake will occur between 5 and 1 years from now. 9: Accidents occur at a busy intersection at the rate of three per year. What is the probability that it will be at least one year before the next accident at the intersection? Compute the answer using the following two methods: a) Let X bethenumberof accidents inthe nextyear. Find thedistributionof X and calculate P(X = ). b) Let T be the amount of time until the next accident. Find the distribution of T and calculate P(T > 1).