6.2. Discrete Probability Distributions



Similar documents
3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved.

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Chapter 5. Random variables

2. Discrete random variables

The Binomial Distribution

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

Chapter 4 Lecture Notes

Statistics 100A Homework 4 Solutions

STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Normal distribution. ) 2 /2σ. 2π σ

LECTURE 16. Readings: Section 5.1. Lecture outline. Random processes Definition of the Bernoulli process Basic properties of the Bernoulli process

Chapter 4. Probability Distributions

Unit 4 The Bernoulli and Binomial Distributions

Homework 4 - KEY. Jeff Brenion. June 16, Note: Many problems can be solved in more than one way; we present only a single solution here.

Binomial Random Variables

Practice problems for Homework 11 - Point Estimation

WHERE DOES THE 10% CONDITION COME FROM?

Question: What is the probability that a five-card poker hand contains a flush, that is, five cards of the same suit?

2 Binomial, Poisson, Normal Distribution

Practice Problems #4

V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE

Ch5: Discrete Probability Distributions Section 5-1: Probability Distribution

4. Continuous Random Variables, the Pareto and Normal Distributions

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL

SOLUTIONS: 4.1 Probability Distributions and 4.2 Binomial Distributions

Lecture 5 : The Poisson Distribution

Notes on Continuous Random Variables

Concepts of Probability

The Binomial Probability Distribution

Some special discrete probability distributions

STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables

Binomial Random Variables. Binomial Distribution. Examples of Binomial Random Variables. Binomial Random Variables

5. Probability Calculus

Sample Questions for Mastery #5

E3: PROBABILITY AND STATISTICS lecture notes

Important Probability Distributions OPRE 6301

Chapter 5. Discrete Probability Distributions

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Probability: Terminology and Examples Class 2, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

CA200 Quantitative Analysis for Business Decisions. File name: CA200_Section_04A_StatisticsIntroduction

Math 461 Fall 2006 Test 2 Solutions

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 13. Random Variables: Distribution and Expectation

We rst consider the game from the player's point of view: Suppose you have picked a number and placed your bet. The probability of winning is

The normal approximation to the binomial

Chapter 3 RANDOM VARIATE GENERATION

Probability Calculator

PROBABILITY SECOND EDITION

ST 371 (IV): Discrete Random Variables

The normal approximation to the binomial

REPEATED TRIALS. The probability of winning those k chosen times and losing the other times is then p k q n k.

Random variables, probability distributions, binomial random variable

SCHOOL OF ENGINEERING & BUILT ENVIRONMENT. Mathematics

Chapter 9 Monté Carlo Simulation

Math 431 An Introduction to Probability. Final Exam Solutions

Lecture 7: Continuous Random Variables

Joint Exam 1/P Sample Exam 1

Introduction to the Practice of Statistics Fifth Edition Moore, McCabe

6 PROBABILITY GENERATING FUNCTIONS

Applied Reliability Applied Reliability

Probability Distributions

Chapter 5: Normal Probability Distributions - Solutions

1 Sufficient statistics

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Feb 7 Homework Solutions Math 151, Winter Chapter 4 Problems (pages )

PROBABILITY AND SAMPLING DISTRIBUTIONS

People have thought about, and defined, probability in different ways. important to note the consequences of the definition:

Statistics 100A Homework 8 Solutions

AMS 5 CHANCE VARIABILITY

The Normal Approximation to Probability Histograms. Dice: Throw a single die twice. The Probability Histogram: Area = Probability. Where are we going?

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

3.2 Roulette and Markov Chains

e.g. arrival of a customer to a service station or breakdown of a component in some system.

Math/Stats 342: Solutions to Homework

Statistics 100A Homework 3 Solutions

Probability and Statistics Vocabulary List (Definitions for Middle School Teachers)

MAS108 Probability I

Math 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 2 Solutions

Binomial random variables (Review)

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

Mathematical goals. Starting points. Materials required. Time needed

STRUTS: Statistical Rules of Thumb. Seattle, WA

TEST 2 STUDY GUIDE. 1. Consider the data shown below.

Normal Approximation. Contents. 1 Normal Approximation. 1.1 Introduction. Anthony Tanbakuchi Department of Mathematics Pima Community College

Chapter 4 & 5 practice set. The actual exam is not multiple choice nor does it contain like questions.

MAT 155. Key Concept. September 27, S5.5_3 Poisson Probability Distributions. Chapter 5 Probability Distributions

Lecture 10: Depicting Sampling Distributions of a Sample Proportion

MA 1125 Lecture 14 - Expected Values. Friday, February 28, Objectives: Introduce expected values.

5. Continuous Random Variables

Lecture 2 Binomial and Poisson Probability Distributions

10.2 Series and Convergence

Binomial Probability Distribution

Math 425 (Fall 08) Solutions Midterm 2 November 6, 2008

AP Stats - Probability Review

13.0 Central Limit Theorem

Section 5 Part 2. Probability Distributions for Discrete Random Variables

Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80)

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Maximum Likelihood Estimation

Chapter 5 Discrete Probability Distribution. Learning objectives

Transcription:

6.2. Discrete Probability Distributions Discrete Uniform distribution (diskreetti tasajakauma) A random variable X follows the dicrete uniform distribution on the interval [a, a+1,..., b], if it may attain each of these values with equal probability. We have then: P (X =x) = F (x) = 1 n x a+1 n for a x b 0 otherwise, 0 for x<a for a x b 1 for x>b, E(X) = 1 2 (a + b) andv (X) =n2 1 12, where n = b a + 1 denotes the number of values that X may take. Example. Throwing a dice: E(X) = 1+6 =3.5 andv (X) = 62 1 2 12 thesameresultsaswefoundearlier. 2.92, 174

Bernoulli distribution (Bernoulli-jakaumaa) Let X be a random variable with possible values 0 and 1, and let P (X =1) = p. That is, the probability distribution of X is x 0 1 P (X = x) 1 p p or written as a mathematical formula p P (X =x) = x (1 p) 1 x, if x [0, 1] 0, otherwise. We denote X Ber(p), meaning \X is Bernoulli distributed with parameter p". An example of a Bernoulli random variable is the result of a toss of a coin with Head, say, equal to one and Tail equal to zero. Note. p is called the probability of "success", and q =1 p the probability of "failure". E[X] =0 p 0 (1 p) 1 +1 p(1 p) 0 = p V [X]=E[X 2 ] (E[X]) 2 =p p 2 =p(1 p)=pq 175

Sums of independent Bernoulli trials Assume that the probability that your computer will fail to perform a certain calculation within a limited time is P (F )=0.1, such that the probability of success is P (S) = 0.9. We may model this as a Bernoulli random variable with p =0.9 andq =0.1. Now suppose you let your computer try solving di erent but equally di±cult calculations (same probability of success p) three times and you're interested in the number of successes X within these 3 trials. Note that X may be thought of as the sum of three independent Bernoulli trials with success probability p. The 2 2 2 = 8 possible outcomes of the sequences of sucesses and failures may be systematically arranged as follows FFF FFS FSS SSS FSF SFS SFF SSF X =0 X =1 X =2 X =3 176

We wish to nd the probability distribution of X. By independence: P (X =0) = P (FFF )=0.1 0.1 0.1 =0.001, P (X =3) = P (SSS) =0.9 0.9 0.9 =0.729. Similiarly, P (FFS)=P (FSF)=P (SFF)= 0.1 0.1 0.9=0.009, P (FSS)=P (SFS)=P (SSF)= 0.1 0.9 0.9=0.081. In order to nd P (X =1) and P (X =2) we note that their constituting events are mutually disjoint and that there are 3 = 3 such events for the event X =1, and 1 3 = 3 such events for the event X =2, 2 such that we may write: 3 P (X =1) = 1 3 P (X =2) = 2 0.1 0.1 0.9 =0.027, 0.1 0.9 0.9 =0.243. All these probabilities may be expressed as: 3 P (X =x) = 0.9 x 0.1 3 x, x =0, 1, 2, 3. x 177

Binomial distribution (Binomijakauma) Suppose a Bernoulli trial with success probability p is repeated independently n times (e.g. tossing a coin n times). Then the probability of x successes, x =0, 1,...,n can be calculated with the binomial distribution: where n x P (X =x) = = n x p x (1 p) n x, x =0, 1,...,n 0, otherwise n!, m!=m(m 1)(m 2) 1. x!(n x)! A binomial variable random variable is a sum of independent Bernoulli variables, that is, X = n i=1 X i where X i Ber(p). So μ =E[X] = n i=1 E[X i ]= n i=1 p = np, and σ 2 = V [X] = σ = n i=1 V [X i ]=np(1 p) =npq, np(1 p) = npq. 178

Example: (Football pool betting continued.) Football pool betting consists of guessing the results of 13 matches, for each of which there are 3 possible outcomes. If the guess is completely random, then the probability of guessing the result of any single match right is p = 1 and the number of correctly guessed 3 matches is binomially distributed with parameters n = 13 and p = 1 3 (X Bin(13, 1 )). Then: 3 E(X) =13 1 1 4.33, V(X) =13 3 3 2 3 2.89 and, for example, P (X =10)= 13 10 13 10 1 2 0.001435 10 3 3 Example. Two percents of a product are defective. If a lot of 100 items are ordered what is the probability that there are no defective items? What is the probability that there are at least two defective items? Let X denote the number of defective items in the lot. Then possible values of X are X =0, 1,...,100. Then X Bin(100, 0.02), such that 100 P (X =0) = (0.02) 0 (0.98) 100 =(0.98) 100 0.1326 0 P (X 2) = 1 P (X 1) = 1 (P (X =0)+P (X =1)) =1 (0.98) 100 P (X=0) 100(0.02)(0.98) 99 P (X=1) 0.2707 0.5967 Note. The expected value is E [X] =100 0.02 = 2. 179

Getting binomial probabilities from Excel If the number of trials n is large, the calculation of binomial probabilities can become quite tedious. Excel can calculate both binomial probabilities b(x; n, p) =P (X =x) = n p x (1 p) n x x and cumulative binomial probabilities B(x; n, p) =P (X x) = x k=0 n k p k (1 p) n k using the syntax BINOMDIST(n,x,[cumulative]). Example: b(5; 18, 0.2) =BINOMDIST(5,18,0.2,FALSE), B(2; 16, 0.05) =BINOMDIST(2,16,0.05,TRUE). 180

Hypergeometric Distribution (Hypergeometrinen jakauma) Suppose we are interested in the number of defective items in a sample of n items from a lot containing N units, of which a are defective. The probability that the rst drawing will yield a defective item is N a, but for the second drawing it is N 1 a 1 or a N 1, depending on whether or not the rst unit was defective. Thus, the trials are not independent and the binomial distribution does not apply! The binomial distribution would apply if we do sampling with replacement (takaisinpanootanta), namely, if each unit selected for the sample is replaced before the next sample is drawn. For sampling without replacement note that x sucesses (defectives) can be chosen in a x ways, the n x failures can be chosen in N a n x ways, such that x sucesses and n x failures can be chosen in a N a x n x ways. Also, n objects can be chosen from a set of N objects in N n ways. 181

The probability of getting x successes in a sample of size n is therefore P (X =x) = a N a x n x N n, x =0, 1,...,n; where x cannot exceed a and n x cannot exceed N a. This equation de nes the hypergeometric distribution HG(N, a, n) with sample size n, population size N, anda sucesses in the population. Hypergeometric probabilities may be calculated in Excel using the command HYPGEOMDIST. The expected value and variance of the hypergeometric distribution are N n E(X) =np and V (X) =npq N 1 with p = a/n and q =1 p. For N n the hypergeometric distribution approaches the binomial distribution with parameters n and p = a/n. As a rule of thumb the binomial approximation starts being useful for N 10n. 182

Example. Consider a box with 20 balls, ve of which are black. Taking out 6 balls without replacement, what is the probability of picking 2 black balls? X := number of black balls HG(20, 5, 6): P (X =2)= 5 20 5 2 6 2 20 6 0.352 If each ball was returned before picking the next ball, then X Bin(n, p) withn =6and p =5/20 = 1/4, such that: P (X =2)= 6 1 3 4 0.297 2 42 4 183

Poisson distribution (Poisson-jakauma) The Poisson distribution is used for modelling the occurence of events during a xed time interval or the number of rare successes in a very large number of trials, such as: - the number of misprints on a bookpage, - the number of goals during a football match, - the number of telephone calls during a xed time interval. Poisson distribution: X Poi(λ) P (X = x) = λx x! e λ, for x =0, 1, 2,... andzerootherwise,whereλ denotes the average number of events per unit time interval, and e =2.71828... is the base of the natural logarithm. Excel Syntax: POISSON(x, λ,[cumulative]). The expected value and variance are E[X] =λ and V [X] =λ. 184

Example. Suppose that in a certain area there are on average ve tra±c accidents per month, that is X=number of tra±c accidents Poi(5), such that E(X) = V (X) = 5. The probability of 4 accidents in a given month is then P (X =4)= 54 4! e 5 0.1755. The Poisson distribution can be used as an approximation for the binomial distribution if p is "small" and n is large (rules of thumb: p 0.05 and n 20). Then λ = np is used. Example. Using the Poisson approximation for our earlier example of the binomial distribution we get for X Bin(100, 0.02) with λ = np = 100 0.02 = 2: and P (X =0) 20 0! e 2 0.135 P (X 2) = 1 P (X =0) P (X =1) 1 e 2 2e 2 0.594 This is pretty close to the exact results found earlier: P (X =0)=0.133 and P (X 2) = 0.597. 185

The following conditions must be ful lled for the Poisson distribution to be applicable: 1. The number of events in nonoverlapping time intervals must be independent. 2. The probability of occurence must be the same in all time intervals of the same length. 3. The probability of more than one event occuring during a short interval must be small relative to the occurence of only one event. 4. The probability of an event occuring in a short interval must be approximately proportional to the interval's length. Note. The parameter λ is interval-speci c in the sense that it denotes the average number of occurrences during whatever is chosen as a unit time interval. It may therefore sometimes be necessary to rst transform the average number of occurences to be measured in the applicable time unit. 186

Geometric Distribution (Geometrinen jakauma) Cosider repeating a Bernoulli trial with success probability p, untilthe rstsuccessoc- curs, and denote with X the number of the rst successful trial. Then X follows a geometric distribution with parameter p, denoted by X Geo(p). Its density function is P (X =x) = p(1 p) x 1, for x =1, 2,... 0, otherwise. E(X) = 1 p and V (X) = 1 p p 2. Example. A gambler decides to play lotto every week until the rst time he guesses 7 out of 39 numbers right. What is the expected waiting time for this to happen? We learned earlier that there are 39 7 = 15 380 937 possibilities to draw 7 out of 39 numbers. The waiting time for the lottery winnings is therefore geometrically distributed with parameter p = 1/15 380 937, such that: E(X) = 1 p = 15 380 937 weeks 295 787 years. 187