# , for x = 0, 1, 2, 3,... (4.1) (1 + 1/n) n = b x /x! = e b, x=0

Save this PDF as:

Size: px
Start display at page:

Download ", for x = 0, 1, 2, 3,... (4.1) (1 + 1/n) n = 2.71828... b x /x! = e b, x=0"

## Transcription

1 Chapter 4 The Poisson Distribution 4.1 The Fish Distribution? The Poisson distribution is named after Simeon-Denis Poisson ( ). In addition, poisson is French for fish. In this chapter we will study a family of probability distributions for a countably infinite sample space, each member of which is called a Poisson Distribution. Recall that a binomial distribution is characterized by the values of two parameters: n and p. A Poisson distribution is simpler in that it has only one parameter, which we denote by θ, pronounced theta. (Many books and websites use λ, pronounced lambda, instead of θ.) The parameter θ must be positive: θ > 0. Below is the formula for computing probabilities for the Poisson. P(X = x) = e θ θ x, for x = 0, 1, 2, 3,.... (4.1) x! In this equation, e is the famous number from calculus, e = lim n (1 + 1/n) n = You might recall from the study of infinite series in calculus, that for any real number b. Thus, b x /x! = e b, x=0 P(X = x) = e θ θ x /x! = 1. x=0 x=0 Thus, we see that Formula 4.1 is a mathematically valid way to assign probabilities to the nonnegative integers. The mean of the Poisson is its parameter θ; i.e. µ = θ. This can be proven using calculus and a similar argument shows that the variance of a Poisson is also equal to θ; i.e. σ 2 = θ and σ = θ. 33

2 When I write X Poisson(θ) I mean that X is a random variable with its probability distribution given by the Poisson with parameter value θ. I ask you for patience. I am going to delay my explanation of why the Poisson distribution is important in science. Poisson probabilities can be computed by hand with a scientific calculator. Alternative, you can go to the following website: I will give an example to illustrate the use of this site. Let X Poisson(θ). The website calculates two probabilities for you: P(X = x) and P(X x). You must give as input your value of θ and your desired value of x. Suppose that I have X Poisson(10) and I am interested in P(X = 8). I go to the site and type 8 in the box labeled Poisson random variable, and I type 10 in the box labeled Average rate of success. I click on the Calculate box and the site gives me the following answers: P(X = 8) = (Appearing as Poisson probability ) and P(X 8) = (Appearing as Cumulative Poisson probability ). From this last equation and the complement rule, I get P(X 9) = P(X > 8) = 1 P(X 8) = = It can be shown that if θ 5 the Poisson distribution is strongly skewed to the right, whereas if θ 25 it s probability histogram is approximately symmetric and bell-shaped. This last statement suggests that we might use the snc to compute approximate probabilities for the Poisson, provided θ is large. For example, suppose that X Poisson(25) and I want to calculate P(X 30). We will use a modification of the method we learned for the binomial. First, we note that µ = 25 and σ = 25 = 5. Using the continuity correction, we replace P(X 30) with P(X 29.5). We now standardize: P(X 29.5) = P(Z ( )/5) = P(Z 0.90). Finally, we approximate this probability for Z by using the snc and obtain With the help of the website, I find that the exact probability is To summarize, to approximate P(X x) for X Poisson(θ), Calculate z = (x 0.5 θ)/ θ. Find the area under the snc to the right of z. If θ is unknown we can use the value of X to estimate it. The point estimate is x and, following the presentation for the binomial, we can use the snc to obtain an approximate confidence interval for θ. The result is: x ± z x. 34

3 Here is an example of its use. Ralph assumes that X has a Poisson distribution, but does not know the value of θ. He observes x = 30. His point estimate of the mean is 30 and his 95% confidence interval is 30 ± = 30 ± 10.7 = [19.3, 40.7]. We will now investigate the accuracy of the snc approximation. Suppose that, in fact, θ = 40. The 95% confidence interval will be correct if, and only if, X 1.96 X 40 X X. After algebra, this becomes (31 X 55). The probability of this event, from the website, is , which is pretty close to the desired I repeated this analysis (calculating the exact probability that the CI is correct) for several values of θ; my results are below. θ: Exact Prob. of Correct Interval In my opinion, the snc approximation works adequately for θ 40. If you believe that θ might be smaller than 40 (and evidence of this would be if X was smaller than 40), then you might want to use an exact method, as I illustrated for the binomial. There is a web site that will do this for you; go to: This site can be used for one- or two-sided CI s. Here is an example. Bart assumes that X Poisson(θ) but does not know the value of θ. He observes X = 3 and wants to obtain: The two-sided 95% CI for θ; and The upper one-sided 95% CI for θ. I will use the website to find Bart s CI s. I type 3 (the value of X) into the Observed Events: box and click on compute. (I don t need to specify the confidence level b/c the 95% two-sided CI is the default answer for this site.) I get [0.6187, ] as the exact two-sided 95% CI for θ. For the one-sided CI, I scroll down and type 5 in the upper tail box and 0 in the lower tail box. Then I scroll up and hit compute. I get the CI: [ , ]. This is clearly a computer error round-off error b/c the lower bound must be 0. So, the answer is that is the 95% upper bound for θ. 35

4 4.2 Poisson Approximation to the Binomial Earlier I promised that I would provide some motivation for studying the Poisson distribution. We have seen that for the binomial, if n is moderately large and p is not too close to 0 (remember, we don t worry about p being close to 1) then the snc gives good approximations to binomial probabilities. In this section we will see that if p is close to 0 and n is large, the Poisson can be used to approximate the binomial. I will show you the derivation of this fact below. If you have not studied calculus and limits, you might find it to be too difficult to follow. This proof will not be on any exam in this course. Remember, if X Bin(n, p), then for a fixed value of x, P(X = x) = n! x!(n x)! px q n x. Now, replace p in this formula by θ/n. In my limit argument below, as n grows, θ will remain fixed which means that p = θ/n will become smaller. We get: P(X = x) = θx x! [ n! (n x)!n x (1 θ/n) x](1 θ/n)n. Now the term in the square brackets: n! (n x)!n x (1 θ/n) x, converges (i.e. gets closer and closer) to 1 as n, so it can be ignored for large n. As shown in calculus, as n, (1 θ/n) n converges to e θ. The result follows. In the old days this result was very useful. For very large n and small p and computations performed by hand, the Poisson might be preferred to working with the binomial. For example, if X Bin(1000,0.003) we get the following exact and approximate probabilities. 36

5 Exact Approximate x P(X = x) P(X = x) Next, we will consider estimation. Suppose that n = 10,000 and there are x = 10 successes observed. The website for the exact binomial confidence interval gives [0.0005, ] for the 95% two-sided confidence interval for p. Alternatively, one can treat X as Poisson in which case the 95% two-sided confidence interval for θ is [4.7954, ]. But remember that the relationship between binomial and Poisson requires us to write p = θ/n; thus, a confidence interval for p, in this example, is the same as a confidence interval for θ/ Thus, by using the Poisson approximation, we get that [0.0005, ] is the 95% two-sided confidence interval for p. That is, to four digits after the decimal point, the two answers agree. Now, I would understand if you feel, Why should we learn to do the confidence interval for p two ways? Fair enough; but computers ideally do more than just give us answers to specific questions; they let us learn about patterns in answers. For example, suppose X Poisson(θ) and we observe X = 0. From the website, the 95% one-sided confidence interval for θ is [0, ]. Why is this interesting? Well, I have said that we don t care about cases where p = 0. But sometimes we might hope for p = 0. Borrowing from the movie, Armageddon, let every day be a trial and the day is a success if the Earth is hit by a asteroid/meteor that destroys all human life. Obviously, throughout human habitation of this planet there have been no successes. Given 0 successes in n trials, the above answer indicates that we are 95% confident that p /n. Just don t ask me exactly what n equals. Or how I know that the trials are i.i.d. 4.3 The Poisson Process The binomial distribution is appropriate for counting successes in n i.i.d. trials. For p small and n large, the binomial can be well approximated by the Poisson. Thus, it is not too surprising to learn that the Poisson is also a model for counting successes. Consider a process evolving in time in which at random times successes occur. What does this possibly mean? Perhaps the following picture will help. 37

6 O O O O O O O O In this picture, observation begins at time t = 0 and time passing is denoted by moving to the right on the number line. At certain times, a success will occur, denoted by the letter O placed on the number line. Here are some examples of such processes. 1. A target is placed near radioactive material and whenever a radioactive particle hits the target we have a success. 2. An intersection is observed. A success is the occurrence of an accident. 3. A hockey (soccer) game is watched. A success occurs whenever a goal is scored. 4. On a remote stretch of highway, a success occurs when a vehicle passes. The idea is that the times of occurrences of successes cannot be predicted with certainty. We would like, however, to be able to calculate probabilities. To do this, we need a mathematical model, much like our mathematical model for BT. Our model is called the Poisson Process. A careful mathematical presentation and derivation is beyond the goals of this course. Here are the basic ideas: 1. The number of successes in disjoint intervals are independent of each other. For example, in a Poisson Process, the number of successes in the interval [0, 3] is independent of the number of successes in the interval [5, 6]. 2. The probability distribution of the number of successes counted in any time interval only depends on the length of the interval. For example, the probability of getting exactly five successes is the same for interval [0, 2.5] as it is for interval [3.5, 6.0]. 3. Successes cannot be simultaneous. With these assumptions, it turns out that the probability distribution of the number of successes in any interval of time is the Poisson distribution with parameter θ, where θ = λ w, where w > 0 is the length of the interval and λ > 0 is a feature of the process, often called its rate. I have presented the Poisson Process as occurring in one dimension time. It also can be applied if the one dimension is, say, distance. For example, a researcher could be walking along a path and at unpredictable places find successes. Also, the Poisson Process can be extended to two or three dimensions. For example, in two dimensions a researcher could be searching a field for a certain plant or animal that is deemed a success. In three dimensions a researcher could be searching a volume of air, water or dirt looking for something of interest. The modification needed for two or three dimensions is quite simple: the process still has a rate, again called λ, and now the number of successes in an area or volume has a Poisson distribution with θ equal to the rate multiplied by the area or volume, whichever is appropriate. 38

7 4.4 Independent Poissons Earlier we learned that if X 1, X 2,...,X n are i.i.d. dichotomous outcomes (success or failure), then we can calculate probabilities for the sum of these guys X: X = X 1 + X X n. Probabilities for X are given by the binomial distribution. There is a similar result for the Poisson, but the conditions are actually weaker. The interested reader can think about how the following fact is implied by the Poisson Process. Suppose that for i = 1, 2, 3,..., n, the random variable X i Poisson(θ i ) and that the sequence of X i s are independent. (If all of the θ i s are the same, then we have i.i.d. The point is that we don t need the i.d., just the independence.) Define n θ + = θ i. i=1 The math result is that X Poisson(θ + ). B/c of this result we will often (as I have done above), but not always, pretend that we have one Poisson random variable, even if, in reality, we have a sum of independent Poisson random variable. I will illustrate what I mean with an estimation example. Suppose that Cathy observes 10 i.i.d. Poisson random variables, each with parameter θ. She can summarize the ten values she obtains and simply work with their sum, X, remembering that X Poisson(10θ). Cathy can then calculate a CI for 10θ and convert it to a CI for θ. For example, suppose that Cathy observes a total of 92 when she sums her 10 values. B/c 92 is so large, I will use the snc to obtain a two-sided 95% CI for 10θ. It is: 92 ± = 92 ± = [73.200, ]. Thus, the two-sided 95% CI for θ is [7.320, ]. BTW, the exact CI for 10θ is [74.165, ]. Well, the snc answer is really not very good. 4.5 Why Bother? Suppose that we plan to observe an i.i.d. sequence of random variables and that each random variable has for possible values: 0, 1, 2,.... (This scenario frequently occurs in science.) In this chapter I have suggested that we assume that each rv has a Poisson distribution. But why? What do we gain? Why not just do the following? Define p 0 = P(X = 0), p 1 = P(X = 1), p 2 = P(X = 2),..., where there is now a sequence of probabilities known only to nature. As a researcher we can try to estimate this sequence. This question is an example of a, if not the, fundamental question a researcher always considers: How much math structure should we impose on a problem? Certainly, the Poisson leads to 39

8 values for p 0, p 1, p 2,.... The difference is that with the Poisson we impose a structure on these probabilities, whereas in the general case we do not impose a structure. As with many things in human experience, many people are too extreme on this issue. Some people put too much faith in the Poisson (or other assumed structures) and cling to it even when the data make its continued assumption ridiculous; others claim the moral high ground and proclaim: I don t make unnecessary assumptions. I cannot give you any rules for how to behave; instead, I will give you an extended example of how answers change when we change assumptions. Let us consider a Poisson Process in two dimensions. For concreteness, imagine you are in a field searching for a plant/insect that you don t particularly like; i.e. you will be happiest if there are none. Thus, you might want to know the numerical value of P(X = 0). Of course, P(X = 0) is what we call p 0 and for the Poisson it is equal to e θ. Suppose it is true (i.e. this is what Nature knows) that X Poisson(0.6931) which makes Suppose further that we have two researchers: P(X = 0) = e = Researcher A assumes Poisson with unknown θ. Researcher B assumes no parametric structure; i.e. B wants to know p 0. Note that both researchers want to get an estimate of for P(X = 0). Suppose that the two researchers observe the same data, namely n = 10 trials. Who will do better? Well, we answer this question by simulating the data. I used my computer to simulate n = 10 i.i.d. trials from the Poisson(0.6931) and obtained the following data: 1, 0, 1, 0, 3, 1, 2, 0, 0, 4. Researcher B counts four occurrences of 0 in the sample and estimates P(X = 0) to be 4/10 = 0.4. Researcher A estimates θ by the mean of the 10 numbers: 12/10 = 1.2 and then estimates P(X = 0) by e 1.2 = In this one simulated data set, each researcher s estimate is too low and Researcher B does better than A. One data set, however, is not conclusive. So, I simulated 999 more data sets of size n = 10 to obtain a total of 1000 simulated data sets. In this simulation, sometimes A did better, sometimes B did better. Statisticians try to decide who does better overall. First, we look at how each researcher did on average. If you average the 1,000 estimates for A you get and for B you get Surprisingly, B, who makes fewer assumptions, is, on average, closer to the truth. When we find a result in a simulation study that seems surprising we should wonder whether it is a false alarm caused by the approximate nature of simulation answers. While I cannot explain why at this point, I will simply say that this is not a false alarm. A consequence of assuming Poisson is that, especially for small values of n, there can be some bias in the mean performance of an estimate. By contrast, the fact that the mean of the estimates by B exceeds 0.5 is not meaningful; i.e. B s method does not possess bias. I will still conclude that A is better than B, despite the bias; I will now describe the basis for this conclusion. 40

9 From the point-of-view of Nature, who knows the truth, every estimate value has an error: e = estimate minus truth. In this simulation the error e is the estimate minus 0.5. Now errors can be positive or negative. Also, trying to make sense of 1000 errors is too difficult; we need a way to summarize them. Statisticians advocate averaging the errors after making sure that the negatives and positives don t cancel. We have two preferred ways of doing this: Convert each error to an absolute error by taking its absolute value. Convert each error to a squared error by squaring it. For my simulation study, the mean absolute error is for A and for B. B/c there is a minimum theoretical value of 0 for the mean absolute error, it makes sense to summarize this difference by saying that the mean absolute error for A is 14.2% smaller than it is for B. This 14.2% is my preferred measure and why I conclude that A is better than B. As we will see, statisticians like to square errors, although justifying this in an intuitive way is a bit difficult. I will just remark that for this simulation study, the mean squared error for A is and for B it is (B/c all of the absolute errors are 0.5 or smaller, squaring the errors make them smaller.) To revisit the issue of bias, I repeated the above simulation study, but now with n = 100. The mean of the estimates for A is and for B is These discrepancies from 0.5 are not meaningful; i.e. there is no bias. 41

10 42

### Normal distribution. ) 2 /2σ. 2π σ

Normal distribution The normal distribution is the most widely known and used of all distributions. Because the normal distribution approximates many natural phenomena so well, it has developed into a

### 4. Continuous Random Variables, the Pareto and Normal Distributions

4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random

### Random variables, probability distributions, binomial random variable

Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that

### 6.4 Normal Distribution

Contents 6.4 Normal Distribution....................... 381 6.4.1 Characteristics of the Normal Distribution....... 381 6.4.2 The Standardized Normal Distribution......... 385 6.4.3 Meaning of Areas under

### Lecture 5 : The Poisson Distribution

Lecture 5 : The Poisson Distribution Jonathan Marchini November 10, 2008 1 Introduction Many experimental situations occur in which we observe the counts of events within a set unit of time, area, volume,

### Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume

### Introduction to Hypothesis Testing

I. Terms, Concepts. Introduction to Hypothesis Testing A. In general, we do not know the true value of population parameters - they must be estimated. However, we do have hypotheses about what the true

### It is time to prove some theorems. There are various strategies for doing

CHAPTER 4 Direct Proof It is time to prove some theorems. There are various strategies for doing this; we now examine the most straightforward approach, a technique called direct proof. As we begin, it

### Notes on Continuous Random Variables

Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes

### Chapter 5. The Goodness-of-Fit Test. 5.1 Dice, Genetics and Computers

Chapter 5 The Goodness-of-Fit Test 5.1 Dice, Genetics and Computers The CM of casting a die was introduced in Chapter 1. I argued that, in my opinion, it is always reasonable to assume that successive

### If f is a 1-1 correspondence between A and B then it has an inverse, and f 1 isa 1-1 correspondence between B and A.

Chapter 5 Cardinality of sets 51 1-1 Correspondences A 1-1 correspondence between sets A and B is another name for a function f : A B that is 1-1 and onto If f is a 1-1 correspondence between A and B,

### CALCULATIONS & STATISTICS

CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents

### Calculating Logarithms By Hand

Calculating Logarithms By Hand W. Blaine Dowler June 14, 010 Abstract This details methods by which we can calculate logarithms by hand. 1 Definition and Basic Properties A logarithm can be defined as

### 1 Sufficient statistics

1 Sufficient statistics A statistic is a function T = rx 1, X 2,, X n of the random sample X 1, X 2,, X n. Examples are X n = 1 n s 2 = = X i, 1 n 1 the sample mean X i X n 2, the sample variance T 1 =

### 7 Hypothesis testing - one sample tests

7 Hypothesis testing - one sample tests 7.1 Introduction Definition 7.1 A hypothesis is a statement about a population parameter. Example A hypothesis might be that the mean age of students taking MAS113X

### MAT2400 Analysis I. A brief introduction to proofs, sets, and functions

MAT2400 Analysis I A brief introduction to proofs, sets, and functions In Analysis I there is a lot of manipulations with sets and functions. It is probably also the first course where you have to take

### Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note 11

CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According

### Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur Module No. #01 Lecture No. #15 Special Distributions-VI Today, I am going to introduce

### CHAPTER 3 Numbers and Numeral Systems

CHAPTER 3 Numbers and Numeral Systems Numbers play an important role in almost all areas of mathematics, not least in calculus. Virtually all calculus books contain a thorough description of the natural,

### Chapter 7 Part 2. Hypothesis testing Power

Chapter 7 Part 2 Hypothesis testing Power November 6, 2008 All of the normal curves in this handout are sampling distributions Goal: To understand the process of hypothesis testing and the relationship

### 99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, 99.42 cm

Error Analysis and the Gaussian Distribution In experimental science theory lives or dies based on the results of experimental evidence and thus the analysis of this evidence is a critical part of the

### Standard Deviation Estimator

CSS.com Chapter 905 Standard Deviation Estimator Introduction Even though it is not of primary interest, an estimate of the standard deviation (SD) is needed when calculating the power or sample size of

### Important Probability Distributions OPRE 6301

Important Probability Distributions OPRE 6301 Important Distributions... Certain probability distributions occur with such regularity in real-life applications that they have been given their own names.

### Chapter 4. Probability and Probability Distributions

Chapter 4. robability and robability Distributions Importance of Knowing robability To know whether a sample is not identical to the population from which it was selected, it is necessary to assess the

### Some Notes on Taylor Polynomials and Taylor Series

Some Notes on Taylor Polynomials and Taylor Series Mark MacLean October 3, 27 UBC s courses MATH /8 and MATH introduce students to the ideas of Taylor polynomials and Taylor series in a fairly limited

### WHERE DOES THE 10% CONDITION COME FROM?

1 WHERE DOES THE 10% CONDITION COME FROM? The text has mentioned The 10% Condition (at least) twice so far: p. 407 Bernoulli trials must be independent. If that assumption is violated, it is still okay

### TRANSCRIPT: In this lecture, we will talk about both theoretical and applied concepts related to hypothesis testing.

This is Dr. Chumney. The focus of this lecture is hypothesis testing both what it is, how hypothesis tests are used, and how to conduct hypothesis tests. 1 In this lecture, we will talk about both theoretical

### STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables

Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables Discrete vs. continuous random variables Examples of continuous distributions o Uniform o Exponential o Normal Recall: A random

### The Normal Curve. The Normal Curve and The Sampling Distribution

Discrete vs Continuous Data The Normal Curve and The Sampling Distribution We have seen examples of probability distributions for discrete variables X, such as the binomial distribution. We could use it

### Probability and Statistics

CHAPTER 2: RANDOM VARIABLES AND ASSOCIATED FUNCTIONS 2b - 0 Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be

### CHAPTER 2. Inequalities

CHAPTER 2 Inequalities In this section we add the axioms describe the behavior of inequalities (the order axioms) to the list of axioms begun in Chapter 1. A thorough mastery of this section is essential

### MAT140: Applied Statistical Methods Summary of Calculating Confidence Intervals and Sample Sizes for Estimating Parameters

MAT140: Applied Statistical Methods Summary of Calculating Confidence Intervals and Sample Sizes for Estimating Parameters Inferences about a population parameter can be made using sample statistics for

### Hypothesis Testing COMP 245 STATISTICS. Dr N A Heard. 1 Hypothesis Testing 2 1.1 Introduction... 2 1.2 Error Rates and Power of a Test...

Hypothesis Testing COMP 45 STATISTICS Dr N A Heard Contents 1 Hypothesis Testing 1.1 Introduction........................................ 1. Error Rates and Power of a Test.............................

PROBLEM SET 1 For the first three answer true or false and explain your answer. A picture is often helpful. 1. Suppose the significance level of a hypothesis test is α=0.05. If the p-value of the test

### Chi-Square Tests. In This Chapter BONUS CHAPTER

BONUS CHAPTER Chi-Square Tests In the previous chapters, we explored the wonderful world of hypothesis testing as we compared means and proportions of one, two, three, and more populations, making an educated

### Math 4310 Handout - Quotient Vector Spaces

Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable

### Inferential Statistics

Inferential Statistics Sampling and the normal distribution Z-scores Confidence levels and intervals Hypothesis testing Commonly used statistical methods Inferential Statistics Descriptive statistics are

### 5 =5. Since 5 > 0 Since 4 7 < 0 Since 0 0

a p p e n d i x e ABSOLUTE VALUE ABSOLUTE VALUE E.1 definition. The absolute value or magnitude of a real number a is denoted by a and is defined by { a if a 0 a = a if a

### Probability Distributions

Learning Objectives Probability Distributions Section 1: How Can We Summarize Possible Outcomes and Their Probabilities? 1. Random variable 2. Probability distributions for discrete random variables 3.

### Chicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011

Chicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011 Name: Section: I pledge my honor that I have not violated the Honor Code Signature: This exam has 34 pages. You have 3 hours to complete this

### IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have

### Hypothesis tests, confidence intervals, and bootstrapping

Hypothesis tests, confidence intervals, and bootstrapping Business Statistics 41000 Fall 2015 1 Topics 1. Hypothesis tests Testing a mean: H0 : µ = µ 0 Testing a proportion: H0 : p = p 0 Testing a difference

### the number of organisms in the squares of a haemocytometer? the number of goals scored by a football team in a match?

Poisson Random Variables (Rees: 6.8 6.14) Examples: What is the distribution of: the number of organisms in the squares of a haemocytometer? the number of hits on a web site in one hour? the number of

### Chapter 2, part 2. Petter Mostad

Chapter 2, part 2 Petter Mostad mostad@chalmers.se Parametrical families of probability distributions How can we solve the problem of learning about the population distribution from the sample? Usual procedure:

### Standard Deviation Calculator

CSS.com Chapter 35 Standard Deviation Calculator Introduction The is a tool to calculate the standard deviation from the data, the standard error, the range, percentiles, the COV, confidence limits, or

### Combinatorics: The Fine Art of Counting

Combinatorics: The Fine Art of Counting Week 7 Lecture Notes Discrete Probability Continued Note Binomial coefficients are written horizontally. The symbol ~ is used to mean approximately equal. The Bernoulli

### Chapter 5: Normal Probability Distributions - Solutions

Chapter 5: Normal Probability Distributions - Solutions Note: All areas and z-scores are approximate. Your answers may vary slightly. 5.2 Normal Distributions: Finding Probabilities If you are given that

### 37.3. The Poisson distribution. Introduction. Prerequisites. Learning Outcomes

The Poisson distribution 37.3 Introduction In this block we introduce a probability model which can be used when the outcome of an experiment is a random variable taking on positive integer values and

### MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete

### Maximum Likelihood Estimation

Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for

### Hypothesis Testing. Learning Objectives. After completing this module, the student will be able to

Hypothesis Testing Learning Objectives After completing this module, the student will be able to carry out a statistical test of significance calculate the acceptance and rejection region calculate and

### Calculus for Middle School Teachers. Problems and Notes for MTHT 466

Calculus for Middle School Teachers Problems and Notes for MTHT 466 Bonnie Saunders Fall 2010 1 I Infinity Week 1 How big is Infinity? Problem of the Week: The Chess Board Problem There once was a humble

### The Binomial Probability Distribution

The Binomial Probability Distribution MATH 130, Elements of Statistics I J. Robert Buchanan Department of Mathematics Fall 2015 Objectives After this lesson we will be able to: determine whether a probability

### ORDERS OF ELEMENTS IN A GROUP

ORDERS OF ELEMENTS IN A GROUP KEITH CONRAD 1. Introduction Let G be a group and g G. We say g has finite order if g n = e for some positive integer n. For example, 1 and i have finite order in C, since

### Lies My Calculator and Computer Told Me

Lies My Calculator and Computer Told Me 2 LIES MY CALCULATOR AND COMPUTER TOLD ME Lies My Calculator and Computer Told Me See Section.4 for a discussion of graphing calculators and computers with graphing

### Probability Generating Functions

page 39 Chapter 3 Probability Generating Functions 3 Preamble: Generating Functions Generating functions are widely used in mathematics, and play an important role in probability theory Consider a sequence

### Hypothesis testing. c 2014, Jeffrey S. Simonoff 1

Hypothesis testing So far, we ve talked about inference from the point of estimation. We ve tried to answer questions like What is a good estimate for a typical value? or How much variability is there

### Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 3: Discrete Uniform Distribution Binomial Distribution Sections 3-5, 3-6 Special discrete random variable distributions we will cover

### Models for Discrete Variables

Probability Models for Discrete Variables Our study of probability begins much as any data analysis does: What is the distribution of the data? Histograms, boxplots, percentiles, means, standard deviations

### 1 Limiting distribution for a Markov chain

Copyright c 2009 by Karl Sigman Limiting distribution for a Markov chain In these Lecture Notes, we shall study the limiting behavior of Markov chains as time n In particular, under suitable easy-to-check

### MATH 2300 review problems for Exam 3 ANSWERS

MATH 300 review problems for Exam 3 ANSWERS. Check whether the following series converge or diverge. In each case, justify your answer by either computing the sum or by by showing which convergence test

### MATH 10: Elementary Statistics and Probability Chapter 7: The Central Limit Theorem

MATH 10: Elementary Statistics and Probability Chapter 7: The Central Limit Theorem Tony Pourmohamad Department of Mathematics De Anza College Spring 2015 Objectives By the end of this set of slides, you

### Probability Distributions

CHAPTER 5 Probability Distributions CHAPTER OUTLINE 5.1 Probability Distribution of a Discrete Random Variable 5.2 Mean and Standard Deviation of a Probability Distribution 5.3 The Binomial Distribution

### Probability and Statistics Vocabulary List (Definitions for Middle School Teachers)

Probability and Statistics Vocabulary List (Definitions for Middle School Teachers) B Bar graph a diagram representing the frequency distribution for nominal or discrete data. It consists of a sequence

### You flip a fair coin four times, what is the probability that you obtain three heads.

Handout 4: Binomial Distribution Reading Assignment: Chapter 5 In the previous handout, we looked at continuous random variables and calculating probabilities and percentiles for those type of variables.

### Lecture Notes Module 1

Lecture Notes Module 1 Study Populations A study population is a clearly defined collection of people, animals, plants, or objects. In psychological research, a study population usually consists of a specific

### Week 3&4: Z tables and the Sampling Distribution of X

Week 3&4: Z tables and the Sampling Distribution of X 2 / 36 The Standard Normal Distribution, or Z Distribution, is the distribution of a random variable, Z N(0, 1 2 ). The distribution of any other normal

### Introduction to the Practice of Statistics Fifth Edition Moore, McCabe

Introduction to the Practice of Statistics Fifth Edition Moore, McCabe Section 5.1 Homework Answers 5.7 In the proofreading setting if Exercise 5.3, what is the smallest number of misses m with P(X m)

### 1.7 Graphs of Functions

64 Relations and Functions 1.7 Graphs of Functions In Section 1.4 we defined a function as a special type of relation; one in which each x-coordinate was matched with only one y-coordinate. We spent most

### MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform

MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish

### Random Variable: A function that assigns numerical values to all the outcomes in the sample space.

STAT 509 Section 3.2: Discrete Random Variables Random Variable: A function that assigns numerical values to all the outcomes in the sample space. Notation: Capital letters (like Y) denote a random variable.

### Sampling Distributions and the Central Limit Theorem

135 Part 2 / Basic Tools of Research: Sampling, Measurement, Distributions, and Descriptive Statistics Chapter 10 Sampling Distributions and the Central Limit Theorem In the previous chapter we explained

### Chapter 4. Probability Distributions

Chapter 4 Probability Distributions Lesson 4-1/4-2 Random Variable Probability Distributions This chapter will deal the construction of probability distribution. By combining the methods of descriptive

### Financial Mathematics and Simulation MATH 6740 1 Spring 2011 Homework 2

Financial Mathematics and Simulation MATH 6740 1 Spring 2011 Homework 2 Due Date: Friday, March 11 at 5:00 PM This homework has 170 points plus 20 bonus points available but, as always, homeworks are graded

### Hypothesis Testing or How to Decide to Decide Edpsy 580

Hypothesis Testing or How to Decide to Decide Edpsy 580 Carolyn J. Anderson Department of Educational Psychology University of Illinois at Urbana-Champaign Hypothesis Testing or How to Decide to Decide

### So let us begin our quest to find the holy grail of real analysis.

1 Section 5.2 The Complete Ordered Field: Purpose of Section We present an axiomatic description of the real numbers as a complete ordered field. The axioms which describe the arithmetic of the real numbers

### Probabilities and Random Variables

Probabilities and Random Variables This is an elementary overview of the basic concepts of probability theory. 1 The Probability Space The purpose of probability theory is to model random experiments so

### Chi-Square Test. Contingency Tables. Contingency Tables. Chi-Square Test for Independence. Chi-Square Tests for Goodnessof-Fit

Chi-Square Tests 15 Chapter Chi-Square Test for Independence Chi-Square Tests for Goodness Uniform Goodness- Poisson Goodness- Goodness Test ECDF Tests (Optional) McGraw-Hill/Irwin Copyright 2009 by The

### SYSM 6304: Risk and Decision Analysis Lecture 3 Monte Carlo Simulation

SYSM 6304: Risk and Decision Analysis Lecture 3 Monte Carlo Simulation M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu September 19, 2015 Outline

### Math/Stats 425 Introduction to Probability. 1. Uncertainty and the axioms of probability

Math/Stats 425 Introduction to Probability 1. Uncertainty and the axioms of probability Processes in the real world are random if outcomes cannot be predicted with certainty. Example: coin tossing, stock

### The Math. P (x) = 5! = 1 2 3 4 5 = 120.

The Math Suppose there are n experiments, and the probability that someone gets the right answer on any given experiment is p. So in the first example above, n = 5 and p = 0.2. Let X be the number of correct

### Report of for Chapter 2 pretest

Report of for Chapter 2 pretest Exam: Chapter 2 pretest Category: Organizing and Graphing Data 1. "For our study of driving habits, we recorded the speed of every fifth vehicle on Drury Lane. Nearly every

### Basic Proof Techniques

Basic Proof Techniques David Ferry dsf43@truman.edu September 13, 010 1 Four Fundamental Proof Techniques When one wishes to prove the statement P Q there are four fundamental approaches. This document

### 3 Multiple Discrete Random Variables

3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f

### Sampling and Hypothesis Testing

Population and sample Sampling and Hypothesis Testing Allin Cottrell Population : an entire set of objects or units of observation of one sort or another. Sample : subset of a population. Parameter versus

### 5.1 Identifying the Target Parameter

University of California, Davis Department of Statistics Summer Session II Statistics 13 August 20, 2012 Date of latest update: August 20 Lecture 5: Estimation with Confidence intervals 5.1 Identifying

### Taylor Polynomials and Taylor Series Math 126

Taylor Polynomials and Taylor Series Math 26 In many problems in science and engineering we have a function f(x) which is too complicated to answer the questions we d like to ask. In this chapter, we will

### The normal approximation to the binomial

The normal approximation to the binomial The binomial probability function is not useful for calculating probabilities when the number of trials n is large, as it involves multiplying a potentially very

### Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?

### Binomial lattice model for stock prices

Copyright c 2007 by Karl Sigman Binomial lattice model for stock prices Here we model the price of a stock in discrete time by a Markov chain of the recursive form S n+ S n Y n+, n 0, where the {Y i }

### Summarizing Data: Measures of Variation

Summarizing Data: Measures of Variation One aspect of most sets of data is that the values are not all alike; indeed, the extent to which they are unalike, or vary among themselves, is of basic importance

### Objectives. Materials

Activity 5 Exploring Infinite Series Objectives Identify a geometric series Determine convergence and sum of geometric series Identify a series that satisfies the alternating series test Use a graphing

### Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 20

CS 70 Discrete Mathematics and Probability Theory Fall 009 Satish Rao, David Tse Note 0 Infinity and Countability Consider a function (or mapping) f that maps elements of a set A (called the domain of

### Test of Hypotheses. Since the Neyman-Pearson approach involves two statistical hypotheses, one has to decide which one

Test of Hypotheses Hypothesis, Test Statistic, and Rejection Region Imagine that you play a repeated Bernoulli game: you win \$1 if head and lose \$1 if tail. After 10 plays, you lost \$2 in net (4 heads

### We now explore a third method of proof: proof by contradiction.

CHAPTER 6 Proof by Contradiction We now explore a third method of proof: proof by contradiction. This method is not limited to proving just conditional statements it can be used to prove any kind of statement

### AP Physics 1 and 2 Lab Investigations

AP Physics 1 and 2 Lab Investigations Student Guide to Data Analysis New York, NY. College Board, Advanced Placement, Advanced Placement Program, AP, AP Central, and the acorn logo are registered trademarks

### Solutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG820). December 15, 2012.

Solutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG8). December 15, 12. 1. (3p) The joint distribution of the discrete random variables X and

### 1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as