Binomial Sampling and the Binomial Distribution



Similar documents
4. Continuous Random Variables, the Pareto and Normal Distributions

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

Lecture 2 Binomial and Poisson Probability Distributions

CA200 Quantitative Analysis for Business Decisions. File name: CA200_Section_04A_StatisticsIntroduction

6.3 Conditional Probability and Independence

WHERE DOES THE 10% CONDITION COME FROM?

Week 4: Standard Error and Confidence Intervals

Descriptive Statistics and Measurement Scales

Introduction to Hypothesis Testing

Unit 4 The Bernoulli and Binomial Distributions

3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved.

The normal approximation to the binomial

Characteristics of Binomial Distributions

The Binomial Probability Distribution

Point and Interval Estimates

Lecture Notes Module 1

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

6.4 Normal Distribution

Fairfield Public Schools

Random variables, probability distributions, binomial random variable

Estimation and Confidence Intervals

Basics of Statistical Machine Learning

AMS 5 CHANCE VARIABILITY

Means, standard deviations and. and standard errors

CHAPTER 2 Estimating Probabilities

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

An Introduction to Basic Statistics and Probability

5.1 Radical Notation and Rational Exponents

Normal distribution. ) 2 /2σ. 2π σ

Chapter 5. Random variables

Chapter 4. Probability and Probability Distributions

6 PROBABILITY GENERATING FUNCTIONS

The Binomial Distribution

SOLUTIONS: 4.1 Probability Distributions and 4.2 Binomial Distributions

Figure 1. A typical Laboratory Thermometer graduated in C.

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

CALCULATIONS & STATISTICS

Stat 20: Intro to Probability and Statistics

2 ESTIMATION. Objectives. 2.0 Introduction

PROBABILITY AND SAMPLING DISTRIBUTIONS

5.1 Identifying the Target Parameter

You flip a fair coin four times, what is the probability that you obtain three heads.

Chapter 3. Distribution Problems. 3.1 The idea of a distribution The twenty-fold way

AP: LAB 8: THE CHI-SQUARE TEST. Probability, Random Chance, and Genetics

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

Chapter 4 Lecture Notes

WEEK #23: Statistics for Spread; Binomial Distribution

A Few Basics of Probability

Reflections on Probability vs Nonprobability Sampling

Simple Random Sampling

Week 3&4: Z tables and the Sampling Distribution of X

STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE

Multiple Imputation for Missing Data: A Cautionary Tale

Section 6-5 Sample Spaces and Probability

Lecture 3: Continuous distributions, expected value & mean, variance, the normal distribution

Statistical tests for SPSS

Probability and Statistics Vocabulary List (Definitions for Middle School Teachers)

Statistics courses often teach the two-sample t-test, linear regression, and analysis of variance

A POPULATION MEAN, CONFIDENCE INTERVALS AND HYPOTHESIS TESTING

COMP6053 lecture: Relationship between two variables: correlation, covariance and r-squared.

LAB : THE CHI-SQUARE TEST. Probability, Random Chance, and Genetics

Chapter 5. Discrete Probability Distributions

Math 58. Rumbos Fall Solutions to Review Problems for Exam 2

2. Discrete random variables

Sample Size and Power in Clinical Trials

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Standard Deviation Estimator

The Math. P (x) = 5! = = 120.

SIMULATION STUDIES IN STATISTICS WHAT IS A SIMULATION STUDY, AND WHY DO ONE? What is a (Monte Carlo) simulation study, and why do one?

Descriptive statistics Statistical inference statistical inference, statistical induction and inferential statistics

Decision Making under Uncertainty

Simple Regression Theory II 2010 Samuel L. Baker

Probability Distributions

Exploratory Data Analysis

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

CONTENTS OF DAY 2. II. Why Random Sampling is Important 9 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE

2DI36 Statistics. 2DI36 Part II (Chapter 7 of MR)

1. How different is the t distribution from the normal?

CHI-SQUARE: TESTING FOR GOODNESS OF FIT

Lecture 19: Chapter 8, Section 1 Sampling Distributions: Proportions

Problem sets for BUEC 333 Part 1: Probability and Statistics

Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY

Introduction to Statistics for Psychology. Quantitative Methods for Human Sciences

CONTINGENCY TABLES ARE NOT ALL THE SAME David C. Howell University of Vermont

The normal approximation to the binomial

Normal Distribution as an Approximation to the Binomial Distribution

Chapter 8 Hypothesis Testing Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing

Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Section 14 Simple Linear Regression: Introduction to Least Squares Regression

The Null Hypothesis. Geoffrey R. Loftus University of Washington

Types of Error in Surveys

17. SIMPLE LINEAR REGRESSION II

Basic Probability Concepts

TEACHER NOTES MATH NSPIRED

Introduction to Probability

REPEATED TRIALS. The probability of winning those k chosen times and losing the other times is then p k q n k.

2 Binomial, Poisson, Normal Distribution

Transcription:

Binomial Sampling and the Binomial Distribution Characterized by two mutually exclusive events." Examples: GENERAL: {success or failure} {on or off} {head or tail} {zero or one} BIOLOGY: {dead or alive} {captured or not captured} {reported or not reported} These events are outcomes" from a single trial." Binomial or Bernoulli trials. For n trials one has y successes." This is standard, general symbolism. Then yis an integer, 0 Ÿ y Ÿ n. The binomial parameter, denoted p, is the probability of success; thus, the probability of failure is 1 p or often denoted as q. Denoting success or failure to p is arbitrary and makes no difference. Obviously, p+ q = 1 because the events are assumed mutually exclusive and exhaustive (a coin must be a head or a tail and cannot remain resting on its edge!). This is a good place to point out that these are only (approximating) models, and not full truth. Truth is not a model. A model, by definition, is an approximation (and not truth). In fact, out of a few million coin flips, one or more coins will remain standing on edge that is truth. The binomial model does not allow such reality. The idea is more easily understood for the outcomes dead" or alive." In many cases, it is clear that an animal is either dead or alive; however, such a simple classification might be difficult in cases where an animal is close to death." Even here, the binomial model might be a very good approximation to truth or full reality. We are, of course, interested in models that are very good approximations! Of course, p is continuous and able to take any value between between 0 and 1 and including 0 and 1. 0 Ÿp Ÿ1. It is likewise somewhat obvious that an estimator of the probability of success is merely ^p = y/n = number of successes/number of trials. The estimator ^p is unbiased; some other useful quantities are: E( y) = np Here is an example where the expectation is symbolized we will employ this in many ways starting with lecture 4. Let n = 100 flips of a fair coin (thus p = 0.5). Then E( y ) = 100 0.5 =

50. This was a case where the expectation of a statistic y was used. This procedure is common in modeling data. var( y) = npq = np(1 p) var( ^p) = ( pq)/ n ^var( ^ p) = ( ^^ pq)/ n ^se( ^p) = È( ^^ pq)/ n = È^var( ^p). You should recognize these from your earlier classes in statistics. These results show that statistics such as y = number of successes also have expected values. The use of hats" to denote estimators or estimates is important. One must no confuse as estimate (e.g., 119) with the parameter (e.g., 188, but usually not known to the investigator). The expression var( ^p ) denotes the sampling variance; the uncertainty associated with the use of an estimator and sample data. Finally, do not think that the standard error of the estimator, ^p in this case, is anything other than the square root of the variance of the estimator ^p. Note too, that ^var( ^p) = (se( ^ ^p )) #. Beginning classes in statistics often use simplified notation to ease learning; however, occasionally this becomes a hindrance in more advanced work. For example, often the population mean. is estimated by the sample mean and denoted as -x ; this seems awkward and we will merely use ^.. The estimate of the population variance 5 # is often denoted by s # or S # and this seems particularly poor, at least for FW663. Consider the usual simple example from ST001 class. A random sample of n = 30 pigs are drawn and weighed exactly. The sample data are the pig weights w ", w #, w $,..., w $!. An estimate of the variability in pig weights across the pigs sampled is the usual standard deviation, Estimated standard deviation = ^5 = Ê! ( w w- )# n 1. 3 This is an estimate of the population standard deviation,. Because a random 5 sample of the population was taken, the sample standard deviation can be taken as a valid measure of the variation in pig weights in the population. Note, if only pigs were weighed close to the road, then the standard deviation would be a measure of the variation in the sample of 30 pigs; without a valid way to make an inductive inference to the population of pigs.

Now, instead of wanting a measure of variation in the weights of pigs in a population (based on a sample), suppose one wanted to estimate some parameter in the pig population? The obvious parameter here is the mean weight. (or, perhaps. A to denote it is the mean weight). It turns out that the estimator of this mean is ^. =! w n 3 and this estimator is unbiased, asymptotically minimum variance and asymptotically normal (it is an MLE" see below). Of course, we need a measure of its precision or repeatability; this is the estimated sampling variance, or its estimated standard error, ^var( ^. ) =!( w w- )# nn ( 1) ^se(. ^) = È^var(. ^). 3 Whenever we need a measure of the precision of an estimator, we turn to the sampling variance or standard error (and perhaps coefficients of variation or confidence intervals). Note, these measures of precision or uncertainty are, themselves, only estimates (note the hat" to indicate this). If a measure of the variation in the population members is desired, one is interested in estimating the population standard deviation, 5. The standard deviation does not change with sample size; it is an innate value of the population. It has nothing to do with sampling, except that large sample might often permit a better estimate of this population parameter. You might turn to a random page in a scientific journal in your field of interest and ask if these concepts about population variation 5 # versus sampling variation var( ) ^ ) are made clear in tables! Quantities such as the sampling variance are parameters and they have estimators. For example, in the case of the binomial model, the sampling variance is and its estimator is var( ^p ) = p (1 p )/ n ^var( ^p ) = ^ p (1 ^ pn )/.

This ^ notation might seem irritating at first, but it becomes essential in real world problems. Why? Perhaps the estimate of the sampling variance is severely negatively biased in a particular situation; then one must worry about confidence interval coverage being less than the nominal level (e.g., 0.90). It is easy to blur the distinction between a parameter and its estimate. For example, consider the person who has just weighted 50 male and 50 female pigs and computed their sample means:. ^ = 34Kg and ^ 7. 0 = 37Kg. Are these sample means significantly" different, he asks? No t or F test is needed to answer this question; of course they are different! 34 Á 37. Perhaps he really meant to ask if the estimated sample means provided any evidence that the population parameters (. 7 and. 0) differed by sex? Of course, he did. The var( ^ p) is really shorthand for var( ^ p ± model), a measure of the sampling variation (uncertainty) of ^p as an estimator of true p, given a model. If sample size Å _, var( p^ ) goes to 0. If true p varies from coin to coin, var( p^ ) tells us nothing at all about that other source of variation. Here, there are two variance components. Consider a population" of loaded coins, each with its own probability, p, of landing heads up and p varies over coins (i.e., individual heterogeneity"). Then there is a population variance of p, namely 5 2 p. Draw a sample of coins of size k, flip each one ntimes, get ^ p, ^ p ^ p " 2, á, k. ^var( ^p ) tells us nothing about. The variation among the ^p, ^p ^ i 5 2 p " 2, á, pk reflects both variation among the set of true p's ( 5 2 :) and the k sampling variances (each one conditional on p of that coin), var( ^i p model), iœ1, 2, á, k. Here, there is a sampling variance and a variance across individuals. We will see other examples of variance components." Things to ponder: Do var( p^ ) and 5 # differ? Why? How does one know which estimator to use? Are there general methods for finding such estimators? What is the sampling variance of some parameter if a census is conducted? Why is it called a sampling variance? Review some old quizzes as a way to bring up other issues and consider them.

Summary Binomial Random Variables: 1. n identical trials (e.g. a flip of 20 pennies) 2. each Bernoulli trial results in one of two mutually exclusive outcomes (e.g., head or tail) 3. the Prob{success} = p on each trial and this remains constant 4. Trials are independent events 5. y= number of successes; the random variable after n trials. 6. p is the probability of success 7. ^pis the estimatorof the probability of success; ^p= y/n. 8. ^p = 0.73 is an estimate. 9. The precision of the estimator is measured by ^var( ^p) = ( ^^ pq)/ n, the estimated sampling variance (or, the square root of ^var( p^ ), the estimated standard error ^se( ^p )). Extended Example: Consider n= 100 unfair pennies where the underlying binomial parameter is p= 0.70 (we know this value in this example). In a prior FW-663 class 11 students each made 50 surveys each involving flipping 100 unfair pennies. Thus, we have the results of 11x50=550 surveys, where each survey involved flipping 100 pennies. For each survey, we have n=100, y= the number of heads observed. Clearly, = (H) = 0.7, = 0.3, and + =1. The estimator is =. The histogram below ^ p p q p q p y/n shows the frequency for the 550 independent surveys.

Frequency 60 50 40 30 20 10 0 54 56 58 60 62 64 66 68 70 72 74 76 78 80 82 84 Number of Heads Note that y is a random variable and has a probability distribution (as above). It is actually a discrete random variable ( y cannot be 37.54 heads). In fact, we will see that this binomial random variable is approximately normally distributed under certain conditions. Does the distribution of y above look somewhat normal"? ^- p = E( ^ ^p) = 0.699 and bias is estimated as E( ^ ^p) p = 0.699 0.70 = -0.001. This estimate of the bias suggests that bias is trivial; in fact, we know that the bias of this estimator is 0. Our estimate of bias is quite good in this case. Note, percent relative bias (PRB) and relative bias (RB) are often useful: PRB = 100 [E( ^^ )) )] ) and RB =[E( ^^ )) )] ).

Binomial Probability Function This function is of passing interest on our way to an understanding of likelihood and loglikehood functions. We will usually denote probability functions as f and, in this case, fy () which is strictly positive and a function of the random variable y, the number of successes observed in n trials. We will return to a coin flipping survey where the outcomes are head (H) with probability por tail (T) with probability 1 p. The binomial probability function is fy ( ± n, p ) = Š n y p C (1 p) 8 C. The left hand side is read the probability of observing y, given nflips with underlying parameter p. Thus, it is assumed that we know the exact values of n and p; only y is a random variable. How is this useful? Let us start with a coin presumed to be fair" i.e., p = 0.5. Let n = 11 trials or flips and yis the number of heads. The outcomes are below: HHTHHHTTHHT; thus y = 7 = the number of heads (H). The order is not important, thus the outcomes could have been written as HHHHHHHTTTT and, in terms of probabilities (which are multiplicative under the assumption of independence), we could write C 8 C ( "" ( p (1 p) as p (1 p). Of course, p is assumed to be 0.5 if the coin is fair. Then, part of the probability function is written simply as 0.5 ( (1 0.5)"" (. The full probability function, including the binomial coefficient, is then 11 ( "" ( y fy± ( 11, 0.5) = Š 0.5 (1 0.5)

This function involves only one variable, y= 7. This can be evaluated numerically; fy± ( 11, 0.5) = (330) x 0.5 "" = 0.16. So, in summary, the probability of getting 7 heads out of 11 flips of a fair coin is 0.16. Common sense leads one to believe that this is not an unusual event, the mathematics allows one to quantify the issue. Th ability to quantify probabilities is especially important when the model and data are much more complex (as we shall soon see). In such case, intuition is often of little help. The theory of probability was well advanced 200 years ago so, by now, you can only imagine that many probability functions exist and are useful in a huge array of practical problems. However, probability functions are of little direct interest in biology because we rarely know the parameters. In fact, biologists have a reverse problem they have data (the n and the y) but do not know the parameters (the p). This leads to the likelihood function!

Binomial distribution for 8œ""ß :œ!þ&þ 0.35 0.3 0.25 Probability 0.2 0.15 0.1 0.05 0 0 1 2 3 4 5 6 7 8 9 10 11 y Binomial distribution for 8œ""ß :œ!þ"&þ 0.35 0.3 0.25 Probability 0.2 0.15 0.1 0.05 0 0 1 2 3 4 5 6 7 8 9 10 11 y

Comparison of binomial distributions for 8œ""ß :œ!þ&(red) and (green). 8œ""ß :œ!þ"& Probability 0.35 0.3 0.25 0.2 0.15 0.1 0.05 0 0 1 2 3 4 5 6 7 8 9 10 11 y