Notes on the Negative Binomial Distribution



Similar documents
1 Sufficient statistics

Negative Binomial. Paul Johnson and Andrea Vieux. June 10, 2013

2WB05 Simulation Lecture 8: Generating random variables

INSURANCE RISK THEORY (Problems)


Statistical Machine Learning from Data

1 Prior Probability and Posterior Probability

Statistics 100A Homework 4 Solutions

6 PROBABILITY GENERATING FUNCTIONS

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

Maximum Likelihood Estimation

Math 461 Fall 2006 Test 2 Solutions

Notes on Continuous Random Variables

4. Continuous Random Variables, the Pareto and Normal Distributions

Zeros of a Polynomial Function

Determining distribution parameters from quantiles

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

6.2 Permutations continued

Practice Problems for Homework #6. Normal distribution and Central Limit Theorem.

Probability Generating Functions

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page

Joint Exam 1/P Sample Exam 1

Part 2: One-parameter models

Zeros of Polynomial Functions

Practice Problems #4

1.1 Introduction, and Review of Probability Theory Random Variable, Range, Types of Random Variables CDF, PDF, Quantiles...

Lecture 5 : The Poisson Distribution

Lecture 8: More Continuous Random Variables

Probability Calculator

Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails

Final Mathematics 5010, Section 1, Fall 2004 Instructor: D.A. Levin

Finance 400 A. Penati - G. Pennacchi Market Micro-Structure: Notes on the Kyle Model

Bayesian Statistics in One Hour. Patrick Lam

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

Experimental Design. Power and Sample Size Determination. Proportions. Proportions. Confidence Interval for p. The Binomial Test

Binomial lattice model for stock prices

Practice problems for Homework 11 - Point Estimation

Aggregate Loss Models

Normal distribution. ) 2 /2σ. 2π σ

Polynomials. Jackie Nicholas Jacquie Hargreaves Janet Hunter

MATH2740: Environmental Statistics

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

2 Binomial, Poisson, Normal Distribution

The Exponential Distribution

ECE302 Spring 2006 HW4 Solutions February 6,

Chapter 5. Random variables

Gaussian Conjugate Prior Cheat Sheet

Chapter 3 RANDOM VARIATE GENERATION

Random variables, probability distributions, binomial random variable

Principle of Data Reduction

Beta Distribution. Paul Johnson and Matt Beverlin June 10, 2013

The normal approximation to the binomial

Math 151. Rumbos Spring Solutions to Assignment #22

Section 5.1 Continuous Random Variables: Introduction

Polynomials and Factoring

Chapter 7. Statistical Models of Operational Loss

HYPOTHESIS TESTING: POWER OF THE TEST

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Inference of Probability Distributions for Trust and Security applications

Bayesian Model Averaging Continual Reassessment Method BMA-CRM. Guosheng Yin and Ying Yuan. August 26, 2009

The Dirichlet-Multinomial and Dirichlet-Categorical models for Bayesian inference

Chapter 3 Section 6 Lesson Polynomials

WHERE DOES THE 10% CONDITION COME FROM?

Standard Deviation Estimator

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Questions and Answers

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL

SOLUTIONS: 4.1 Probability Distributions and 4.2 Binomial Distributions

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

1 The Brownian bridge construction

AFM Ch.12 - Practice Test

Math 431 An Introduction to Probability. Final Exam Solutions

JUST THE MATHS UNIT NUMBER 1.8. ALGEBRA 8 (Polynomials) A.J.Hobson

Basics of Statistical Machine Learning

Approximation of Aggregate Losses Using Simulation

Modelling the Scores of Premier League Football Matches

Basic Bayesian Methods

Homework 4 - KEY. Jeff Brenion. June 16, Note: Many problems can be solved in more than one way; we present only a single solution here.

A Predictive Probability Design Software for Phase II Cancer Clinical Trials Version 1.0.0

Exam C, Fall 2006 PRELIMINARY ANSWER KEY

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Chapter 4 Lecture Notes

Zeros of Polynomial Functions

Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics

Lecture 8. Confidence intervals and the central limit theorem

12.5: CHI-SQUARE GOODNESS OF FIT TESTS

MAS2317/3317. Introduction to Bayesian Statistics. More revision material

A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University

Zeros of Polynomial Functions

Master s Theory Exam Spring 2006

ISyE 6761 Fall 2012 Homework #2 Solutions

SPARE PARTS INVENTORY SYSTEMS UNDER AN INCREASING FAILURE RATE DEMAND INTERVAL DISTRIBUTION

Sampling Distributions

16. THE NORMAL APPROXIMATION TO THE BINOMIAL DISTRIBUTION

PROBABILITY AND SAMPLING DISTRIBUTIONS

Transcription:

Notes on the Negative Binomial Distribution John D. Cook October 28, 2009 Abstract These notes give several properties of the negative binomial distribution. 1. Parameterizations 2. The connection between the negative binomial distribution and the binomial theorem 3. The mean and variance 4. The negative binomial as a Poisson with gamma mean 5. Relations to other distributions 6. Conjugate prior 1 Parameterizations There are a couple variations of the negative binomial distribution. The first version counts the number of the trial at which the rth success occurs. With this version, 1 P (X 1 = p, r) = p r (1 p). r 1 for integer r. Here 0 < p < 1 and r is a positive integer. The second version counts the number of failures before the rth success. With this version, r + 1 P (X 2 = p, r) = p r (1 p). 1

for integer 0. If X 1 is a negative binomial random variable according to the first definition, then X 2 = X 1 r is a negative binomial according to the second definition. We will standardize on this second version for the remainder of these notes. One advantage to this version is that the range of is non-negative integers. Also, the definition can be more easily etended to all positive real values of r since there is no factor of r in the bottom of the binomial coefficient. Rather than parameterizing the negative binomial in terms of r and p, considering µ and σ as derived quantities, we could consider µ and σ primary and r and p derived. In this case and r = µ2 σ 2 µ p = r r + µ. The factor 1/r is a sort of clumping parameter. As r, the negative binomial converges in distribution to the Poisson as noted below. This fact is suggested by the variance approaching the mean as r. Viewing µ and σ as primary, we ignore the combinatorial motivation for defining the negative binomial and instead view it simply as a model for count data. A little algebra shows that σ 2 = µ + 1 r µ2. Thus the variance is always larger than the mean for the negative binomial. Since the Poisson requires the mean and variance to be equal, it is unsuitable for data with greater variance than mean; the negative binomial may be appropriate in such settings. 2 What s negative about a negative binomial? What s negative about the negative binomial distribution? What is binomial about it? This is not always clearly eplained. The name comes from applying the general form of the binomial theorem with a negative eponent. 2

1 = p r p = p r (1 q) = p r The th term in the series above is p r ( q) = ( 1) p r q = =0 ( q) r + 1 p r (1 p) which is the probability that X = where X negative binomial with parameters r and p. 3 Mean and variance The negative binomial distribution with parameters r and p has mean µ = r(1 p)/p and variance σ 2 = r(1 p)/p 2 = µ + 1 r µ2. 4 Hierarchical Poisson-gamma distribution In the first section of these notes we saw that the negative binomial distribution can be seen as an etension of the Poisson distribution that allows for greater variance. Here we eamine another derivation of the negative binomial distribution that makes the connection with the Poisson more eplicit. Suppose X Λ is a Poisson random variable and Λ is a gamma(α, β) random variable. We create a new kind of random variable by starting with a Poisson but making it more variable by allowing the mean parameter to itself be random. P (X = ) = 1 Γ(α)β α 0 e λ λ λ α 1 e λ/β dλ! 3

= = = 1!Γ(α)β α λ α+ 1 e λ(1+1/β) dλ 0 ( 1 β Γ( + 1)Γ(α)β α Γ(α + ) β + 1 ( α + 1 ) ( 1 β + 1 ) α ( 1 1 β + 1 ) α+ ) Therefore the marginal distribution of X is negative binomial with r = α and p = 1/(β + 1). If we change the parameters in the distribution of Λ so that the mean stays the same but the distribution becomes more concentrated, we epect the distribution of X to become more like that of the Poisson. That is in fact what happens. The mean of Λ is λ = αβ and the variance is αβ 2. If we let α while keeping β = λ/α, the variance of Λ goes to 0. The distribution on X converges to a Poisson distribution because as noted in Section 5.4 below, r and p 1 while keeping the mean constant. 5 Relation to other distributions Throughout this section, assume X has a negative binomial distribution with parameters r and p. 5.1 Geometric A negative binomial distribution with r = 1 is a geometric distribution. Also, the sum of r independent Geometric(p) random variables is a negative binomial(r, p) random variable. 5.2 Negative binomial If each X i is distributed as negative binomial(r i, p) then X i is distributed as negative binomial( r i, p). 4

5.3 Binomial Let Y binomial(n, p). Then P (Y < r 1) = P (X > n r). 5.4 Poisson If r and p 1 as µ stays constant, P (X = ) converges to e µ µ /!, the density for a Poisson(µ) distribution. 5.5 Beta F () = I p (r, + 1) = P (Y < p) where Y beta(r, + 1). 5.6 Gamma If p 0 as r stays constant, px converges in distribution to a gamma distribution with shape r and scale 1. 6 Conjugate prior If the likelihood function for an observation is negative binomial(r, p) and p is distributed a priori as Beta(a, b) then the posterior distribution for p is Beta(a + r, b + ). Note that this is the same as having observed r successes and failures with a binomial(r +, p) likelihood. All that matters from a Bayesian perspective is that r successes were observed and failures. This document available at http://www.johndcook.com/negative binomial.pdf. 5