Maximum Likelihood Estimation



Similar documents
Practice problems for Homework 11 - Point Estimation

1 Prior Probability and Posterior Probability

Exact Confidence Intervals

Principle of Data Reduction

1 Sufficient statistics

Efficiency and the Cramér-Rao Inequality

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

Basics of Statistical Machine Learning

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

INSURANCE RISK THEORY (Problems)

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page


A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails

5. Continuous Random Variables

Section 6.1 Joint Distribution Functions

Properties of Future Lifetime Distributions and Estimation

Lecture Notes on Elasticity of Substitution

Section 5.1 Continuous Random Variables: Introduction

A LOGNORMAL MODEL FOR INSURANCE CLAIMS DATA

Math 319 Problem Set #3 Solution 21 February 2002

2WB05 Simulation Lecture 8: Generating random variables

Determining distribution parameters from quantiles

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Stats on the TI 83 and TI 84 Calculator

Bayesian Statistics in One Hour. Patrick Lam

Final Mathematics 5010, Section 1, Fall 2004 Instructor: D.A. Levin

Chapter 6: Point Estimation. Fall Probability & Statistics

Lecture 7: Continuous Random Variables

Practice with Proofs

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Math 461 Fall 2006 Test 2 Solutions

Hypothesis Testing for Beginners

Reliability and Risk Analysis. Analysis of Model Tasks from Selected Application Areas Using a Computer (software R, Excel)

Estimating an ARMA Process

MATH 10: Elementary Statistics and Probability Chapter 5: Continuous Random Variables

1 Inner Products and Norms on Real Vector Spaces

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

Dongfeng Li. Autumn 2010

Lecture 8. Confidence intervals and the central limit theorem

Notes on the Negative Binomial Distribution

Part 2: One-parameter models

The Dirichlet-Multinomial and Dirichlet-Categorical models for Bayesian inference

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents

Chapter 3 RANDOM VARIATE GENERATION

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Exponential Distribution

The Exponential Distribution

Note on the EM Algorithm in Linear Regression Model

Credit Risk Models: An Overview

Separation Properties for Locally Convex Cones

Joint Exam 1/P Sample Exam 1

Inner Product Spaces

Lecture 3: Linear methods for classification

Domain of a Composition

STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables

Maximum Likelihood Estimation by R

Notes on Continuous Random Variables

Nonparametric adaptive age replacement with a one-cycle criterion

4. Continuous Random Variables, the Pareto and Normal Distributions

Statistical Machine Learning

Definition and Properties of the Production Function: Lecture

1.3. DOT PRODUCT If θ is the angle (between 0 and π) between two non-zero vectors u and v,

TOPIC 4: DERIVATIVES

6.2 Permutations continued

Probability Theory. Florian Herzog. A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T..

Inequalities of Analysis. Andrejs Treibergs. Fall 2014

Stat 5102 Notes: Nonparametric Tests and. confidence interval

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

VISUALIZATION OF DENSITY FUNCTIONS WITH GEOGEBRA

Parametric Survival Models

Lecture 8: More Continuous Random Variables

1 The EOQ and Extensions

E3: PROBABILITY AND STATISTICS lecture notes

Aggregate Loss Models

Statistics 100A Homework 8 Solutions

THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Generating Random Variables and Stochastic Processes

SOCIETY OF ACTUARIES/CASUALTY ACTUARIAL SOCIETY EXAM C CONSTRUCTION AND EVALUATION OF ACTUARIAL MODELS EXAM C SAMPLE QUESTIONS

1.1 Introduction, and Review of Probability Theory Random Variable, Range, Types of Random Variables CDF, PDF, Quantiles...

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS

Probability Generating Functions

Probability Calculator

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

Conductance, the Normalized Laplacian, and Cheeger s Inequality

Bayesian logistic betting strategy against probability forecasting. Akimichi Takemura, Univ. Tokyo. November 12, 2012

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Pattern Analysis. Logistic Regression. 12. Mai Joachim Hornegger. Chair of Pattern Recognition Erlangen University

STAT 830 Convergence in Distribution

Econometrics Simple Linear Regression

Is a Brownian motion skew?

The Standard Normal distribution

Survival Distributions, Hazard Functions, Cumulative Hazards

Generating Random Variables and Stochastic Processes

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

CHAPTER 2 Estimating Probabilities

Transcription:

Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for an unknown parameter θ. It was introduced by R. A. Fisher, a great English mathematical statistician, in 1912. Maximum likelihood estimation (MLE) can be applied in most problems, it has a strong intuitive appeal, and often yields a reasonable estimator of θ. Furthermore, if the sample is large, the method will yield an excellent estimator of θ. For these reasons, the method of maximum likelihood is probably the most widely used method of estimation in statistics. Suppose that the random variables X 1,, X n form a random sample from a distribution f(x θ); if X is continuous random variable, f(x θ) is pdf, if X is discrete random variable, f(x θ) is point mass function. We use the given symbol to represent that the distribution also depends on a parameter θ, where θ could be a real-valued unknown parameter or a vector of parameters. For every observed random sample x 1,, x n, we define f(x 1,, x n θ) = f(x 1 θ) f(x n θ) (1) If f(x θ) is pdf, f(x 1,, x n θ) is the joint density function; if f(x θ) is pmf, f(x 1,, x n θ) is the joint probability. Now we call f(x 1,, x n θ) as the likelihood function. As we can see, the likelihood function depends on the unknown parameter θ, and it is always denoted as L(θ). Suppose, for the moment, that the observed random sample x 1,, x n came from a discrete distribution. If an estimate of θ must be selected, we would certainly not consider any value of θ for which it would have been impossible to obtain the data x 1,, x n that was actually observed. Furthermore, suppose that the probability f(x 1,, x n θ) of obtaining the actual observed data x 1,, x n is very high when θ has a particular value, say θ = θ 0, and is very small for every other value of θ. Then we would naturally estimate the value of θ to be θ 0. When the sample comes from a continuous distribution, it would again be natural to try to find a value of θ for which the probability density f(x 1,, x n θ) is large, and to use this value as an estimate of θ. For any given observed data x 1,, x n, we are led by this reasoning to consider a value of θ for which the likelihood function L(θ) is a maximum and to use this value as an estimate of θ. 1

2 The meaning of maximum likelihood is as follows. We choose the parameter that makes the likelihood of having the obtained data at hand maximum. With discrete distributions, the likelihood is the same as the probability. We choose the parameter for the density that maximizes the probability of the data coming from it. Theoretically, if we had no actual data, maximizing the likelihood function will give us a function of n random variables X 1,, X n, which we shall call maximum likelihood estimate ˆθ. When there are actual data, the estimate takes a particular numerical value, which will be the maximum likelihood estimator. MLE requires us to maximum the likelihood function L(θ) with respect to the unknown parameter θ. From Eqn. 1, L(θ) is defined as a product of n terms, which is not easy to be maximized. Maximizing L(θ) is equivalent to maximizing log L(θ) because log is a monotonic increasing function. We define log L(θ) as log likelihood function, we denote it as l(θ), i.e. n l(θ) = log L(θ) = log f(x i θ) = log f(x i θ) Maximizing l(θ) with respect to θ will give us the MLE estimation. 2 Examples Example 1: Suppose that X is a discrete random variable with the following probability mass function: where 0 θ 1 is a parameter. The following 10 independent observations X 0 1 2 3 P (X) 2θ/3 θ/3 2(1 θ)/3 (1 θ)/3 were taken from such a distribution: (3,0,2,1,3,2,1,0,2,1). What is the maximum likelihood estimate of θ. Solution: Since the sample is (3,0,2,1,3,2,1,0,2,1), the likelihood is L(θ) = P (X = 3)P (X = 0)P (X = 2)P (X = 1)P (X = 3) P (X = 2)P (X = 1)P (X = 0)P (X = 2)P (X = 1) (2) Substituting from the probability distribution given above, we have n L(θ) = P (X i θ) = ( 2θ 3 ) 2 ( θ 3 ) 3 ( ) 3 ( 2(1 θ) 1 θ 3 3 Clearly, the likelihood function L(θ) is not easy to maximize. ) 2

3 Let us look at the log likelihood function l(θ) = log L(θ) = log P (X i θ) = 2 (log 2 ) 3 + log θ + 3 (log 1 ) 3 + log θ + 3 (log 2 ) 3 + log(1 θ) + 2 (log 1 ) 3 + log(1 θ) = C + 5 log θ + 5 log(1 θ) where C is a constant which does not depend on θ. It can be seen that the log likelihood function is easier to maximize compared to the likelihood function. Let the derivative of l(θ) with respect to θ be zero: dl(θ) dθ = 5 θ 5 1 θ = 0 and the solution gives us the MLE, which is ˆθ = 0.5. We remember that the method of moment estimation is ˆθ = 5/12, which is different from MLE. Example 2: Suppose X 1, X 2,, X n are i.i.d. random variables with density function f(x σ) = 1 exp ( x ) 2σ σ, please find the maximum likelihood estimate of σ. Solution: The log-likelihood function is Let the derivative with respect to θ be zero: [ l(σ) = log 2 log σ X ] i σ [ l (σ) = 1 σ + X ] i σ 2 and this gives us the MLE for σ as ˆσ = = n σ + n X i σ 2 = 0 n X i Again this is different from the method of moment estimation which is n ˆσ = n X 2 i 2n Example 3: Use the method of moment to estimate the parameters µ and σ for the normal density f(x µ, σ 2 ) = 1 } (x µ)2 exp, 2πσ 2σ 2

4 based on a random sample X 1,, X n. Solution: In this example, we have two unknown parameters, µ and σ, therefore the parameter θ = (µ, σ) is a vector. We first write out the log likelihood function as l(µ, σ) = [ log σ 1 ] 1 log 2π 2 2σ (X 2 i µ) 2 = n log σ n 1 log 2π 2 2σ 2 Setting the partial derivative to be 0, we have (X i µ) 2 l(µ, σ) µ = 1 n (X σ 2 i µ) = 0 l(µ, σ) σ = n σ + n σ 3 (X i µ) 2 = 0 Solving these equations will give us the MLE for µ and σ: ˆµ = X and ˆσ = 1 n (X i X) 2 This time the MLE is the same as the result of method of moment. From these examples, we can see that the maximum likelihood result may or may not be the same as the result of method of moment. Example 4: The Pareto distribution has been used in economics as a model for a density function with a slowly decaying tail: f(x x 0, θ) = θx θ 0x θ 1, x x 0, θ > 1 Assume that x 0 > 0 is given and that X 1, X 2,, X n is an i.i.d. sample. Find the MLE of θ. Solution: The log-likelihood function is l(θ) = log f(x i θ) = (log θ + θ log x 0 (θ + 1) log X i ) = n log θ + nθ log x 0 (θ + 1) log X i Let the derivative with respect to θ be zero: dl(θ) dθ = n θ + n log x 0 log X i = 0

5 Solving the equation yields the MLE of θ: ˆθ MLE = 1 log X log x 0 Example 5: Suppose that X 1,, X n form a random sample from a uniform distribution on the interval (0, θ), where of the parameter θ > 0 but is unknown. Please find MLE of θ. Solution: The pdf of each observation has the following form: 1/θ, for 0 x θ (3) Therefore, the likelihood function has the form 1, for 0 x L(θ) = θ n i θ (i = 1,, n) It can be seen that the MLE of θ must be a value of θ for which θ x i for i = 1,, n and which maximizes 1/θ n among all such values. Since 1/θ n is a decreasing function of θ, the estimate will be the smallest possible value of θ such that θ x i for i = 1,, n. This value is θ = max(x 1,, x n ), it follows that the MLE of θ is ˆθ = max(x 1,, X n ). It should be remarked that in this example, the MLE ˆθ does not seem to be a suitable estimator of θ. We know that max(x 1,, X n ) < θ with probability 1, and therefore ˆθ surely underestimates the value of θ. Example 6: Suppose again that X 1,, X n form a random sample from a uniform distribution on the interval (0, θ), where of the parameter θ > 0 but is unknown. However, suppose now we write the density function as 1/θ, for 0 < x < θ (4) We will prove that in this case, the MLE for θ does not exist. Proof: The only difference between Eqn. 3 and Eqn. 4 is that the value of the pdf at the two endpints 0 and θ has been changed by replacing the weak inequalities in Eqn. 3 with strict inequalities in Eqn. 4. Either equation could be used as the pdf of the uniform distribution. However, if Eqn. 4 is used as the pdf, then an MLE of θ will be a value of θ for which θ > x i for i = 1,, n and which maximizes 1/θ n among all such values. It should be noted that the possible values of θ no longer include the value θ = max(x 1,, x n ), since θ must be strictly greater than each observed value x i for i = 1,, n. Since θ can be chosen arbitrarily close to the value max(x 1,, x n ) but cannot be chosen equal to this value, it follows that the MLE of θ does not exist in this case.

6 Example 5 and 6 illustrate one shortcoming of the concept of an MLE. We know that it is irrelevant whether the pdf of the uniform distribution is chosen to be equal to 1/θ over the open interval 0 < x < θ or over the closed interval 0 x θ. Now, however, we see that the existence of an MLE depends on this typically irrelevant and unimportant choice. This difficulty is easily avoided in Example 5 by using the pdf given by Eqn. 3 rather than that given by Eqn. 4. In many other problems, as well, in which there is a difficulty of this type in regard to the existence of an MLE, the difficulty can be avoided simply by choosing one particular appropriate version of the pdf to represent the given distribution. Example 7: Suppose that X 1,, X n form a random sample from a uniform distribution on the interval (θ, θ + 1), where the value of the parameter θ is unknown ( < θ < ). Clearly, the density function is 1, for θ x θ + 1 We will see that the MLE for θ is not unique. Proof: In this example, the likelihood function is 1, for θ xi θ + 1 (i = 1,, n) L(θ) = The condition that θ x i for i = 1,, n is equivalent to the condition that θ min(x 1,, x n ). Similarly, the condition that x i θ + 1 for i = 1,, n is equivalent to the condition that θ max(x 1,, x n ) 1. Therefore, we can rewrite the the likelihood function as 1, for max(x1,, x L(θ) = n ) 1 θ min(x 1,, x n ) Thus, we can select any value in the interval [max(x 1,, x n ) 1, min(x 1,, x n )] as the MLE for θ. Therefore, the MLE is not uniquely specified in this example. 3 Exercises Exercise 1: Let X 1,, X n be an i.i.d. sample from a Poisson distribution with parameter λ, i.e., P (X = x λ) = λx e λ. x! Please find the MLE of the parameter λ. Exercise 2: Let X 1,, X n be an i.i.d. sample from an exponential distribution with the density function f(x β) = 1 β e x β, with 0 x <.

7 Please find the MLE of the parameter β. Exercise 3: Gamma distribution has a density function as f(x α, λ) = 1 Γ(α) λα x α 1 e λx, with 0 x <. Suppose the parameter α is known, please find the MLE of λ based on an i.i.d. sample X 1,, X n. Exercise 4: Suppose that X 1,, X n form a random sample from a distribution for which the pdf f(x θ) is as follows: θx θ 1, for 0 < x < 1 0, for x 0 Also suppose that the value of θ is unknown (θ > 0). Find the MLE of θ. Exercise 5: Suppose that X 1,, X n form a random sample from a distribution for which the pdf f(x θ) is as follows: 1 2 e x θ for < x < Also suppose that the value of θ is unknown ( < θ < ). Find the MLE of θ. Exercise 6: Suppose that X 1,, X n form a random sample from a distribution for which the pdf f(x θ) is as follows: e θ x, for x > θ 0, for x θ Also suppose that the value of θ is unknown ( < θ < ). a) Show that the MLE of θ does not exist. b) Determine another version of the pdf of this same distribution for which the MLE of θ will exist, and find this estimate. Exercise 7: Suppose that X 1,, X n form a random sample from a uniform distribution on the interval (θ 1, θ 2 ), with the pdf as follows: 1 θ 2 θ 1, for θ 1 x θ 2 Also suppose that the values of θ 1 and θ 2 are unknown ( < θ 1 < θ 2 < ). Find the MLE s of θ 1 and θ 2.