Practice problems for Homework 11 - Point Estimation



Similar documents
Maximum Likelihood Estimation

1 Sufficient statistics

Math 461 Fall 2006 Test 2 Solutions

5. Continuous Random Variables

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

Statistics 100A Homework 4 Solutions

Section 5.1 Continuous Random Variables: Introduction

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Practice Problems for Homework #6. Normal distribution and Central Limit Theorem.

Notes on Continuous Random Variables

Principle of Data Reduction

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Basics of Statistical Machine Learning

1 Prior Probability and Posterior Probability

Math 425 (Fall 08) Solutions Midterm 2 November 6, 2008

Efficiency and the Cramér-Rao Inequality

2WB05 Simulation Lecture 8: Generating random variables

Math 151. Rumbos Spring Solutions to Assignment #22

4. Continuous Random Variables, the Pareto and Normal Distributions

2. Discrete random variables

Statistics 100A Homework 8 Solutions

Math 431 An Introduction to Probability. Final Exam Solutions

Homework 4 - KEY. Jeff Brenion. June 16, Note: Many problems can be solved in more than one way; we present only a single solution here.

The Exponential Distribution

Chapter 5. Random variables

Exact Confidence Intervals

ECE302 Spring 2006 HW4 Solutions February 6,

An Introduction to Basic Statistics and Probability

Probability Generating Functions

Practice Problems #4

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

Statistics 100A Homework 7 Solutions

STAT 830 Convergence in Distribution

ISyE 6761 Fall 2012 Homework #2 Solutions

Lecture 6: Discrete & Continuous Probability and Random Variables

Exponential Distribution

東 海 大 學 資 訊 工 程 研 究 所 碩 士 論 文

Chapter 4 Lecture Notes

Chapter 6: Point Estimation. Fall Probability & Statistics

I. Pointwise convergence

A LOGNORMAL MODEL FOR INSURANCE CLAIMS DATA

INSURANCE RISK THEORY (Problems)

STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE

Lecture 13: Martingales

Normal distribution. ) 2 /2σ. 2π σ

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

Lecture 3: Continuous distributions, expected value & mean, variance, the normal distribution

ECE302 Spring 2006 HW5 Solutions February 21,

Lecture 10: Depicting Sampling Distributions of a Sample Proportion

SOCIETY OF ACTUARIES/CASUALTY ACTUARIAL SOCIETY EXAM C CONSTRUCTION AND EVALUATION OF ACTUARIAL MODELS EXAM C SAMPLE QUESTIONS

Final Mathematics 5010, Section 1, Fall 2004 Instructor: D.A. Levin

5.1 Identifying the Target Parameter

WHERE DOES THE 10% CONDITION COME FROM?

Lecture 8. Confidence intervals and the central limit theorem

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Math 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 1 Solutions

Bayesian Statistics in One Hour. Patrick Lam

6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:30-9:30 PM. SOLUTIONS

Homework # 3 Solutions

STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables

3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved.

Feb 28 Homework Solutions Math 151, Winter Chapter 6 Problems (pages )

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

Aggregate Loss Models

Chapter 4 Statistical Inference in Quality Control and Improvement. Statistical Quality Control (D. C. Montgomery)

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Joint Exam 1/P Sample Exam 1

Estimation and Confidence Intervals

e.g. arrival of a customer to a service station or breakdown of a component in some system.

ACTS 4301 FORMULA SUMMARY Lesson 1: Probability Review. Name f(x) F (x) E[X] Var(X) Name f(x) E[X] Var(X) p x (1 p) m x mp mp(1 p)

Section 6.1 Joint Distribution Functions

Stats on the TI 83 and TI 84 Calculator

Math/Stats 342: Solutions to Homework

6 PROBABILITY GENERATING FUNCTIONS

Probability Calculator

Linear Classification. Volker Tresp Summer 2015

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

1.1 Introduction, and Review of Probability Theory Random Variable, Range, Types of Random Variables CDF, PDF, Quantiles...

6.2. Discrete Probability Distributions

Tenth Problem Assignment

Statistics. Aims of this course... Schedules... Recommended books... Keywords... Notation...

Notes on the Negative Binomial Distribution

Binomial random variables

University of California, Los Angeles Department of Statistics. Random variables

Exam C, Fall 2006 PRELIMINARY ANSWER KEY

Permutation & Non-Parametric Tests

Statistical Machine Learning from Data

Chapter 2. Hypothesis testing in one population

Probability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X

Tests of Hypotheses Using Statistics

2 Binomial, Poisson, Normal Distribution

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference

Econometrics Simple Linear Regression

Introduction to Probability

The Normal Distribution. Alan T. Arnholt Department of Mathematical Sciences Appalachian State University

Transcription:

Practice problems for Homework 11 - Point Estimation 1. (10 marks) Suppose we want to select a random sample of size 5 from the current CS 3341 students. Which of the following strategies is the best: a) Pick 5 students from the first row. b) Pick 5 of your friends from the class. c) Post a note on elearning and get the first 5 students who respond. d) Wait near the door and pick the first 5 students who enter the class. e) Assign a number to each student in the class and use a random number generator to pick 5 students.. (10 marks) Consider the following results of 10 tosses of a coin: H, T, T, T, T, H, T, H, T, T a) Estimate the probability of head (H) for this coin. b) Estimate the standard error of your estimate. 3. (10 marks) Suppose the following data shows the number of the problems from the Practice Problems Set attempted in the past week by 10 randomly selected students: a) Find the sample mean. b) Find the sample variance., 4, 0, 7, 1,, 0, 3,, 1. c) Estimate the mean number of practice problems attempted by a student in the past week. d) Estimate the standard error of the estimated mean. 4. (10 marks) Let X 1,..., X n denote a random sample from a Uniform (0, θ) distribution, with θ > 0 as the unknown parameter. Let X denote the sample mean. a) Is X unbiased for θ? Explain your answer. b) Find an unbiased estimator for θ. c) Find the variance of the estimator in the previous part. d) Suggest an alternative estimator for θ. 1

5. (10 marks) A sample of 3 observations, (X 1 = 0.4, X = 0.7, X 3 = 0.9) is collected from a continuous distribution with density Estimate θ a. by the method of moments; b. by the method of maximum likelihood. 6. (10 marks) (P. 648, #3, first part) f(x) = θx θ 1 for 0 < x < 1. The memory residence times of 13,171 jobs were measured, and the sample mean was found to be 0.05 s and the sample variance, 0.00674 s. Assuming that the memory residence time is gamma-distributed, estimate its parameters r and λ using the method of moments. 7. (10 marks) Derive method of moments and maximum-likelihood estimators for: a) parameter p based on a Bernoulli(p) sample of size n. b) parameter p based on a Binomial(N, p) sample of size n. Compute your estimators if the observed sample is (3, 6,, 0, 0, 3) and N = 10. c) parameter λ based on a Poisson(λ) sample of size n. d) parameters a and b based on a Uniform (a, b) sample of size n. e) parameter µ based on a Normal(µ, σ ) sample of size n with known variance σ and unknown mean µ. f) parameter σ based on a Normal(µ, σ ) sample of size n with known mean µ and unknown variance σ. g) parameters (µ, σ ) based on a Normal(µ, σ ) sample of size n with unknown mean µ and variance σ.

Solutions: 1.. 3. Strategy (e) is the best as this is the only strategy that ensures that each possible sample of 5 students is equally likely to be selected. The other strategies lead to convenience samples. Let X denote the toss of a single coin. Further, let X = 1 if a head results, and X = 0 if a tail results. This X is a Bernoulli (p) random variable, where p denotes the probability of head. Let ˆp denote the estimator of p. a) The estimated value of p is ˆp = (1 + 0 + 0 +... + 1 + 0 + 0)/10 = 0.3. b) The estimated standard error of ˆp is ˆp(1 ˆp)/n = 0.3(0.7)/10 = 0.14. a) X = n X i/n = ( + 4 +... + + 1)/10 =. b) S = n (X i X) /(n 1) = ( (.) + (4.) +... + (.) + (1.) ) /(10 1) = 4.4 c) The estimate is X =.. d) Estimated standard error of X is S/ n = 4.4/10 = 0.66 4. Let X denote a Uniform (0, θ) random variable. a) No. This is because expectation of E(X) is θ/, which implies that E(X) = E(X) = θ/ θ. b) The estimator X is unbiased since E(X) = θ, for all θ. c) var(x) = 4var(X) = 4var(X)/n = 4θ /(1n) = θ /(3n). d) Since θ denotes the maximum value in the population, a natural estimator is maximum value in the sample, i.e., maxx 1,..., X n }. 5. a. Method of moments. Compute 3

equate it to µ 1 = E(X) = 1 0 xf(x)dx = 1 0 θx θ dx = θxθ+1 θ + 1 x=1 x=0 = θ θ + 1, and solve for θ, m 1 = X = 0.4 + 0.7 + 0.9 3 = 3 θ θ + 1 = 3 = ˆθ = b. Method of maximum likelihood. The joint density is Take logarithm, f(x 1,..., X n ) = 3 θx θ 1 i 3 3 ln f(x 1,..., X n ) = ln θ + (θ 1) ln x i } = 3 ln θ + (θ 1) ln x i Take derivative, equate to 0, and solve for θ, ln f θ = 3 3 θ + ln x i = 0, 3 ˆθ = 3/ ln x i = 3/(ln 0.4 + ln 0.7 + ln 0.9) =.1766 6. For the Gamma distribution, E(X) = r/λ and V ar(x) = r/λ. So, the first two moments are µ 1 = r/λ and µ = (r/λ ) + (r/λ). 4

Also, the first two sample moments are M 1 = X and M = 1 n X i. Then, the method of moments estimators can be obtained by solving the system of equations r/λ = M 1 with respect to r and λ. µ = (r/λ ) + (r/λ) = M r/λ = M 1 (r/λ ) + (r/λ) = M r/λ = M 1 r/λ + M 1 = M r = M 1 /(M M 1 ) λ = M 1 /(M M 1 ) Notice that M M 1 = 1 n (Xi X) = (n 1)s /n = (13170)(0.00674)/(13171) = 0.00673. Then we can compute the estimates, ˆr = X /0.00673 = 0.37, ˆλ = X/0.00673 = 7.437. It is easier to use the \bf central} second moments, s and V ar(x) instead of E(X ) and M. This way, we solve the system of equations E(X) = r/λ = X V ar(x) = r/λ = s. Solving this system in terms of r and λ, we immediately get the method of moment estimates ˆr = X /s = 0.37, ˆλ = X/s = 7.436. 7. # 1a (Bernoulli) This is a special case of #1(b) when N = 1. 5

Following #1(b), both methods result in ˆp = X. # 1b (Binomial) Method of moments. To estimate parameter p of Binomial(N, p) distribution, we recall that µ 1 = E(X) = Np. There is only one unknown parameter p, hence we write one equation. Equate µ 1 = E(X) = Np to X and solve: ˆp = X/N. Maximum likelihood. Start with the joint p.m.f. f(x 1,..., X n ) = n ( N X i ) p X i (1 p) N X i, then ln f(x 1,..., X n ) = ( N X i ) + X i ln p + (N X i ) ln(1 p), p ln f(x 1,..., X n ) = n 1 X n i 1 (N X i) p 1 p (find roots of the derivative in order to maximize the density). Solving this equation, = n X p nn n X 1 p = 0, Answer: ˆp = X/N. (1 p) X = p(n X) p = X/N. # 1(c) (Poisson) Method of moments. To estimate parameter λ of Poisson(λ) distribution, we recall that µ 1 = E(X) = λ. There is only one unknown parameter, hence we write one equation, µ 1 = λ = m 1 = X. 6

Solving it for λ, we obtain ˆλ = X, the method of moments estimator of λ. Maximum likelihood. The p.m.f. of Poisson distribution is P (x) = e λ λx x!, and its logarithm is ln P (x) = λ + x ln λ ln(x!). Thus, we need to maximize ln P (X 1,..., X n ) = ( λ + X i ln λ) + C = nλ + ln λ X i, where C = ln(x!) is a constant that does not contain the unknown parameter λ. Find the critical point(s) of this log-likelihood. Differentiating it and equating its derivative to 0, we get This equation has only one solution λ ln P (X 1,..., X n ) = n + 1 λ ˆλ = 1 n X i = X. X i = 0. Since this is the only critical point, and since the likelihood vanishes (converges to 0) as λ 0 or λ, we conclude that ˆλ is the maximizer. Therefore, it is the maximum likelihood estimator of λ. For the Poisson distribution, the method of moments and the method of maximum likelihood returned the same estimator, ˆλ = X. # 1d (Uniform) 7

Method of moments. Use two moments, µ 1 = (a + b)/ and, for example, µ = V ar(x) = (b a) /1. Find â and ˆb by solving the system a+b = X (b a) = S â = X S 1 1, ˆb = X + S 1 Maximum likelihood. The joint density f(x 1,..., X n ) = ( 1 b a) n if a X 1,..., X n b 0 otherwise is monotonically increasing in a and decreasing in b. It is maximized at the largest value of a and the smallest value of b where this density is not 0. These are â = min(x i ) and ˆb = max(x i ) # 1e (Normal with unknown mean) Method of moments. Equate µ to X and trivially obtain ˆµ = X. Maximum likelihood. The joint density is Then f(x 1,..., X n ) = n 1 σ π exp (X i µ) σ }. ln f(x 1,..., X n ) = n ln(σ π) (X i µ) σ = n ln(σ π) nµ X i µ + Xi σ is a parabola in terms of µ, and it is maximized at ˆµ = X i /n = X. # 1f (Normal with unknown variance) Method of moments. The first population moment is not a function of σ, thus we equate the second (central, for simplicity) moment σ to the second moment S and obtain ˆσ = S. 8

Since µ is known, we can also use the estimator instead of S. ˆσ = 1 n (X i µ) Maximum likelihood. Differentiate ln f(x 1,..., X n ) from #1e with respect to σ, σ ln f(x 1,..., X n ) = n σ + (Xi µ) = 0, σ 3 (find roots of the derivative in order to maximize the density) so that #1 g (Normal with both parameters unknown) Method of moments. (Xi µ) ˆσ =. n When both µ and σ are unknown, we equate the first and second (central, for simplicity) moments and trivially obtain ˆµ = X and ˆσ = S so that ˆσ = S. Maximum likelihood. First, maximizing ln f(x 1,..., X n ) from # 1e in µ, we again obtain ˆµ = X regardless of the value of σ. Then, substitute the obtained maximizer into ln f(x 1,..., X n ) for the unknown value of µ and maximize the resulting function in terms of σ. We get the same answer as in (d) with the unknown parameter µ replaced by X, (Xi ˆσ = X) = S. n 9