Econ 514: Probability and Statistics. Lecture 9: Point estimation. Point estimators

Size: px
Start display at page:

Download "Econ 514: Probability and Statistics. Lecture 9: Point estimation. Point estimators"

Transcription

1 Econ 514: Probability and Statistics Lecture 9: Point estimation Point estimators In Lecture 7 we discussed the setup of a study of the income distribution in LA. Regarding the population we considered the following possibilities No assumption on the population distribution (except finite population mean and variance): nonparametric approach. Population distribution has density f(x; θ), i.e. the density is known up to a vector of parameters: parametric approach. In this lecture we take the parametric approach. 1

2 We have a random sample X 1,..., X n from a distribution with density f(x; θ). Upon observation we have the numbers x 1,..., x n and we use these numbers to estimate θ. Definition: A (point) estimator of θ is a statistic ˆθ = t(x 1,..., X n ). Example: If the population distribution is N(µ, σ 2 ), then θ 1 = µ and θ 2 = σ 2, we estimate θ with the estimators ˆθ 1 = X n and ˆθ 2 = S 2 n. We could write ˆµ and ˆσ 2 for these estimators. This raises two questions: How do we find (good) estimators? How do we evaluate estimators, i.e. how do we decide that an estimator is a good estimator? 2

3 Finding estimators Method of moments Any statistic t(x 1,..., X n ) can be an estimator for θ. Natural estimators have a relation to θ. Example: In N(µ, σ 2 ) population we have µ = E(X) and σ 2 = var(x) so that the sample mean and variance are natural estimators. General procedure: Let θ be a K vector of parameters θ 1 θ =. θ K 3

4 Let for r = 1,..., K E(X r ) = µ r (θ 1,..., θ K ) (1) Consider the system of equations 1 n Xi r = µ r (θ 1,..., θ K ) (2) n that equates the sample and population r-th moments. If (3) has a unique solution for θ, then by the strong law of large numbers (4) has a unique solution if n. Definition: If (3) has a unique solution for θ, then a solution ˆθ of (4) is called the Method of moments (MM) estimator of θ. 4

5 Example: X 1,..., X n is a random sample from the exponential distribution with density f(x; λ) = λe λx 0 < x < = 0 otherwise We have E(X) = 1 λ and the MM estimator is the solution to 1 n X i = 1ˆλ n or ˆλ = 1 X n Also E(X 2 ) = 2 λ 2 so that another MM estimator is 2 ˆλ = 1 n n X2 i Conclusion: MM estimators are not unique. 5

6 Theorem: If (3) has a unique solution θ, the functions µ r are continuous for r = 1,..., K and E( X K ) <, then p ˆθ n θ Proof: By the (weak) law of large numbers for r = 1,..., K n p E(X r ) so that from (4) X r i µ r (ˆθ n ) p E(X r ) By the continuous mapping theorem this implies that ˆθ n has a limit that is equal to θ, because the solution of (3) is unique.. 6

7 Maximum likelihood Example Population density, e.g. coin tossing. f(x; p) = p x (1 p) 1 x x = 0, 1 = 0 otherwise We do 3 independent tosses which is a random sample X 1, X 2, X 3. For the observed values x 1, x 2, x 3 we have Pr(X 1 = x 1, X 2 = x 2, X 3 = x 3 ) = = 3 p x i (1 p) 1 x i = p 3 x i (1 p) 3 3 x i Note that X i is a sufficient statistic for p. We think that p = 1 2 or p =

8 Possible observations and their probability 3 x i p = 1 4 p = If we observe x 1 = x 2 = x 3 = 0 would you choose ˆp = 1 4 or ˆp = 1 2? An obvious method to select the estimate is to maximize the probability of the observed sample: ˆp = 1 4 = 1 2,, 3 x i = 0, 1 3 x i = 2, 3 This is a function of the observations only and hence an estimator of p. Note that ˆp is a function of the sufficient statistic. 8

9 General case In the example the estimator maximized the joint density of the random sample. If X 1,..., X n a random sample from a population with density f(x; θ), the joint density is n f(x 1,..., x n ; θ) = f(x i ; θ) = L(θ; x 1,..., x n ) If we consider this as a function of θ for fixed (observed) x 1,..., x n, this function is called the likelihood function. The maximum likelihood estimator (MLE) is ˆθ = arg max L(θ; x) θ with (note the abuse of notation) x 1 x =. x n 9

10 The solution is a function of x 1,..., x n Before observation ˆθ = t(x 1,..., x n ) ˆθ = t(x 1,..., X n ) Because the maximizing value is unaffected by monotone transformations of the maximand ˆθ = arg max ln L(θ; x) θ ˆθ = arg min θ L(θ; x) For a random sample the loglikelihood is n ln L(θ; x) = ln f(x i ; θ) Finding the first-order condition is easier with the loglikelihood and the set of equations (as many as parameters) ln L (θ; x) = 0 θ are called the likelihood equations. The derivative of the loglikelihood ln L θ (θ; x) is called the score function. 10

11 Example Same population as in previous example, but random sample of size n. Likelihood function n L(p; x) = f(x i ; p) = n p x i (1 p) 1 x i = p n x i (1 p) n n x i Loglikelihood function n ln L(p; x) = x i ln p + (n n x i ) ln(1 p) Score function ln L p (p; x) = 1 p n 1 n 1 p (n x i ) MLE ˆp = 1 n n x i = x n 11

12 Invariance of MLE Let L(θ; x) be a likelihood function and define τ = h(θ). Define A τ = {θ h(θ) = τ} and define If ˆθ is the MLE of θ, then with equality if ˆτ = h(ˆθ). L(τ; x) = sup θ A τ L(θ; x) sup L(τ; x) L(ˆθ; x) τ Conclusion: The MLE of τ is ˆτ = h(ˆθ). This is called the invariance property of the MLE. 12

13 Evaluation of estimators Sampling distribution Population parameter θ is unknown, but often there are restrictions on θ, i.e. θ Θ with Θ R K. Θ is called the parameter space. Example: In N(µ, σ 2 ) Θ = {(µ, σ 2 ) < µ <, σ 2 > 0} Because and estimator is a statistic ˆθ = t(x 1,..., X n ) it has a sampling distribution derived from the joint distribution of X 1,..., X n n f(x 1,..., x n ; θ) = f(x i ; θ) 13

14 Ideally this sampling distribution should be concentrated in θ irrespective what population value of θ is. Example: Random sample of size n from N(θ, 1) with Θ = {θ < θ < }. For estimator ˆθ = X n we have for the sampling distribution ˆθ N(θ, 1 n ) For sampling distribution: E(ˆθ) = θ and Var(ˆθ) 0 if n. 14

15 Compare sampling distributions in graph. Ranking of estimators ˆθ 1 dominates ˆθ 2. ˆθ 1 dominates ˆθ 3. What about ˆθ 1 and ˆθ 4? 15

16 Performance measure: MSE Ranking is possible if the performance of an estimator is captured by a single number. That is always somewhat arbitrary. Estimation error is ˆθ θ and squared estimation error (ˆθ θ) 2 treats positive and negative errors in the same way, penalizes large errors more. The mean squared error (MSE) is the average squared error in the sampling distribution of ˆθ ] MSE(ˆθ, θ) = E [(ˆθ θ) 2 The MSE depends in general on the population parameter θ. 16

17 Unbiased estimators Definition: An estimator ˆθ is unbiased if for all θ Θ E(ˆθ) = θ Consider MSE [ ( ) ] 2 MSE(ˆθ, θ) = E (ˆθ E(ˆθ)) + (E(ˆθ) θ) = = E [ ] [ ] (ˆθ E(ˆθ)) 2 + +(E(ˆθ) θ) 2 +2E (ˆθ E(ˆθ))(E(ˆθ) θ) = = Var(ˆθ) + (E(ˆθ) θ) 2 E(ˆθ) θ is the bias of the estimator, so that we have MSE(ˆθ, θ) = Var(ˆθ) + bias 2 and if the estimator is unbiased then MSE(ˆθ, θ) = Var(ˆθ) 17

18 If we only consider unbiased estimators ranking by MSE is ranking by variance. Example: For a random sample from N(θ, 1) and ˆθ = X n E(ˆθ) = θ, i.e. the estimator is unbiased. MSE(ˆθ, θ) = Var(X n ) = 1 n independent of θ. 18

19 Best estimators Definition: Let X 1,..., X n be a random sample from a population with density f(x; θ). The estimator ˆθ is a uniformly minimum variance unbiased (UMVU)estimator if for all θ Θ (i) E(ˆθ) = θ (ii) For all unbiased estimators θ, Var( θ) Var(ˆθ). Instead of looking for UMVU estimators directly we use a different approach (i) Find a lower bound on the variance of all unbiased estimators ˆθ (ii) Verify that for an estimator the variance is equal to the lower bound. That estimator is UMVU. 19

20 Theorem (Cramér-Rao): Let X 1,..., X n be a random sample from a population with density f(x; θ) with theta a scalar parameter. Then for all unbiased estimators ˆθ of θ 1 Var(ˆθ) [ ( ) ] 2 ne ln f θ (X; θ) Proof: We consider the case of a density w.r.t. Lebesgue measure. For all θ Θ n... f(x i ; θ)dx 1... dx n 1... t(x 1,..., x n ) n f(x i ; θ)dx 1... dx n θ We want to interchange differentation w.r.t. θ and integration. From Lecture 2 a sufficient condition is that f(x; θ) θ M(x) with M(x) integrable (check this!). 20

21 If this (or other sufficient condition) holds ( n )... f(x i ; θ) dx 1... dx n 0 (3) θ and ( n )... t(x 1,..., x n ) f(x i ; θ) dx 1... dx n 1 θ (4) Hence ( n ) 1... t(x 1,..., x n ) f(x i ; θ) dx 1... dx n = θ =... (t(x 1,..., x n ) θ) θ ( ln ) n n f(x i ; θ) f(x i ; θ)dx 1... dx n = 21

22 [ ( )] = E (t(x 1,..., X n ) θ) n ln f(x i ; θ) (5) θ E [(t(x 1,..., X n ) θ) 2] ( ( )) 2 E n ln f(x i ; θ) θ so that Var(ˆθ) 1 [ ( E θ ( n ln f(x i; θ)) ) ] (6) 2 In (5) we used the Cauchy-Schwartz inequality (see Lecture 6). 22

23 Because X 1,..., X n are independently and identically, abbreviated i.i.d. distributed ( ( n )) 2 E ln f(x i ; θ) = θ ( n = E because from [ ] ln f E θ (X 1; θ) = ) 2 [ ( ) ] 2 ln f θ (X i; θ) ln f = ne θ (X 1; θ) f θ (x; θ)dx = θ for the expectation of the cross-products [ ln f E θ (X i; θ) ln f ] θ (X j; θ) = 0. f(x; θ)dx = 0 23

24 The lower bound on the variance of an unbiased estimator is called the Cramér-Rao lower bound. If X 1,..., X n is not a random sample, but has joint density f(x 1,..., x n ; θ), then the lower bound is 1 [ ( E θ ln f(x 1,..., X n ; θ) ) ] 2 The inequality in (5) is an equality if and only if θ ln L(θ; x 1,..., x n ) = c(θ)(t(x 1,..., x n ) θ) Hence the estimator has a variance that reaches the lower bound if this equation holds. 24

25 Example: For an N(θ, 1) population f(x; θ) = 1 2π e 1 2 (x θ)2 Hence and E ln f(x; θ) = 1 2 ln(2π) 1 (x θ)2 2 ln f θ (x; θ) = x θ [ ( ) ] 2 ln f θ (X 1; θ) = E[(X 1 θ) 2 ] = 1 Conclusion: Cramér-Rao lower bound is 1 n. The estimator ˆθ = X n is unbiased and has a variance equal to this bound. ˆθ = X n is UMVU. 25

26 Information matrix We have ln f... θ (x 1,..., x n ; θ)dx 1... dx n 0 Differentiating with respect to θ and interchanging differentiation and integration (suggest a sufficient condition that allows this) 2 ln f 0... (x θ 2 1,..., x n ; θ)f(x 1,..., x n ; θ)dx 1... dx n + ln f +... θ (x 1,..., x n ; θ) f θ (x 1,..., x n ; θ)dx 1... dx n = [ 2 ] [ ( ) ] 2 ln f ln f = E (X θ 2 1,..., X n ; θ) +E θ (X 1,..., X n ; θ) 26

27 Conclusion: [ ( ) ] 2 ln f E θ (X 1,..., X n ; θ) = E [ 2 ] ln f (X θ 2 1,..., X n ; θ) The left-hand side is called Fisher information matrix. The equation shows that this is the variance of the score function and that it is equal to minus the expected value of the Hessian of the min-loglikelihood. 27

28 Example 1: Random sample of size n from N(µ, σ 2 ). The sample variance S 2 n = 1 n 1 is an unbiased estimator of σ 2. It can be shown i = 1 n (X i X n ) 2 Var(S 2 n) = 2σ4 n 1 The second derivative of the log density w.r.t. σ 2 is 2 ( ) 1 (σ 2 ) ln 1 2 σ2 2π e 2σ 2 (x µ 0) 2 = 1 2σ (x µ 0) 2 4 σ 6 By the information matrix equality the information matrix is [ E 1 2σ + (X µ 0) 2 ] = 1 4 σ 6 2σ 4 and the Cramér-Rao lower bound is 2σ 4 Conclusion: The MSE of S 2 n is strictly greater than the lower bound. n 28

29 The MLE of σ 2 is ˆσ 2 = 1 i = 1 n (X i X n ) 2 = n 1 n n S2 n and Var(ˆσ 2 ) = (n 1)2 n 2 Var(S 2 n) = E(ˆσ 2 ) = n 1 n σ2 < σ 2 2(n 1) n 2 σ 4 Conclusion: The MLE is biased. However MSE(S 2 n; σ 2 ) = 2σ4 n 1 > 2n 1 σ 4 = MSE(ˆσ 2 ; σ 2 ) n 2 Consider the case that µ is known. We have ln L = n σ 2 2σ + n 1 n (x 2 2σ 4 i µ) 2 = n = n 2σ 4 ( 1 n ) n (x i µ) 2 σ 2 = c(σ 2 )( σ 2 σ 2 ) Conclusion: If µ is known the estimator σ 2 = 1 n (x i µ) 2 n is unbiased and reaches the Cramér-Rao lower bound. 29

30 Example 2 Random sample of size n from a uniform distribution with density Because f(x; θ) = 1 θ I [0,θ](x) 2 ln f θ 2 (x; θ) = 1 θ 2 the Cramér-Rao lower bound seems to be θ2 n. Estimator θ = max{x 1,..., X n }. The statistic T = max{x 1,..., X n } has density f T (t) = nyn 1, 0 < y < θ θ n = 0, otherwise We find E( θ) = E(T ) = n n + 1 θ so that ˆθ = n+1 θ n is an unbiased estimator with Var(ˆθ) = 1 n(n + 2) θ2 < θ2 n 30

31 It seems that we have an unbiased estimator with a variance that is smaller than the lower bound. Problem is θ 0 f(x; θ)dx 1 Taking the derivative w.r.t. θ f(θ; θ) + θ 0 f (x; θ)dx θ θ 0 f (x; θ)dx θ i.e. we cannot interchange differentiation and integration. 31

32 Uniqueness of UMVU estimators and Rao-Blackwell theorem We show that UMVU estimators are unique. Let ˆθ be UMVU and let θ be another unbiased estimator of θ. Define a third estimator θ = ˆθ+t(ˆθ θ) with t some real number. We have Var(θ ) = Var(ˆθ) + t 2 Var(ˆθ θ) + 2tCov(ˆθ θ, ˆθ) This variance is minimal if t = Cov(ˆθ θ, ˆθ) Var(ˆθ θ) Because ˆθ is UMVU, t has to be 0 (otherwise we would have an estimator with a variance smaller than the lower bound). 32

33 Hence so that 0 = Cov(ˆθ θ, ˆθ) = Var(ˆθ) Cov(ˆθ, θ) Var( θ ˆθ) = Var(ˆθ)+Var( θ) 2Cov(ˆθ, θ) = Var( θ) Var(ˆθ) (7) Conclusions: If θ is also UMVU, then Var( θ ˆθ) = 0 and hence ˆθ = θ with probability 1. (7) simplifies the calculation of the variance of the difference of an unbiased and and UMVU estimator. 33

34 The next theorem shows how to improve an unbiased estimator. Rao-Blackwell theorem: If θ is an unbiased estimator of θ and T is a sufficient statistic for θ, then is unbiased and E(ˆθ) = E( θ T ) Var(ˆθ) Var( θ) Proof: For all θ Θ by the law of iterated expectations E(ˆθ) = E(E( θ T )) = E( θ) = θ and by the law of iterated variances Var( θ) = Var(E( θ T ))+E(Var( θ T )) = Var(ˆθ)+E(Var( θ T )) Var(ˆθ) Finally, because T is sufficient E( θ T ) does not depend on θ. 34

35 Asymptotic properties of estimators In most cases the sampling distribution of estimators is too complicated to compute mean, variance, and MSE. In that case we use asymptotic analysis, i.e. we let the sample size n. Consider random sample X 1,..., X n and let ˆθ n = t(x 1,..., X n ) be an estimator. What happens to the sampling distribution of ˆθ n if the sample size becomes large? Because if n becomes larger we know more and more of the population, the sampling distribution should behave as in the figure 35

36 36

37 if n = we know the populations and the sampling distribution should be degenerate in θ. Let ˆθ n be the sequence of estimators for increasing sample size. We say that θ n is (weakly) consistent if ˆθ n p θ If the convergence is a.s. we say that the estimator (sequence) is strongly consistent. Why does it not make sense to consider convergence in distribution? Example: For N(θ, σ 2 ) population the sample means is strongly and weakly consistent for θ. 37

38 Large sample behavior of MLE If for all θ Θ ( ) ln f E (X; θ) θ < then Also 1 n E n ( ) ln f θ (X i; θ) p ln f E θ (X 1; θ) ( ) ln f θ (X 1; θ) f = θ so that for θ = θ ( ) ln f E θ (X 1; θ) = f(x; θ) (x; θ) dx f(x; θ) f θ (x; θ)dx = θ f(x; θ)dx = 0 where we have interchanged differentiation and integration, which is allowed e.g. if for θ is a small interval around θ, f(x; θ) < g(x) with g integrable. 38

39 For the MLE ˆθ n 0 = 1 n n This suggests that ln f θ (X i; ˆθ n ) = E ( ) ln f θ (X 1; θ) ˆθ n p θ i.e. the MLE is weakly consistent. 39

40 Asymptotic distribution of MLE By Taylor s theorem 0 = 1 n ln f n θ (X i; ˆθ n ) = 1 n + 1 n n n 2 ln f θ 2 (X i ; θ n )(ˆθ n θ) with θ θ n ˆθ n or ˆθ n θ n θ. Consider the first term on rhs and define Y i = ln f θ (X i; θ) ln f θ (X i; θ)+ then E(Y i ) = 0 and [ ( ) ] 2 ln f Var(Y i ) = E θ (X i; θ) = I(θ) By the Central Limit Theorem 1 n ln f n θ (X i; θ) d N(0, I(θ)) 40

41 Next consider the second term on rhs. Because ˆθ n p θ also θ n θ. Further 1 n 2 ( ln f (X n θ 2 i ; θ) p 2 ) ln f E (X θ 2 1 ; θ) if ( 2 ) ln f E (X θ 2 1 ; θ) < This suggests that 1 n 2 ( ln f (X n θ 2 i ; θ n ) p 2 ) ln f E (X θ 2 1 ; θ) = A(θ) By the Slutsky theorem 0 = 1 n ln f n θ (X i; θ)+ 1 n n 2 ln f θ 2 implies that n(ˆθn θ) d N(A(θ) 1 I(θ)A(θ) 1 ) This is the limit distribution of the MLE. p (X i ; θ n ) n(ˆθ n θ) 41

42 By the information matrix equality A(θ) = I(θ) so that this simplifies to n(ˆθn θ) d N(I(θ) 1 ) Note the we consider the sequence n(ˆθ n θ) which is similar to nx n in the CLT. The variance is equal to the Cramér-Rao lower bound. This means that the limit distribution has a variance equal to the lower bound. We say that the MLE is asymptotically efficient. The asymptotic normal distribution can be used to compute a confidence interval for θ of the form ˆθ n c I(ˆθ n ) n < θ < ˆθ n + c I(ˆθ n ) n 42

43 Bootstrap To obtain the sampling distribution of a statistic for finite n one could use the computer. Consider the sample mean X n = 1 n n X i X n is the mean over the empirical distribution, P n, that assigns probability 1 n to X 1,..., X n X n = xdp n The observed sample x 1,..., x n is a realization of X 1,..., X n. 43

44 Now draw from x 1,..., x n a random sample of size n with replacement x 1,..., x n. This is a draw from the empirical distribution that gives x n = 1 n x i n Do this N times and consider the x n -s as draws from the sampling distribution of X n. The justification of this procedure is that if n then the empirical distribution converges to the population distribution, so that averaging over the empirical distribution is the same in large samples as averaging over the population distribution. This method of approximating the sampling distribution is called the bootstrap method after a tall tale by the (in)famous baron Von Münchhausen ( ). 44

Chapter 6: Point Estimation. Fall 2011. - Probability & Statistics

Chapter 6: Point Estimation. Fall 2011. - Probability & Statistics STAT355 Chapter 6: Point Estimation Fall 2011 Chapter Fall 2011 6: Point1 Estimat / 18 Chap 6 - Point Estimation 1 6.1 Some general Concepts of Point Estimation Point Estimate Unbiasedness Principle of

More information

Basics of Statistical Machine Learning

Basics of Statistical Machine Learning CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar

More information

i=1 In practice, the natural logarithm of the likelihood function, called the log-likelihood function and denoted by

i=1 In practice, the natural logarithm of the likelihood function, called the log-likelihood function and denoted by Statistics 580 Maximum Likelihood Estimation Introduction Let y (y 1, y 2,..., y n be a vector of iid, random variables from one of a family of distributions on R n and indexed by a p-dimensional parameter

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

Practice problems for Homework 11 - Point Estimation

Practice problems for Homework 11 - Point Estimation Practice problems for Homework 11 - Point Estimation 1. (10 marks) Suppose we want to select a random sample of size 5 from the current CS 3341 students. Which of the following strategies is the best:

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for

More information

MATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...

MATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators... MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................

More information

THE CENTRAL LIMIT THEOREM TORONTO

THE CENTRAL LIMIT THEOREM TORONTO THE CENTRAL LIMIT THEOREM DANIEL RÜDT UNIVERSITY OF TORONTO MARCH, 2010 Contents 1 Introduction 1 2 Mathematical Background 3 3 The Central Limit Theorem 4 4 Examples 4 4.1 Roulette......................................

More information

Master s Theory Exam Spring 2006

Master s Theory Exam Spring 2006 Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem

More information

Differentiating under an integral sign

Differentiating under an integral sign CALIFORNIA INSTITUTE OF TECHNOLOGY Ma 2b KC Border Introduction to Probability and Statistics February 213 Differentiating under an integral sign In the derivation of Maximum Likelihood Estimators, or

More information

2DI36 Statistics. 2DI36 Part II (Chapter 7 of MR)

2DI36 Statistics. 2DI36 Part II (Chapter 7 of MR) 2DI36 Statistics 2DI36 Part II (Chapter 7 of MR) What Have we Done so Far? Last time we introduced the concept of a dataset and seen how we can represent it in various ways But, how did this dataset came

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete

More information

STAT 830 Convergence in Distribution

STAT 830 Convergence in Distribution STAT 830 Convergence in Distribution Richard Lockhart Simon Fraser University STAT 830 Fall 2011 Richard Lockhart (Simon Fraser University) STAT 830 Convergence in Distribution STAT 830 Fall 2011 1 / 31

More information

Multivariate Normal Distribution

Multivariate Normal Distribution Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

Metric Spaces. Chapter 7. 7.1. Metrics

Metric Spaces. Chapter 7. 7.1. Metrics Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some

More information

MATHEMATICAL METHODS OF STATISTICS

MATHEMATICAL METHODS OF STATISTICS MATHEMATICAL METHODS OF STATISTICS By HARALD CRAMER TROFESSOK IN THE UNIVERSITY OF STOCKHOLM Princeton PRINCETON UNIVERSITY PRESS 1946 TABLE OF CONTENTS. First Part. MATHEMATICAL INTRODUCTION. CHAPTERS

More information

Sections 2.11 and 5.8

Sections 2.11 and 5.8 Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and

More information

Advanced statistical inference. Suhasini Subba Rao Email: suhasini.subbarao@stat.tamu.edu

Advanced statistical inference. Suhasini Subba Rao Email: suhasini.subbarao@stat.tamu.edu Advanced statistical inference Suhasini Subba Rao Email: suhasini.subbarao@stat.tamu.edu August 1, 2012 2 Chapter 1 Basic Inference 1.1 A review of results in statistical inference In this section, we

More information

Lecture Notes 1. Brief Review of Basic Probability

Lecture Notes 1. Brief Review of Basic Probability Probability Review Lecture Notes Brief Review of Basic Probability I assume you know basic probability. Chapters -3 are a review. I will assume you have read and understood Chapters -3. Here is a very

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

University of Ljubljana Doctoral Programme in Statistics Methodology of Statistical Research Written examination February 14 th, 2014.

University of Ljubljana Doctoral Programme in Statistics Methodology of Statistical Research Written examination February 14 th, 2014. University of Ljubljana Doctoral Programme in Statistics ethodology of Statistical Research Written examination February 14 th, 2014 Name and surname: ID number: Instructions Read carefully the wording

More information

PS 271B: Quantitative Methods II. Lecture Notes

PS 271B: Quantitative Methods II. Lecture Notes PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.

More information

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference 0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures

More information

171:290 Model Selection Lecture II: The Akaike Information Criterion

171:290 Model Selection Lecture II: The Akaike Information Criterion 171:290 Model Selection Lecture II: The Akaike Information Criterion Department of Biostatistics Department of Statistics and Actuarial Science August 28, 2012 Introduction AIC, the Akaike Information

More information

Principle of Data Reduction

Principle of Data Reduction Chapter 6 Principle of Data Reduction 6.1 Introduction An experimenter uses the information in a sample X 1,..., X n to make inferences about an unknown parameter θ. If the sample size n is large, then

More information

Big Data - Lecture 1 Optimization reminders

Big Data - Lecture 1 Optimization reminders Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics

More information

Statistical Theory. Prof. Gesine Reinert

Statistical Theory. Prof. Gesine Reinert Statistical Theory Prof. Gesine Reinert November 23, 2009 Aim: To review and extend the main ideas in Statistical Inference, both from a frequentist viewpoint and from a Bayesian viewpoint. This course

More information

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015 ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)! Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following

More information

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page Errata for ASM Exam C/4 Study Manual (Sixteenth Edition) Sorted by Page 1 Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page Practice exam 1:9, 1:22, 1:29, 9:5, and 10:8

More information

Lecture 3: Linear methods for classification

Lecture 3: Linear methods for classification Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,

More information

Fixed Point Theorems

Fixed Point Theorems Fixed Point Theorems Definition: Let X be a set and let T : X X be a function that maps X into itself. (Such a function is often called an operator, a transformation, or a transform on X, and the notation

More information

Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 )

Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 ) Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 ) and Neural Networks( 類 神 經 網 路 ) 許 湘 伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 35 13 Examples

More information

Detekce změn v autoregresních posloupnostech

Detekce změn v autoregresních posloupnostech Nové Hrady 2012 Outline 1 Introduction 2 3 4 Change point problem (retrospective) The data Y 1,..., Y n follow a statistical model, which may change once or several times during the observation period

More information

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010 Simulation Methods Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Simulation Methods 15.450, Fall 2010 1 / 35 Outline 1 Generating Random Numbers 2 Variance Reduction 3 Quasi-Monte

More information

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?

More information

Random graphs with a given degree sequence

Random graphs with a given degree sequence Sourav Chatterjee (NYU) Persi Diaconis (Stanford) Allan Sly (Microsoft) Let G be an undirected simple graph on n vertices. Let d 1,..., d n be the degrees of the vertices of G arranged in descending order.

More information

1. Let A, B and C are three events such that P(A) = 0.45, P(B) = 0.30, P(C) = 0.35,

1. Let A, B and C are three events such that P(A) = 0.45, P(B) = 0.30, P(C) = 0.35, 1. Let A, B and C are three events such that PA =.4, PB =.3, PC =.3, P A B =.6, P A C =.6, P B C =., P A B C =.7. a Compute P A B, P A C, P B C. b Compute P A B C. c Compute the probability that exactly

More information

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability CS 7 Discrete Mathematics and Probability Theory Fall 29 Satish Rao, David Tse Note 8 A Brief Introduction to Continuous Probability Up to now we have focused exclusively on discrete probability spaces

More information

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,

More information

Estimating an ARMA Process

Estimating an ARMA Process Statistics 910, #12 1 Overview Estimating an ARMA Process 1. Main ideas 2. Fitting autoregressions 3. Fitting with moving average components 4. Standard errors 5. Examples 6. Appendix: Simple estimators

More information

Lecture 13: Martingales

Lecture 13: Martingales Lecture 13: Martingales 1. Definition of a Martingale 1.1 Filtrations 1.2 Definition of a martingale and its basic properties 1.3 Sums of independent random variables and related models 1.4 Products of

More information

Performance measures for Neyman-Pearson classification

Performance measures for Neyman-Pearson classification Performance measures for Neyman-Pearson classification Clayton Scott Abstract In the Neyman-Pearson (NP) classification paradigm, the goal is to learn a classifier from labeled training data such that

More information

Lecture 8. Confidence intervals and the central limit theorem

Lecture 8. Confidence intervals and the central limit theorem Lecture 8. Confidence intervals and the central limit theorem Mathematical Statistics and Discrete Mathematics November 25th, 2015 1 / 15 Central limit theorem Let X 1, X 2,... X n be a random sample of

More information

Exact Nonparametric Tests for Comparing Means - A Personal Summary

Exact Nonparametric Tests for Comparing Means - A Personal Summary Exact Nonparametric Tests for Comparing Means - A Personal Summary Karl H. Schlag European University Institute 1 December 14, 2006 1 Economics Department, European University Institute. Via della Piazzuola

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

To give it a definition, an implicit function of x and y is simply any relationship that takes the form:

To give it a definition, an implicit function of x and y is simply any relationship that takes the form: 2 Implicit function theorems and applications 21 Implicit functions The implicit function theorem is one of the most useful single tools you ll meet this year After a while, it will be second nature to

More information

Analysis of Reliability and Warranty Claims in Products with Age and Usage Scales

Analysis of Reliability and Warranty Claims in Products with Age and Usage Scales Analysis of Reliability and Warranty Claims in Products with Age and Usage Scales J.F. Lawless M.J. Crowder K.-A. Lee University of Waterloo Imperial College London University of Waterloo Abstract Failures

More information

Stat 5102 Notes: Nonparametric Tests and. confidence interval

Stat 5102 Notes: Nonparametric Tests and. confidence interval Stat 510 Notes: Nonparametric Tests and Confidence Intervals Charles J. Geyer April 13, 003 This handout gives a brief introduction to nonparametrics, which is what you do when you don t believe the assumptions

More information

Non Parametric Inference

Non Parametric Inference Maura Department of Economics and Finance Università Tor Vergata Outline 1 2 3 Inverse distribution function Theorem: Let U be a uniform random variable on (0, 1). Let X be a continuous random variable

More information

Sums of Independent Random Variables

Sums of Independent Random Variables Chapter 7 Sums of Independent Random Variables 7.1 Sums of Discrete Random Variables In this chapter we turn to the important question of determining the distribution of a sum of independent random variables

More information

MSc. Econ: MATHEMATICAL STATISTICS, 1995 MAXIMUM-LIKELIHOOD ESTIMATION

MSc. Econ: MATHEMATICAL STATISTICS, 1995 MAXIMUM-LIKELIHOOD ESTIMATION MAXIMUM-LIKELIHOOD ESTIMATION The General Theory of M-L Estimation In orer to erive an M-L estimator, we are boun to make an assumption about the functional form of the istribution which generates the

More information

CHAPTER 2 Estimating Probabilities

CHAPTER 2 Estimating Probabilities CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a

More information

An extension of the factoring likelihood approach for non-monotone missing data

An extension of the factoring likelihood approach for non-monotone missing data An extension of the factoring likelihood approach for non-monotone missing data Jae Kwang Kim Dong Wan Shin January 14, 2010 ABSTRACT We address the problem of parameter estimation in multivariate distributions

More information

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}. Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian

More information

Lecture 21. The Multivariate Normal Distribution

Lecture 21. The Multivariate Normal Distribution Lecture. The Multivariate Normal Distribution. Definitions and Comments The joint moment-generating function of X,...,X n [also called the moment-generating function of the random vector (X,...,X n )]

More information

Adaptive Online Gradient Descent

Adaptive Online Gradient Descent Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650

More information

Numerical methods for American options

Numerical methods for American options Lecture 9 Numerical methods for American options Lecture Notes by Andrzej Palczewski Computational Finance p. 1 American options The holder of an American option has the right to exercise it at any moment

More information

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have

More information

I. Pointwise convergence

I. Pointwise convergence MATH 40 - NOTES Sequences of functions Pointwise and Uniform Convergence Fall 2005 Previously, we have studied sequences of real numbers. Now we discuss the topic of sequences of real valued functions.

More information

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1]. Probability Theory Probability Spaces and Events Consider a random experiment with several possible outcomes. For example, we might roll a pair of dice, flip a coin three times, or choose a random real

More information

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12 This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Tail inequalities for order statistics of log-concave vectors and applications

Tail inequalities for order statistics of log-concave vectors and applications Tail inequalities for order statistics of log-concave vectors and applications Rafał Latała Based in part on a joint work with R.Adamczak, A.E.Litvak, A.Pajor and N.Tomczak-Jaegermann Banff, May 2011 Basic

More information

BANACH AND HILBERT SPACE REVIEW

BANACH AND HILBERT SPACE REVIEW BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but

More information

Lecture 6: Discrete & Continuous Probability and Random Variables

Lecture 6: Discrete & Continuous Probability and Random Variables Lecture 6: Discrete & Continuous Probability and Random Variables D. Alex Hughes Math Camp September 17, 2015 D. Alex Hughes (Math Camp) Lecture 6: Discrete & Continuous Probability and Random September

More information

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

More information

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max

More information

Maximum likelihood estimation of mean reverting processes

Maximum likelihood estimation of mean reverting processes Maximum likelihood estimation of mean reverting processes José Carlos García Franco Onward, Inc. jcpollo@onwardinc.com Abstract Mean reverting processes are frequently used models in real options. For

More information

Testing predictive performance of binary choice models 1

Testing predictive performance of binary choice models 1 Testing predictive performance of binary choice models 1 Bas Donkers Econometric Institute and Department of Marketing Erasmus University Rotterdam Bertrand Melenberg Department of Econometrics Tilburg

More information

**BEGINNING OF EXAMINATION** The annual number of claims for an insured has probability function: , 0 < q < 1.

**BEGINNING OF EXAMINATION** The annual number of claims for an insured has probability function: , 0 < q < 1. **BEGINNING OF EXAMINATION** 1. You are given: (i) The annual number of claims for an insured has probability function: 3 p x q q x x ( ) = ( 1 ) 3 x, x = 0,1,, 3 (ii) The prior density is π ( q) = q,

More information

5 Numerical Differentiation

5 Numerical Differentiation D. Levy 5 Numerical Differentiation 5. Basic Concepts This chapter deals with numerical approximations of derivatives. The first questions that comes up to mind is: why do we need to approximate derivatives

More information

Interpreting Kullback-Leibler Divergence with the Neyman-Pearson Lemma

Interpreting Kullback-Leibler Divergence with the Neyman-Pearson Lemma Interpreting Kullback-Leibler Divergence with the Neyman-Pearson Lemma Shinto Eguchi a, and John Copas b a Institute of Statistical Mathematics and Graduate University of Advanced Studies, Minami-azabu

More information

Chapter 3: The Multiple Linear Regression Model

Chapter 3: The Multiple Linear Regression Model Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

. (3.3) n Note that supremum (3.2) must occur at one of the observed values x i or to the left of x i.

. (3.3) n Note that supremum (3.2) must occur at one of the observed values x i or to the left of x i. Chapter 3 Kolmogorov-Smirnov Tests There are many situations where experimenters need to know what is the distribution of the population of their interest. For example, if they want to use a parametric

More information

4. Continuous Random Variables, the Pareto and Normal Distributions

4. Continuous Random Variables, the Pareto and Normal Distributions 4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random

More information

x a x 2 (1 + x 2 ) n.

x a x 2 (1 + x 2 ) n. Limits and continuity Suppose that we have a function f : R R. Let a R. We say that f(x) tends to the limit l as x tends to a; lim f(x) = l ; x a if, given any real number ɛ > 0, there exists a real number

More information

Notes from Week 1: Algorithms for sequential prediction

Notes from Week 1: Algorithms for sequential prediction CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking

More information

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014 Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about

More information

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding

More information

1 Maximum likelihood estimation

1 Maximum likelihood estimation COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N

More information

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur Module No. #01 Lecture No. #15 Special Distributions-VI Today, I am going to introduce

More information

Reject Inference in Credit Scoring. Jie-Men Mok

Reject Inference in Credit Scoring. Jie-Men Mok Reject Inference in Credit Scoring Jie-Men Mok BMI paper January 2009 ii Preface In the Master programme of Business Mathematics and Informatics (BMI), it is required to perform research on a business

More information

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March 2015. Due:-March 25, 2015.

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March 2015. Due:-March 25, 2015. Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment -3, Probability and Statistics, March 05. Due:-March 5, 05.. Show that the function 0 for x < x+ F (x) = 4 for x < for x

More information

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL STATIsTICs 4 IV. RANDOm VECTORs 1. JOINTLY DIsTRIBUTED RANDOm VARIABLEs If are two rom variables defined on the same sample space we define the joint

More information

Game Theory: Supermodular Games 1

Game Theory: Supermodular Games 1 Game Theory: Supermodular Games 1 Christoph Schottmüller 1 License: CC Attribution ShareAlike 4.0 1 / 22 Outline 1 Introduction 2 Model 3 Revision questions and exercises 2 / 22 Motivation I several solution

More information

GLM, insurance pricing & big data: paying attention to convergence issues.

GLM, insurance pricing & big data: paying attention to convergence issues. GLM, insurance pricing & big data: paying attention to convergence issues. Michaël NOACK - michael.noack@addactis.com Senior consultant & Manager of ADDACTIS Pricing Copyright 2014 ADDACTIS Worldwide.

More information

NOV - 30211/II. 1. Let f(z) = sin z, z C. Then f(z) : 3. Let the sequence {a n } be given. (A) is bounded in the complex plane

NOV - 30211/II. 1. Let f(z) = sin z, z C. Then f(z) : 3. Let the sequence {a n } be given. (A) is bounded in the complex plane Mathematical Sciences Paper II Time Allowed : 75 Minutes] [Maximum Marks : 100 Note : This Paper contains Fifty (50) multiple choice questions. Each question carries Two () marks. Attempt All questions.

More information

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics Undergraduate Notes in Mathematics Arkansas Tech University Department of Mathematics An Introductory Single Variable Real Analysis: A Learning Approach through Problem Solving Marcel B. Finan c All Rights

More information

Managing uncertainty in call centers using Poisson mixtures

Managing uncertainty in call centers using Poisson mixtures Managing uncertainty in call centers using Poisson mixtures Geurt Jongbloed and Ger Koole Vrije Universiteit, Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam, The Netherlands

More information

A General Approach to Variance Estimation under Imputation for Missing Survey Data

A General Approach to Variance Estimation under Imputation for Missing Survey Data A General Approach to Variance Estimation under Imputation for Missing Survey Data J.N.K. Rao Carleton University Ottawa, Canada 1 2 1 Joint work with J.K. Kim at Iowa State University. 2 Workshop on Survey

More information

Schatten van de extreme value index voor afgeronde data (Engelse titel: Estimation of the extreme value index for imprecise data)

Schatten van de extreme value index voor afgeronde data (Engelse titel: Estimation of the extreme value index for imprecise data) Technische Universiteit Delft Faculteit Elektrotechniek, Wiskunde en Informatica Delft Institute of Applied Mathematics Schatten van de extreme value index voor afgeronde data (Engelse titel: Estimation

More information

5. Continuous Random Variables

5. Continuous Random Variables 5. Continuous Random Variables Continuous random variables can take any value in an interval. They are used to model physical characteristics such as time, length, position, etc. Examples (i) Let X be

More information

Introduction to Probability

Introduction to Probability Introduction to Probability EE 179, Lecture 15, Handout #24 Probability theory gives a mathematical characterization for experiments with random outcomes. coin toss life of lightbulb binary data sequence

More information

Notes on metric spaces

Notes on metric spaces Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.

More information

Linda Staub & Alexandros Gekenidis

Linda Staub & Alexandros Gekenidis Seminar in Statistics: Survival Analysis Chapter 2 Kaplan-Meier Survival Curves and the Log- Rank Test Linda Staub & Alexandros Gekenidis March 7th, 2011 1 Review Outcome variable of interest: time until

More information

Inner Product Spaces

Inner Product Spaces Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

More information

LS.6 Solution Matrices

LS.6 Solution Matrices LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions

More information