Lecture 4: Estimation of ARIMA models

Size: px
Start display at page:

Download "Lecture 4: Estimation of ARIMA models"

Transcription

1 Lecture 4: Estimation of ARIMA models Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept Dec Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

2 Road map 1 Introduction Overview Framework 2 Sample moments 3 (Nonlinear) Least squares method Least squares estimation Nonlinear least squares estimation Discussion 4 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Generalized method of moments 5 Maximum likelihood estimation Overview Estimation Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

3 Introduction Overview 1. Introduction 1.1. Overview Provide an overview of some estimation methods for linear time series models : Sample moments estimation ; (Nonlinear) Linear least squares method ; Maximum likelihood estimation (Generalized) Method of moments Other methods are available (e.g., Bayesian estimation or Kalman filtering through state-space models)! Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

4 Introduction Overview Broadly speaking, these methods consist in estimating the parameters of interest (autoregressive coefficients, moving average coefficients, and variance of the innovation process) as follows : Optimize Q T (δ) δ s.t. (nonlinear) linear equality and inequality constraints where Q T is the objective function and δ is the vector of parameters of interest. Remark : (Nonlinear) Linear equality and inequality constraints may represent the fundamentalness conditions. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

5 Introduction Overview Examples (Nonlinear) Linear least squares methods aims at minimizing the sum of squared residuals ; Maximum likelihood estimation aims at maximizing the (log-) likelihood function ; Generalized method of moments aims at minimizing the distance between the theoretical moments and zero (using a weighting matrix). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

6 Introduction Overview These methods differ in terms of : Assumptions ; Nature of the model (e.g., MA versus AR) ; Finite and large sample properties ; Etc. High quality software programs (Eviews, SAS, Splus, Stata, etc) are available! However, it is important to know the estimation options (default procedure, optimization algorithm, choice of initial conditions) and to keep in mind that all these estimation techniques do not perform equally and do depend on the nature of the model... Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

7 1.2. Framework Introduction Framework Suppose that (X t ) is correctly specified as an ARIMA(p,d,q) model Φ(L) d X t = Θ(L)ɛ t where ɛ t is a weak white noise (0, σɛ 2 ) and Φ(L) = 1 φ 1 L φ p L p with φ p 0 Θ(L) = 1 + θ 1 L + + θ q L q with θ q 0. Assume that 1 The model order (p,d, and q) is known ; 2 The data has zero mean ; 3 The series is weakly stationary (e.g., after d-difference transformation, etc). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

8 Sample moments 2. Sample moments Let (X 1,, X T ) represent a sample of size T from a covariance-stationary process with E [X t ] = m X for all t Cov [X t, X t h ] = γ X (h) for all t Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

9 Sample moments If nothing is known about the distribution of the time series, (non-parametric) estimation can be provided by : Mean Autocovariance or ˆγ X (h) = 1 T ˆγ X (h) = Autocorrelation T h ˆm X X T = 1 T T X t (X t X T )(X t+ h X T ) T h 1 (X t X T )(X t+ h X T ). T h ˆρ X (h) = ˆγ X (h) ˆγ X (0). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

10 (Nonlinear) Least squares method Least squares estimation 3. (Nonlinear) Least squares method 3.1. Least squares estimation Consider the linear regression model (t = 1,, T ) : y t = x tb 0 + ɛ t where (y t, x t) are i.i.d. (H1), the regressors are stochastic (H2), x t is the t th -row of the X matrix, and the following assumptions hold : 1 Exogeneity (H3) 2 Spherical error terms (H4) 3 Rank condition (H5) E [ɛ t x t ] = 0 for all t; V [ɛ t x t ] = σ 2 0 for all t; Cov [ɛ t, ɛ t x t, x t ] = 0 for t t rank [E (x t x t)] = k. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

11 (Nonlinear) Least squares method Least squares estimation Definition The ordinary least squares estimation of b 0, b ols, defined from the following minimization program ˆb ols = argmin b 0 = argmin b 0 T ɛ 2 t T ( yt x tb ) 2 0 is given by ( T ) 1 ( T ) ˆb ols = x t x t x t y t. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

12 (Nonlinear) Least squares method Least squares estimation Proposition Under H1-H5, the ordinary least squares estimator of b is weakly consistent ˆb ols p T b 0 and is asymptotically normally distributed ) ( l T (ˆbols b 0 N 0 k 1, σ0e 2 [ x t x t ] ) 1. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

13 (Nonlinear) Least squares method Least squares estimation Example : AR(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( φ < 1) : X t = φx t 1 + ɛ t where (ɛ t ) is the innovation process of (X t ). The ordinary least squares estimation of φ is defined to be : Moreover ˆφ ols = ( T t=2 x 2 t 1 ) 1 ( T ) x t 1 x t t=2 ( ) l T ˆφ ols φ N ( 0, 1 φ 2). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

14 (Nonlinear) Least squares method Nonlinear least squares estimation 3.2. Nonlinear least squares estimation Definition Consider the nonlinear regression model defined by (t = 1,, T ) : y t = h(x t; b 0 ) + ɛ t where h is a nonlinear function (w.r.t. b 0 ). The nonlinear least squares estimator of b 0, ˆb nls, satisfies the following minimization program : ˆb nls = argmin b 0 = argmin b 0 T ɛ 2 t T ( yt h(x t; b 0 ) ) 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

15 (Nonlinear) Least squares method Nonlinear least squares estimation Example : MA(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( θ < 1) : X t = ɛ t + θɛ t 1 where (ɛ t ) is the innovation process of (X t ). The ordinary least squares estimation of θ, which is defined by ˆθ ols = argmin θ T (x t θɛ t 1 ) 2, t=2 is not feasible (since ɛ t 1 is not observable!). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

16 (Nonlinear) Least squares method Example : MA(1) estimation (cont d) Conditionally on ɛ 0, x 1 = ɛ 1 + θɛ 0 ɛ 1 = x 1 θɛ 0 Nonlinear least squares estimation x 2 = ɛ 2 + θɛ 1 ɛ 2 = x 2 θɛ 1 ɛ 2 = x 2 θ(x 1 θɛ 0 ). t 2 x t 1 = ɛ t 1 θɛ t 2 ɛ t 1 = ( θ) k x t 1 k + ( θ) t 1 ɛ 0. The (conditional) nonlinear least squares estimation of θ is defined by (assuming that ɛ 0 is negligible) ( ) T t 2 2 ˆθ nls = argmin x t θ ( θ) k x t 1 k θ t=2 k=0 and ) l T (ˆθ nls θ N ( 0, 1 θ 2). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50 k=0

17 (Nonlinear) Least squares method Nonlinear least squares estimation Example : Least squares estimation using backcasting procedure Consider the fundamental representation of a MA(q) : X t = u t Given some initial values δ (0) = u t = ɛ t + θ 1 ɛ t θ q ɛ t q ( ), θ (0) 1,, θ(0) q Compute the unconditional residuals û t û t = X t! Backcast values of ɛ t for t = (q 1),, T using the backward recursion ɛ t = û t θ (0) 1 ɛ t+1 θ q (0) ɛ t+q where the q values for ɛ t beyond the estimation sample are set to zero : ɛ T +1 = = ɛ T +q = 0, and û t = 0 for t < 1. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

18 (Nonlinear) Least squares method Nonlinear least squares estimation Example : Least squares estimation using backcasting procedure (cont d) Estimate the values of ɛ t using a forward recursion ˆɛ t = û t θ (0) 1 ɛ t 1 θ q (0) ɛ t q Minimize the sum of squared residuals using the fitted values ˆɛ t : and find ˆδ (1) = T (X t ˆɛ t θ 1ˆɛ t 1 θ qˆɛ t q ) 2 q+1 ( ). θ (1) 1,, θ(1) q Repeat the backcast step, forward recursion, and minimization procedures until the estimates of the moving average part converge. Remark : The backcast step can be turned off ( ɛ (q 1) = = ɛ 0 = 0) and the forward recursion is only used. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

19 3.3. Discussion (Nonlinear) Least squares method Discussion (Linear and Nonlinear) Least squares estimation are often used in practise. These methods lead to simple algorithms (e.g., backcasting procedure for ARMA models) and can often be used to initialize more sophisticated estimation methods. These methods are opposed to exact methods (see likelihood estimation). Nonlinear estimation requires numerical optimization algorithms : Initial conditions ; Number of iterations and stopping criteria ; Local and global convergence ; Fundamental representation and constraints. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

20 (Generalized) Method of moments Methods of moments and Yule-Walker estimation 4. (Generalized) Method of moments 4.1. Methods of moments and Yule-Walker estimation Definition Suppose there is a set of k conditions S T g (δ) = 0 k 1 where S T R k denotes a vector of theoretical moments, δ R k is a vector of parameters, and g : R k R k defines a (bijective) mapping between S T and δ. The method of moments estimation of δ, ˆδ mm, is defined to be the value of δ such that ) Ŝ T g (ˆδmm = 0 k 1 where Ŝ T is the estimation (empirical counterpart) of S T. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

21 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Example 1 : MA(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( θ < 1) : X t = ɛ t + θɛ t 1 where (ɛ t ) is the innovation process of (X t ). One has S T g(δ) = ( γx (0) (1 + θ 2 )σ 2 ɛ γ X (1) θσ 2 ɛ ) = where δ = (θ, σɛ 2 ) and S T = (γ X (0), γ X (1)) = (E [ Xt 2 ], E [Xt X t 1 ]). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

22 (Generalized) Method of moments Methods of moments and Yule-Walker estimation The method of moment estimation of δ, ˆδ mm, solves ( ) ˆγ Ŝ T g(ˆδ mm ) = X (0) (1 + ˆθ mm)ˆσ 2 ɛ,mm 2 ˆγ X (1) ˆθ mmˆσ ɛ,mm 2 = Therefore (using the constraint θ < 1) ˆφ mm = 1 1 4ˆρ 2 X (1) and 2ˆρ X (1) ˆσ 2 ɛ,mm = 2γ X (0) ˆρ 2X (1) Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

23 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Example 2 : Yule-Walker estimation Let (X t ) be a causal AR(p). The Yule-Walker equations write p E X t k X t φ j X t j = 0 for all k = 0,, p j=1 or S T g(δ) = 0 (p+1) 1 { γx (0) φ γ p = σ 2 ɛ γ p Γ p φ = 0 p 1 where γ p = (γ X (1),, γ X (p)) = (E [X t X t 1 ],, E [X t X t p ]), S T = (γ X (0), γ p), φ = (φ 1,, φ p ), δ = (φ, σ 2 ɛ ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

24 (Generalized) Method of moments Methods of moments and Yule-Walker estimation The method of moment estimation of (φ, σ 2 ɛ ) solves and ˆφ mm = 1 ˆΓ p ˆγ p ˆσ 2 ɛ,mm = ˆγ X (0) ˆφ mmˆγ p. Question : Difference(s) with ordinary least squares estimation of an AR(p)? Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

25 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Discussion It is necessary to use theoretical moments that can be well-estimated from the data (or given the sample size). Method of moment estimation depends on the mapping g 1 and especially the smoothness of the inverse g 1 (need to avoid that the inverse map has a large derivative ). Summarizing the data through the sample moments incurs a loss of information the main exception is when the sample moments are sufficient (in a statistical sense). Instead of reducing a sample of size T to a sample of empirical moments of size k, one may reduce the sample to more empirical moments than there are parameters...the so-called generalized method of moments. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

26 (Generalized) Method of moments 4.2. Generalized method of moments Generalized method of moments Intuition : Consider the linear regression model in an i.i.d. context y t = x tb 0 + ɛ t The exogeneity assumption E [ɛ t X ] implies the k moments conditions (for t = 1,, T ) E [ x t (y t x tb 0 ) ] = 0 k 1. Using the empirical counterpart (analogy principle) 1 T x t (y t x T tb 0 ) = 0 k 1 This is a system of k equations with k unknowns (just-identified) ( T ) 1 ( T ) ˆb ols = x t x t x t y t. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

27 (Generalized) Method of moments Generalized method of moments More generally, suppose that there exists a set of instruments z t R p (with p k) the instruments being (i) uncorrelated with the error terms and (ii) correlated with the explanatory variables such that E [ z t (y t x tb 0 ) ] = 0 p 1 E [h(z t ; δ 0 )] = 0 p 1. Using the empirical counterpart (analogy principle) 1 T T z t (y t x tb 0 ) = 0 p 1 1 T T h(z t ; δ 0 ) = 0 p 1. This system is overidentified for p > k. How can we find an estimation of b 0? Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

28 (Generalized) Method of moments Generalized method of moments The answer is provided by the generalized method of moments, i.e. an estimation method based on estimating equations that imposes the nullity of the expectation of a vector function of the observations and the parameters of interest b 0. The basic idea is to choose a value for δ such that the empirical counterpart is as close as possible to zero. As close as possible to zero means that one minimizes the T distance between 1 T z t (y t x tb 0 ) and 0 p 1 using a weighting matrix (say, W T ). Two issues 1 Choice of instruments ; 2 Choice of the (optimal) weighting matrix. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

29 (Generalized) Method of moments Generalized method of moments Definition Consider the set of moment conditions E [h(z t, δ 0 )] = 0 p 1 where δ 0 R k, p k, and h is a p-dimensional (nonlinear) linear function of the parameters of interest and depends on (i.i.d.) observations. The generalized method of moments estimator of δ 0 associated with W T (symmetric positive definite matrix) is a solution to the problem min δ [ 1 T ] [ ] T 1 T h(z t ; δ) W T h(z t ; δ). T Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

30 (Generalized) Method of moments Generalized method of moments Definition The generalized method of moments estimator of δ 0 solves the first-order conditions [ 1 T T h(z t ; ˆδ ] [ gmm ) 1 W T δ T ] T h(z t ; ˆδ gmm ) = 0 k 1. This estimator is consistent and asymptotically normally distributed. Remark : The first-order conditions correspond to k linear combinations of the moments conditions. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

31 (Generalized) Method of moments Generalized method of moments This estimator depends on W t. There exists a best GMM estimator, i.e. an optimal weighting matrix. The optimal weighting matrix is given by (i.i.d. context) W T = V 1 [h(z t, δ 0 )] E 1 [ h(z t, δ 0 )h(z t, δ 0 ) ] and the sample analog is [ ] 1 1 T h(z t, δ 0 )h(z t, δ 0 ). T It does depend on δ 0! How can we proceed in practise? Two-step GMM or iterated GMM method ; One-step method (Continuously updating estimator, exponential tilting estimator, empirical likelihood estimator, etc). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

32 (Generalized) Method of moments Generalized method of moments Example : AR(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( φ < 1) : X t = φx t 1 + ɛ t. Since (ɛ t ) is the innovation process of (X t ), one has the following moment conditions E [x t k ɛ t ] = 0 for k > 0 Past values x t 1,, x t h can be used as instruments Let z t denote the set of instruments z t = (x t 1,, x t h ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

33 (Generalized) Method of moments Generalized method of moments The moment conditions write where E [h(z t, δ 0 )] = 0 h 1 h(z t, δ 0 ) = z t (x t φ 1 x t 1 ). The empirical counterpart is defined to be x 1 T t 1 T. (x t φ 1 x t 1 ) = 0 h 1. t=h+1 x t h Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

34 (Generalized) Method of moments Generalized method of moments The 2S-GMM estimator of φ is a solution to the problem [ T t=h+1 h(z t ; ˆδ gmm ) δ ] ˆΩ 1 T [ T t=h+1 h(z t ; ˆδ gmm ) ] = 0 h 1. where ˆΩ 1 T is a first-step consistent estimate of the inverse of the variance-covariance matrix of the moments conditions and it accounts for the dependence of the moment conditions. Remark : In a first step, one solves [ T t=h+1 ] [ T ] h(z t ; ˆδ 1,gmm ) I T h(z t ; δ ˆδ 1,gmm ) = 0 h 1. t=h+1 and then compute ˆΩ 1 T (ˆδ 1,gmm ) ˆΩ 1 T. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

35 (Generalized) Method of moments Generalized method of moments Discussion Choice of instruments (optimal instruments?) and of the weighting matrix ; Bias-efficiency trade-off ; Numerical procedures can be cumbersome! Other methods can be written as a GMM estimation. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

36 Maximum likelihood estimation Overview 5. Maximum likelihood estimation 5.1. Overview Consider the linear regression model (t = 1,, T ) : y t = x tb 0 + ɛ t where (y t, x t) are i.i.d., the regressors are stochastic, x t is the t th -row of the X matrix, and It follows that ɛ t x t i.i.d.n (0, σ 2 ɛ ). y t x t i.i.d.n (x tb 0, σ 2 ɛ ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

37 Maximum likelihood estimation Overview One observes a sample of (conditionally) i.i.d. observations (y t, x t). These observations are realizations of random variables with a known distribution but unknown parameters. The basic idea is to maximize the probability of observing these realizations (given a known parametric distribution with unknown parameters). The probability of observing these realizations is provided by the joint density (or pdf) of the observations (or evaluated at the observations). This joint density is the likelihood function, with the exception that it is a function of the unknown parameters (and not the observations!). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

38 5.2. Estimation Maximum likelihood estimation Estimation Definition The exact likelihood function of the linear regression model is defined to be L(y, x; δ 0 ) = T L t (y t, x t ; δ 0 ) where L t is the (exact) likelihood function for observation t L t (y t, x t ; δ 0 ) = f Yt X t (y t x t ; δ 0 ) f Xt (x t ). where f Yt Xt (respectively, f Xt ) is the (conditional) probability density function of Y t X t (respectively, of X t ). Accordingly, the exact log-likelihood function is defined to be l(y, x; δ 0 ) = T l t (y t, x t ; δ 0 ) T logl t (y t, x t ; δ 0 ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

39 Maximum likelihood estimation Estimation The exact likelihood function is given by T T L(y, x; δ 0 ) = f Yt X t (y t x t ; δ 0 ) f Xt (x t ) The second right-hand side term is often neglected (especially if f X does not depend on δ 0 ) and one considers the conditional likelihood function T L(y x; δ 0 ) = f Yt X t (y t x t ; δ 0 ) T ( = σ 2 ɛ 2π ) 1/2 exp ( 1 ) 2σ 2 (y t x tb 0 ) 2 ɛ = ( σɛ2π 2 ) ( ) T /2 exp 1 T 2σɛ 2 (y t x tb 0 ) 2 and the conditional log-likelihood function is given by l(y x; δ 0 ) = T 2 log(σ2 ɛ) T 2 log(2π) 1 T 2σɛ 2 (y t x tb 0 ) 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

40 Maximum likelihood estimation Estimation Definition The maximum likelihood estimator of δ 0 in the linear regression model, which is the solution of the problem max δ L(y x; δ) maxl(y x; δ), is asymptotically unbiased and efficient, and normally distributed ) l T (ˆδml δ 0 N ( 0 k 1, I 1 ) F where I F is the Fisher information matrix. δ Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

41 Maximum likelihood estimation Example : AR(1) estimation Estimation Consider the following AR(1) representation ( φ < 1) X t = φx t 1 + ɛ t. where ɛ t or ɛ t X t 1 i.i.d.n ( 0, σɛ 2 ) Therefore Accordingly, X t X t 1 = x t 1 N ( φx t 1, σɛ 2 ) ) X 1 N ( 0, σ 2 ɛ 1 φ 2 f Xt X t 1 (x t x t 1 ; δ 0 ) = ( σ 2 ɛ 2π ) 1/2 exp ( 1 f X1 (x 1 ; δ 0 ) = 2σ 2 ɛ (x t φx t 1 ) 2 ) ( ) σ 2 1/2 ɛ 1 φ 2 2π exp ( 1 ) φ2 2σɛ 2 x1 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

42 Maximum likelihood estimation Exact (log-) likelihood function Estimation The exact likelihood function is given by T L(x; δ 0 ) = L 1 (x 1 ; δ 0 ) L τ (x τ x τ 1 ; δ 0 ) = f X1 (x 1 ; δ 0 ) = τ=2 T f Xτ X τ 1 (x τ x τ 1 ; δ 0 ) τ=2 ( ) σ 2 1/2 ɛ 1 φ 2 2π exp ( 1 φ2 ( σ 2 ɛ2π ) T 1 2 exp ( 1 2σ 2 ɛ 2σ 2 ɛ x 2 1 ) ) T (x t φx t 1 ) 2. The exact log-likelihood is l(x; δ 0 ) = T 2 log(2π) 1 ( ) σ 2 2 log ɛ 1 φ 2 1 φ2 T 1 log(σ 2 2 ɛ) 1 2σɛ 2 t=2 2σ 2 ɛ T (x t φx t 1 ) 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50 t=2 x 2 1

43 Maximum likelihood estimation Estimation Conditional (log-) likelihood function The conditional likelihood function is given by L(x x 1 ; δ 0 ) = = T L τ (x τ x τ 1 ; δ 0 ) τ=2 T f Xτ X τ 1 (x τ x τ 1 ; δ 0 ) τ=2 = ( σ 2 ɛ 2π ) T 1 2 exp The conditional log-likelihood is ( 1 2σ 2 ɛ ) T (x t φx t 1 ) 2. t=2 l(x x 1 ; δ 0 ) = T 1 log(2π) T 1 log(σɛ 2 ) T 2σɛ 2 (x t φx t 1 ) 2. t=2 Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

44 Maximum likelihood estimation Estimation The two methods lead to different maximum likelihood estimates ˆδ cml ˆδ eml The conditional maximum likelihood estimator of φ is also the ordinary least squares estimator of φ This also holds for any AR(p). ˆφ cml = ˆφ ols The Yule-Walker, conditional maximum likelihood and (ordinary) least squares estimator are asymptotically equivalent. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

45 Discussion Maximum likelihood estimation Estimation Generally the maximum likelihood estimates are built as if the observations are Gaussian, though it is not necessary that (X t ) is Gaussian when doing the estimation. If the model is misspecified, the only difference (to some extent...) is that estimates may be less efficient. In this case, one can use the pseudo- or quasi-maximum likelihood estimation. In the case of linear models, there might be not so much gain of using exact maximum likelihood against conditional likelihood method (or least squares method). Maximum likelihood estimation often requires nonlinear numerical optimization procedures. Estimation of the Fisher information matrix can be cumbersome (especially in finite samples). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

46 Appendix Appendix : ML estimation of a MA(q) One has Step 1 : Writing ɛ = (ɛ 1,, ɛ T ) ɛ 1 q = ɛ 1 q. =. ɛ 0 = ɛ 0 ɛ 1 = x 1 θ 1 ɛ 1 θ q ɛ 1 q ɛ 1 = x 2 θ 1 ɛ 2 θ q ɛ 2 q. =. ɛ T = x T θ 1 ɛ T θ q ɛ T q Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

47 Appendix Therefore ɛ = NX + Zɛ = ( Z N ) ( ɛ X ) with X = (x 1,, x T ), ɛ = (ɛ 1 q, ɛ 2 q,, ɛ 1, ɛ 0 ), and N = ( 0q T A 1 (θ) ) ( Iq, Z = A 2 (θ) where A 1 (θ) is a lower triangular matrix (T T ), A 2 (θ) is a T q matrix, and I q is the identity matrix of order q. ) Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

48 Appendix Step 2 : Determination of the joint density of ɛ The joint probability density function of ɛ is defined to be [ ( ) 2πσ 2 T +q 2 ɛ exp 1 ] 2σɛ 2 (NX + Zɛ ) (NX + Zɛ ). Step 3 : Determination of the marginal density of X The marginal density of X is defined to be ( ) 2πσ 2 T ( 2 ɛ detx X ) [ 1 2 exp 1 ] 2σɛ 2 (NX + Z ˆɛ ) (NX + Z ˆɛ ) with ˆɛ = (Z Z) 1 Z NX. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

49 Appendix Step 4 : Determination of the log-likelihood funcion The log-likelihood function is defined to be l (x; δ) = T 2 log(2π) T 2 log(σ2 ɛ ) T 2 log ( detx X ) S(θ) 2σ 2 ɛ where δ = (θ, σ 2 ɛ ) and S(θ) = (NX + Z ˆɛ ) (NX + Z ˆɛ ). Step 5 : Determination of ˆσ ɛ,ml 2 Using the first-order condition w.r.t. σɛ 2, one has ) S (ˆθ ml ˆσ ɛ,ml 2 =. T Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

50 Appendix Step 6 : Determination of the concentrated log-likelihood funcion The concentrated log-likelihood function is defined to be l (x; θ) = T log [S(θ)] log [ detx X ]. Step 7 : Determination of ˆθ ml ˆθ ml solves Remarks : min θ { T log [S(θ)] + log [ detx X ]}. 1 Least squares methods only focus on the first right-hand side term of the log-likelihood function. Exact methods account for both right-hand side terms. 2 This method generalizes to ARMA(p,q) models. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50

Lecture 2: ARMA(p,q) models (part 3)

Lecture 2: ARMA(p,q) models (part 3) Lecture 2: ARMA(p,q) models (part 3) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC) Univariate time series Sept.

More information

Time Series Analysis

Time Series Analysis Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Identification of univariate time series models, cont.:

More information

Time Series Analysis III

Time Series Analysis III Lecture 12: Time Series Analysis III MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis III 1 Outline Time Series Analysis III 1 Time Series Analysis III MIT 18.S096 Time Series Analysis

More information

Univariate and Multivariate Methods PEARSON. Addison Wesley

Univariate and Multivariate Methods PEARSON. Addison Wesley Time Series Analysis Univariate and Multivariate Methods SECOND EDITION William W. S. Wei Department of Statistics The Fox School of Business and Management Temple University PEARSON Addison Wesley Boston

More information

Time Series Analysis 1. Lecture 8: Time Series Analysis. Time Series Analysis MIT 18.S096. Dr. Kempthorne. Fall 2013 MIT 18.S096

Time Series Analysis 1. Lecture 8: Time Series Analysis. Time Series Analysis MIT 18.S096. Dr. Kempthorne. Fall 2013 MIT 18.S096 Lecture 8: Time Series Analysis MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis 1 Outline Time Series Analysis 1 Time Series Analysis MIT 18.S096 Time Series Analysis 2 A stochastic

More information

Analysis and Computation for Finance Time Series - An Introduction

Analysis and Computation for Finance Time Series - An Introduction ECMM703 Analysis and Computation for Finance Time Series - An Introduction Alejandra González Harrison 161 Email: mag208@exeter.ac.uk Time Series - An Introduction A time series is a sequence of observations

More information

Estimating an ARMA Process

Estimating an ARMA Process Statistics 910, #12 1 Overview Estimating an ARMA Process 1. Main ideas 2. Fitting autoregressions 3. Fitting with moving average components 4. Standard errors 5. Examples 6. Appendix: Simple estimators

More information

Chapter 3: The Multiple Linear Regression Model

Chapter 3: The Multiple Linear Regression Model Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Univariate Time Series Analysis; ARIMA Models

Univariate Time Series Analysis; ARIMA Models Econometrics 2 Spring 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing

More information

Univariate Time Series Analysis; ARIMA Models

Univariate Time Series Analysis; ARIMA Models Econometrics 2 Fall 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Univariate Time Series Analysis We consider a single time series, y,y 2,..., y T. We want to construct simple

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

Some useful concepts in univariate time series analysis

Some useful concepts in univariate time series analysis Some useful concepts in univariate time series analysis Autoregressive moving average models Autocorrelation functions Model Estimation Diagnostic measure Model selection Forecasting Assumptions: 1. Non-seasonal

More information

1 Short Introduction to Time Series

1 Short Introduction to Time Series ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The

More information

3.1 Stationary Processes and Mean Reversion

3.1 Stationary Processes and Mean Reversion 3. Univariate Time Series Models 3.1 Stationary Processes and Mean Reversion Definition 3.1: A time series y t, t = 1,..., T is called (covariance) stationary if (1) E[y t ] = µ, for all t Cov[y t, y t

More information

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its

More information

Introduction to General and Generalized Linear Models

Introduction to General and Generalized Linear Models Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby

More information

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem Chapter Vector autoregressions We begin by taking a look at the data of macroeconomics. A way to summarize the dynamics of macroeconomic data is to make use of vector autoregressions. VAR models have become

More information

Chapter 4: Vector Autoregressive Models

Chapter 4: Vector Autoregressive Models Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...

More information

Multivariate Normal Distribution

Multivariate Normal Distribution Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues

More information

ITSM-R Reference Manual

ITSM-R Reference Manual ITSM-R Reference Manual George Weigt June 5, 2015 1 Contents 1 Introduction 3 1.1 Time series analysis in a nutshell............................... 3 1.2 White Noise Variance.....................................

More information

Sales forecasting # 2

Sales forecasting # 2 Sales forecasting # 2 Arthur Charpentier arthur.charpentier@univ-rennes1.fr 1 Agenda Qualitative and quantitative methods, a very general introduction Series decomposition Short versus long term forecasting

More information

Financial TIme Series Analysis: Part II

Financial TIme Series Analysis: Part II Department of Mathematics and Statistics, University of Vaasa, Finland January 29 February 13, 2015 Feb 14, 2015 1 Univariate linear stochastic models: further topics Unobserved component model Signal

More information

Factor analysis. Angela Montanari

Factor analysis. Angela Montanari Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number

More information

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium

More information

Time Series Analysis

Time Series Analysis Time Series Analysis Autoregressive, MA and ARMA processes Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 212 Alonso and García-Martos

More information

Time Series Analysis in Economics. Klaus Neusser

Time Series Analysis in Economics. Klaus Neusser Time Series Analysis in Economics Klaus Neusser May 26, 2015 Contents I Univariate Time Series Analysis 3 1 Introduction 1 1.1 Some examples.......................... 2 1.2 Formal definitions.........................

More information

PS 271B: Quantitative Methods II. Lecture Notes

PS 271B: Quantitative Methods II. Lecture Notes PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.

More information

Examples. David Ruppert. April 25, 2009. Cornell University. Statistics for Financial Engineering: Some R. Examples. David Ruppert.

Examples. David Ruppert. April 25, 2009. Cornell University. Statistics for Financial Engineering: Some R. Examples. David Ruppert. Cornell University April 25, 2009 Outline 1 2 3 4 A little about myself BA and MA in mathematics PhD in statistics in 1977 taught in the statistics department at North Carolina for 10 years have been in

More information

TIME SERIES ANALYSIS

TIME SERIES ANALYSIS TIME SERIES ANALYSIS L.M. BHAR AND V.K.SHARMA Indian Agricultural Statistics Research Institute Library Avenue, New Delhi-0 02 lmb@iasri.res.in. Introduction Time series (TS) data refers to observations

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

STATISTICA Formula Guide: Logistic Regression. Table of Contents

STATISTICA Formula Guide: Logistic Regression. Table of Contents : Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

State Space Time Series Analysis

State Space Time Series Analysis State Space Time Series Analysis p. 1 State Space Time Series Analysis Siem Jan Koopman http://staff.feweb.vu.nl/koopman Department of Econometrics VU University Amsterdam Tinbergen Institute 2011 State

More information

SYSTEMS OF REGRESSION EQUATIONS

SYSTEMS OF REGRESSION EQUATIONS SYSTEMS OF REGRESSION EQUATIONS 1. MULTIPLE EQUATIONS y nt = x nt n + u nt, n = 1,...,N, t = 1,...,T, x nt is 1 k, and n is k 1. This is a version of the standard regression model where the observations

More information

Wooldridge, Introductory Econometrics, 3d ed. Chapter 12: Serial correlation and heteroskedasticity in time series regressions

Wooldridge, Introductory Econometrics, 3d ed. Chapter 12: Serial correlation and heteroskedasticity in time series regressions Wooldridge, Introductory Econometrics, 3d ed. Chapter 12: Serial correlation and heteroskedasticity in time series regressions What will happen if we violate the assumption that the errors are not serially

More information

Advanced Forecasting Techniques and Models: ARIMA

Advanced Forecasting Techniques and Models: ARIMA Advanced Forecasting Techniques and Models: ARIMA Short Examples Series using Risk Simulator For more information please visit: www.realoptionsvaluation.com or contact us at: admin@realoptionsvaluation.com

More information

Econometrics Simple Linear Regression

Econometrics Simple Linear Regression Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight

More information

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling Jeff Wooldridge NBER Summer Institute, 2007 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of Groups and

More information

TIME SERIES ANALYSIS

TIME SERIES ANALYSIS TIME SERIES ANALYSIS Ramasubramanian V. I.A.S.R.I., Library Avenue, New Delhi- 110 012 ram_stat@yahoo.co.in 1. Introduction A Time Series (TS) is a sequence of observations ordered in time. Mostly these

More information

Chapter 2. Dynamic panel data models

Chapter 2. Dynamic panel data models Chapter 2. Dynamic panel data models Master of Science in Economics - University of Geneva Christophe Hurlin, Université d Orléans Université d Orléans April 2010 Introduction De nition We now consider

More information

Java Modules for Time Series Analysis

Java Modules for Time Series Analysis Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series

More information

From the help desk: Bootstrapped standard errors

From the help desk: Bootstrapped standard errors The Stata Journal (2003) 3, Number 1, pp. 71 80 From the help desk: Bootstrapped standard errors Weihua Guan Stata Corporation Abstract. Bootstrapping is a nonparametric approach for evaluating the distribution

More information

Time Series Analysis

Time Series Analysis Time Series Analysis Andrea Beccarini Center for Quantitative Economics Winter 2013/2014 Andrea Beccarini (CQE) Time Series Analysis Winter 2013/2014 1 / 143 Introduction Objectives Time series are ubiquitous

More information

Statistics in Retail Finance. Chapter 6: Behavioural models

Statistics in Retail Finance. Chapter 6: Behavioural models Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural

More information

Analysis of algorithms of time series analysis for forecasting sales

Analysis of algorithms of time series analysis for forecasting sales SAINT-PETERSBURG STATE UNIVERSITY Mathematics & Mechanics Faculty Chair of Analytical Information Systems Garipov Emil Analysis of algorithms of time series analysis for forecasting sales Course Work Scientific

More information

Time Series Analysis and Forecasting

Time Series Analysis and Forecasting Time Series Analysis and Forecasting Math 667 Al Nosedal Department of Mathematics Indiana University of Pennsylvania Time Series Analysis and Forecasting p. 1/11 Introduction Many decision-making applications

More information

Time Series Analysis

Time Series Analysis JUNE 2012 Time Series Analysis CONTENT A time series is a chronological sequence of observations on a particular variable. Usually the observations are taken at regular intervals (days, months, years),

More information

Chapter 6: Multivariate Cointegration Analysis

Chapter 6: Multivariate Cointegration Analysis Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration

More information

On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information

On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information Finance 400 A. Penati - G. Pennacchi Notes on On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information by Sanford Grossman This model shows how the heterogeneous information

More information

2. Linear regression with multiple regressors

2. Linear regression with multiple regressors 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions

More information

Topic 5: Stochastic Growth and Real Business Cycles

Topic 5: Stochastic Growth and Real Business Cycles Topic 5: Stochastic Growth and Real Business Cycles Yulei Luo SEF of HKU October 1, 2015 Luo, Y. (SEF of HKU) Macro Theory October 1, 2015 1 / 45 Lag Operators The lag operator (L) is de ned as Similar

More information

Discrete Time Series Analysis with ARMA Models

Discrete Time Series Analysis with ARMA Models Discrete Time Series Analysis with ARMA Models Veronica Sitsofe Ahiati (veronica@aims.ac.za) African Institute for Mathematical Sciences (AIMS) Supervised by Tina Marquardt Munich University of Technology,

More information

Multiple Linear Regression in Data Mining

Multiple Linear Regression in Data Mining Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple

More information

Impulse Response Functions

Impulse Response Functions Impulse Response Functions Wouter J. Den Haan University of Amsterdam April 28, 2011 General definition IRFs The IRF gives the j th -period response when the system is shocked by a one-standard-deviation

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Regression III: Advanced Methods

Regression III: Advanced Methods Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models

More information

Maximum likelihood estimation of mean reverting processes

Maximum likelihood estimation of mean reverting processes Maximum likelihood estimation of mean reverting processes José Carlos García Franco Onward, Inc. jcpollo@onwardinc.com Abstract Mean reverting processes are frequently used models in real options. For

More information

Detekce změn v autoregresních posloupnostech

Detekce změn v autoregresních posloupnostech Nové Hrady 2012 Outline 1 Introduction 2 3 4 Change point problem (retrospective) The data Y 1,..., Y n follow a statistical model, which may change once or several times during the observation period

More information

Chapter 7. Univariate Volatility Modeling. 7.1 Why does volatility change?

Chapter 7. Univariate Volatility Modeling. 7.1 Why does volatility change? Chapter 7 Univariate Volatility Modeling Note: The primary references for these notes are chapters 1 and 11 in Taylor (5). Alternative, but less comprehensive, treatments can be found in chapter 1 of Hamilton

More information

Agenda. Managing Uncertainty in the Supply Chain. The Economic Order Quantity. Classic inventory theory

Agenda. Managing Uncertainty in the Supply Chain. The Economic Order Quantity. Classic inventory theory Agenda Managing Uncertainty in the Supply Chain TIØ485 Produkjons- og nettverksøkonomi Lecture 3 Classic Inventory models Economic Order Quantity (aka Economic Lot Size) The (s,s) Inventory Policy Managing

More information

Statistical machine learning, high dimension and big data

Statistical machine learning, high dimension and big data Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,

More information

CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York

CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not

More information

i=1 In practice, the natural logarithm of the likelihood function, called the log-likelihood function and denoted by

i=1 In practice, the natural logarithm of the likelihood function, called the log-likelihood function and denoted by Statistics 580 Maximum Likelihood Estimation Introduction Let y (y 1, y 2,..., y n be a vector of iid, random variables from one of a family of distributions on R n and indexed by a p-dimensional parameter

More information

Software Review: ITSM 2000 Professional Version 6.0.

Software Review: ITSM 2000 Professional Version 6.0. Lee, J. & Strazicich, M.C. (2002). Software Review: ITSM 2000 Professional Version 6.0. International Journal of Forecasting, 18(3): 455-459 (June 2002). Published by Elsevier (ISSN: 0169-2070). http://0-

More information

Time Series Analysis

Time Series Analysis Time Series Analysis Forecasting with ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos (UC3M-UPM)

More information

Rob J Hyndman. Forecasting using. 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1

Rob J Hyndman. Forecasting using. 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1 Rob J Hyndman Forecasting using 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1 Outline 1 Regression with ARIMA errors 2 Example: Japanese cars 3 Using Fourier terms for seasonality 4

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for

More information

11. Time series and dynamic linear models

11. Time series and dynamic linear models 11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd

More information

Traffic Safety Facts. Research Note. Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen

Traffic Safety Facts. Research Note. Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen Traffic Safety Facts Research Note March 2004 DOT HS 809 718 Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen Summary This research note uses

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct

More information

Forecasting model of electricity demand in the Nordic countries. Tone Pedersen

Forecasting model of electricity demand in the Nordic countries. Tone Pedersen Forecasting model of electricity demand in the Nordic countries Tone Pedersen 3/19/2014 Abstract A model implemented in order to describe the electricity demand on hourly basis for the Nordic countries.

More information

Time Series - ARIMA Models. Instructor: G. William Schwert

Time Series - ARIMA Models. Instructor: G. William Schwert APS 425 Fall 25 Time Series : ARIMA Models Instructor: G. William Schwert 585-275-247 schwert@schwert.ssb.rochester.edu Topics Typical time series plot Pattern recognition in auto and partial autocorrelations

More information

Parametric Statistical Modeling

Parametric Statistical Modeling Parametric Statistical Modeling ECE 275A Statistical Parameter Estimation Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 275A SPE Version 1.1 Fall 2012 1 / 12 Why

More information

Centre for Central Banking Studies

Centre for Central Banking Studies Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics

More information

Is the Basis of the Stock Index Futures Markets Nonlinear?

Is the Basis of the Stock Index Futures Markets Nonlinear? University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Engineering and Information Sciences 2011 Is the Basis of the Stock

More information

Lecture 3: Linear methods for classification

Lecture 3: Linear methods for classification Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,

More information

TIME SERIES. Syllabus... Keywords...

TIME SERIES. Syllabus... Keywords... TIME SERIES Contents Syllabus.................................. Books................................... Keywords................................. iii iii iv 1 Models for time series 1 1.1 Time series

More information

Monte Carlo-based statistical methods (MASM11/FMS091)

Monte Carlo-based statistical methods (MASM11/FMS091) Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February 5, 2013 J. Olsson Monte Carlo-based

More information

Spatial Statistics Chapter 3 Basics of areal data and areal data modeling

Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data

More information

Poisson Models for Count Data

Poisson Models for Count Data Chapter 4 Poisson Models for Count Data In this chapter we study log-linear models for count data under the assumption of a Poisson error structure. These models have many applications, not only to the

More information

Online Appendices to the Corporate Propensity to Save

Online Appendices to the Corporate Propensity to Save Online Appendices to the Corporate Propensity to Save Appendix A: Monte Carlo Experiments In order to allay skepticism of empirical results that have been produced by unusual estimators on fairly small

More information

Nonlinear Programming Methods.S2 Quadratic Programming

Nonlinear Programming Methods.S2 Quadratic Programming Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective

More information

ARMA, GARCH and Related Option Pricing Method

ARMA, GARCH and Related Option Pricing Method ARMA, GARCH and Related Option Pricing Method Author: Yiyang Yang Advisor: Pr. Xiaolin Li, Pr. Zari Rachev Department of Applied Mathematics and Statistics State University of New York at Stony Brook September

More information

Sections 2.11 and 5.8

Sections 2.11 and 5.8 Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and

More information

7 Gaussian Elimination and LU Factorization

7 Gaussian Elimination and LU Factorization 7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method

More information

Clustering in the Linear Model

Clustering in the Linear Model Short Guides to Microeconometrics Fall 2014 Kurt Schmidheiny Universität Basel Clustering in the Linear Model 2 1 Introduction Clustering in the Linear Model This handout extends the handout on The Multiple

More information

Master s Theory Exam Spring 2006

Master s Theory Exam Spring 2006 Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem

More information

One-year reserve risk including a tail factor : closed formula and bootstrap approaches

One-year reserve risk including a tail factor : closed formula and bootstrap approaches One-year reserve risk including a tail factor : closed formula and bootstrap approaches Alexandre Boumezoued R&D Consultant Milliman Paris alexandre.boumezoued@milliman.com Yoboua Angoua Non-Life Consultant

More information

Chapter 9: Univariate Time Series Analysis

Chapter 9: Univariate Time Series Analysis Chapter 9: Univariate Time Series Analysis In the last chapter we discussed models with only lags of explanatory variables. These can be misleading if: 1. The dependent variable Y t depends on lags of

More information

Analysis of Mean-Square Error and Transient Speed of the LMS Adaptive Algorithm

Analysis of Mean-Square Error and Transient Speed of the LMS Adaptive Algorithm IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 7, JULY 2002 1873 Analysis of Mean-Square Error Transient Speed of the LMS Adaptive Algorithm Onkar Dabeer, Student Member, IEEE, Elias Masry, Fellow,

More information

problem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved

problem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved 4 Data Issues 4.1 Truncated Regression population model y i = x i β + ε i, ε i N(0, σ 2 ) given a random sample, {y i, x i } N i=1, then OLS is consistent and efficient problem arises when only a non-random

More information

Forecasting the US Dollar / Euro Exchange rate Using ARMA Models

Forecasting the US Dollar / Euro Exchange rate Using ARMA Models Forecasting the US Dollar / Euro Exchange rate Using ARMA Models LIUWEI (9906360) - 1 - ABSTRACT...3 1. INTRODUCTION...4 2. DATA ANALYSIS...5 2.1 Stationary estimation...5 2.2 Dickey-Fuller Test...6 3.

More information

Trend and Seasonal Components

Trend and Seasonal Components Chapter 2 Trend and Seasonal Components If the plot of a TS reveals an increase of the seasonal and noise fluctuations with the level of the process then some transformation may be necessary before doing

More information

Chapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions

Chapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions Chapter 5 Analysis of Multiple Time Series Note: The primary references for these notes are chapters 5 and 6 in Enders (2004). An alternative, but more technical treatment can be found in chapters 10-11

More information

Assessing the Relative Power of Structural Break Tests Using a Framework Based on the Approximate Bahadur Slope

Assessing the Relative Power of Structural Break Tests Using a Framework Based on the Approximate Bahadur Slope Assessing the Relative Power of Structural Break Tests Using a Framework Based on the Approximate Bahadur Slope Dukpa Kim Boston University Pierre Perron Boston University December 4, 2006 THE TESTING

More information

LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as

LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as LOGISTIC REGRESSION Nitin R Patel Logistic regression extends the ideas of multiple linear regression to the situation where the dependent variable, y, is binary (for convenience we often code these values

More information

Statistics Graduate Courses

Statistics Graduate Courses Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.

More information

Time Series Analysis of Aviation Data

Time Series Analysis of Aviation Data Time Series Analysis of Aviation Data Dr. Richard Xie February, 2012 What is a Time Series A time series is a sequence of observations in chorological order, such as Daily closing price of stock MSFT in

More information