Lecture 4: Estimation of ARIMA models
|
|
- Ernest Parrish
- 7 years ago
- Views:
Transcription
1 Lecture 4: Estimation of ARIMA models Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept Dec Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
2 Road map 1 Introduction Overview Framework 2 Sample moments 3 (Nonlinear) Least squares method Least squares estimation Nonlinear least squares estimation Discussion 4 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Generalized method of moments 5 Maximum likelihood estimation Overview Estimation Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
3 Introduction Overview 1. Introduction 1.1. Overview Provide an overview of some estimation methods for linear time series models : Sample moments estimation ; (Nonlinear) Linear least squares method ; Maximum likelihood estimation (Generalized) Method of moments Other methods are available (e.g., Bayesian estimation or Kalman filtering through state-space models)! Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
4 Introduction Overview Broadly speaking, these methods consist in estimating the parameters of interest (autoregressive coefficients, moving average coefficients, and variance of the innovation process) as follows : Optimize Q T (δ) δ s.t. (nonlinear) linear equality and inequality constraints where Q T is the objective function and δ is the vector of parameters of interest. Remark : (Nonlinear) Linear equality and inequality constraints may represent the fundamentalness conditions. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
5 Introduction Overview Examples (Nonlinear) Linear least squares methods aims at minimizing the sum of squared residuals ; Maximum likelihood estimation aims at maximizing the (log-) likelihood function ; Generalized method of moments aims at minimizing the distance between the theoretical moments and zero (using a weighting matrix). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
6 Introduction Overview These methods differ in terms of : Assumptions ; Nature of the model (e.g., MA versus AR) ; Finite and large sample properties ; Etc. High quality software programs (Eviews, SAS, Splus, Stata, etc) are available! However, it is important to know the estimation options (default procedure, optimization algorithm, choice of initial conditions) and to keep in mind that all these estimation techniques do not perform equally and do depend on the nature of the model... Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
7 1.2. Framework Introduction Framework Suppose that (X t ) is correctly specified as an ARIMA(p,d,q) model Φ(L) d X t = Θ(L)ɛ t where ɛ t is a weak white noise (0, σɛ 2 ) and Φ(L) = 1 φ 1 L φ p L p with φ p 0 Θ(L) = 1 + θ 1 L + + θ q L q with θ q 0. Assume that 1 The model order (p,d, and q) is known ; 2 The data has zero mean ; 3 The series is weakly stationary (e.g., after d-difference transformation, etc). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
8 Sample moments 2. Sample moments Let (X 1,, X T ) represent a sample of size T from a covariance-stationary process with E [X t ] = m X for all t Cov [X t, X t h ] = γ X (h) for all t Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
9 Sample moments If nothing is known about the distribution of the time series, (non-parametric) estimation can be provided by : Mean Autocovariance or ˆγ X (h) = 1 T ˆγ X (h) = Autocorrelation T h ˆm X X T = 1 T T X t (X t X T )(X t+ h X T ) T h 1 (X t X T )(X t+ h X T ). T h ˆρ X (h) = ˆγ X (h) ˆγ X (0). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
10 (Nonlinear) Least squares method Least squares estimation 3. (Nonlinear) Least squares method 3.1. Least squares estimation Consider the linear regression model (t = 1,, T ) : y t = x tb 0 + ɛ t where (y t, x t) are i.i.d. (H1), the regressors are stochastic (H2), x t is the t th -row of the X matrix, and the following assumptions hold : 1 Exogeneity (H3) 2 Spherical error terms (H4) 3 Rank condition (H5) E [ɛ t x t ] = 0 for all t; V [ɛ t x t ] = σ 2 0 for all t; Cov [ɛ t, ɛ t x t, x t ] = 0 for t t rank [E (x t x t)] = k. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
11 (Nonlinear) Least squares method Least squares estimation Definition The ordinary least squares estimation of b 0, b ols, defined from the following minimization program ˆb ols = argmin b 0 = argmin b 0 T ɛ 2 t T ( yt x tb ) 2 0 is given by ( T ) 1 ( T ) ˆb ols = x t x t x t y t. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
12 (Nonlinear) Least squares method Least squares estimation Proposition Under H1-H5, the ordinary least squares estimator of b is weakly consistent ˆb ols p T b 0 and is asymptotically normally distributed ) ( l T (ˆbols b 0 N 0 k 1, σ0e 2 [ x t x t ] ) 1. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
13 (Nonlinear) Least squares method Least squares estimation Example : AR(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( φ < 1) : X t = φx t 1 + ɛ t where (ɛ t ) is the innovation process of (X t ). The ordinary least squares estimation of φ is defined to be : Moreover ˆφ ols = ( T t=2 x 2 t 1 ) 1 ( T ) x t 1 x t t=2 ( ) l T ˆφ ols φ N ( 0, 1 φ 2). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
14 (Nonlinear) Least squares method Nonlinear least squares estimation 3.2. Nonlinear least squares estimation Definition Consider the nonlinear regression model defined by (t = 1,, T ) : y t = h(x t; b 0 ) + ɛ t where h is a nonlinear function (w.r.t. b 0 ). The nonlinear least squares estimator of b 0, ˆb nls, satisfies the following minimization program : ˆb nls = argmin b 0 = argmin b 0 T ɛ 2 t T ( yt h(x t; b 0 ) ) 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
15 (Nonlinear) Least squares method Nonlinear least squares estimation Example : MA(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( θ < 1) : X t = ɛ t + θɛ t 1 where (ɛ t ) is the innovation process of (X t ). The ordinary least squares estimation of θ, which is defined by ˆθ ols = argmin θ T (x t θɛ t 1 ) 2, t=2 is not feasible (since ɛ t 1 is not observable!). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
16 (Nonlinear) Least squares method Example : MA(1) estimation (cont d) Conditionally on ɛ 0, x 1 = ɛ 1 + θɛ 0 ɛ 1 = x 1 θɛ 0 Nonlinear least squares estimation x 2 = ɛ 2 + θɛ 1 ɛ 2 = x 2 θɛ 1 ɛ 2 = x 2 θ(x 1 θɛ 0 ). t 2 x t 1 = ɛ t 1 θɛ t 2 ɛ t 1 = ( θ) k x t 1 k + ( θ) t 1 ɛ 0. The (conditional) nonlinear least squares estimation of θ is defined by (assuming that ɛ 0 is negligible) ( ) T t 2 2 ˆθ nls = argmin x t θ ( θ) k x t 1 k θ t=2 k=0 and ) l T (ˆθ nls θ N ( 0, 1 θ 2). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50 k=0
17 (Nonlinear) Least squares method Nonlinear least squares estimation Example : Least squares estimation using backcasting procedure Consider the fundamental representation of a MA(q) : X t = u t Given some initial values δ (0) = u t = ɛ t + θ 1 ɛ t θ q ɛ t q ( ), θ (0) 1,, θ(0) q Compute the unconditional residuals û t û t = X t! Backcast values of ɛ t for t = (q 1),, T using the backward recursion ɛ t = û t θ (0) 1 ɛ t+1 θ q (0) ɛ t+q where the q values for ɛ t beyond the estimation sample are set to zero : ɛ T +1 = = ɛ T +q = 0, and û t = 0 for t < 1. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
18 (Nonlinear) Least squares method Nonlinear least squares estimation Example : Least squares estimation using backcasting procedure (cont d) Estimate the values of ɛ t using a forward recursion ˆɛ t = û t θ (0) 1 ɛ t 1 θ q (0) ɛ t q Minimize the sum of squared residuals using the fitted values ˆɛ t : and find ˆδ (1) = T (X t ˆɛ t θ 1ˆɛ t 1 θ qˆɛ t q ) 2 q+1 ( ). θ (1) 1,, θ(1) q Repeat the backcast step, forward recursion, and minimization procedures until the estimates of the moving average part converge. Remark : The backcast step can be turned off ( ɛ (q 1) = = ɛ 0 = 0) and the forward recursion is only used. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
19 3.3. Discussion (Nonlinear) Least squares method Discussion (Linear and Nonlinear) Least squares estimation are often used in practise. These methods lead to simple algorithms (e.g., backcasting procedure for ARMA models) and can often be used to initialize more sophisticated estimation methods. These methods are opposed to exact methods (see likelihood estimation). Nonlinear estimation requires numerical optimization algorithms : Initial conditions ; Number of iterations and stopping criteria ; Local and global convergence ; Fundamental representation and constraints. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
20 (Generalized) Method of moments Methods of moments and Yule-Walker estimation 4. (Generalized) Method of moments 4.1. Methods of moments and Yule-Walker estimation Definition Suppose there is a set of k conditions S T g (δ) = 0 k 1 where S T R k denotes a vector of theoretical moments, δ R k is a vector of parameters, and g : R k R k defines a (bijective) mapping between S T and δ. The method of moments estimation of δ, ˆδ mm, is defined to be the value of δ such that ) Ŝ T g (ˆδmm = 0 k 1 where Ŝ T is the estimation (empirical counterpart) of S T. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
21 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Example 1 : MA(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( θ < 1) : X t = ɛ t + θɛ t 1 where (ɛ t ) is the innovation process of (X t ). One has S T g(δ) = ( γx (0) (1 + θ 2 )σ 2 ɛ γ X (1) θσ 2 ɛ ) = where δ = (θ, σɛ 2 ) and S T = (γ X (0), γ X (1)) = (E [ Xt 2 ], E [Xt X t 1 ]). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
22 (Generalized) Method of moments Methods of moments and Yule-Walker estimation The method of moment estimation of δ, ˆδ mm, solves ( ) ˆγ Ŝ T g(ˆδ mm ) = X (0) (1 + ˆθ mm)ˆσ 2 ɛ,mm 2 ˆγ X (1) ˆθ mmˆσ ɛ,mm 2 = Therefore (using the constraint θ < 1) ˆφ mm = 1 1 4ˆρ 2 X (1) and 2ˆρ X (1) ˆσ 2 ɛ,mm = 2γ X (0) ˆρ 2X (1) Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
23 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Example 2 : Yule-Walker estimation Let (X t ) be a causal AR(p). The Yule-Walker equations write p E X t k X t φ j X t j = 0 for all k = 0,, p j=1 or S T g(δ) = 0 (p+1) 1 { γx (0) φ γ p = σ 2 ɛ γ p Γ p φ = 0 p 1 where γ p = (γ X (1),, γ X (p)) = (E [X t X t 1 ],, E [X t X t p ]), S T = (γ X (0), γ p), φ = (φ 1,, φ p ), δ = (φ, σ 2 ɛ ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
24 (Generalized) Method of moments Methods of moments and Yule-Walker estimation The method of moment estimation of (φ, σ 2 ɛ ) solves and ˆφ mm = 1 ˆΓ p ˆγ p ˆσ 2 ɛ,mm = ˆγ X (0) ˆφ mmˆγ p. Question : Difference(s) with ordinary least squares estimation of an AR(p)? Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
25 (Generalized) Method of moments Methods of moments and Yule-Walker estimation Discussion It is necessary to use theoretical moments that can be well-estimated from the data (or given the sample size). Method of moment estimation depends on the mapping g 1 and especially the smoothness of the inverse g 1 (need to avoid that the inverse map has a large derivative ). Summarizing the data through the sample moments incurs a loss of information the main exception is when the sample moments are sufficient (in a statistical sense). Instead of reducing a sample of size T to a sample of empirical moments of size k, one may reduce the sample to more empirical moments than there are parameters...the so-called generalized method of moments. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
26 (Generalized) Method of moments 4.2. Generalized method of moments Generalized method of moments Intuition : Consider the linear regression model in an i.i.d. context y t = x tb 0 + ɛ t The exogeneity assumption E [ɛ t X ] implies the k moments conditions (for t = 1,, T ) E [ x t (y t x tb 0 ) ] = 0 k 1. Using the empirical counterpart (analogy principle) 1 T x t (y t x T tb 0 ) = 0 k 1 This is a system of k equations with k unknowns (just-identified) ( T ) 1 ( T ) ˆb ols = x t x t x t y t. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
27 (Generalized) Method of moments Generalized method of moments More generally, suppose that there exists a set of instruments z t R p (with p k) the instruments being (i) uncorrelated with the error terms and (ii) correlated with the explanatory variables such that E [ z t (y t x tb 0 ) ] = 0 p 1 E [h(z t ; δ 0 )] = 0 p 1. Using the empirical counterpart (analogy principle) 1 T T z t (y t x tb 0 ) = 0 p 1 1 T T h(z t ; δ 0 ) = 0 p 1. This system is overidentified for p > k. How can we find an estimation of b 0? Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
28 (Generalized) Method of moments Generalized method of moments The answer is provided by the generalized method of moments, i.e. an estimation method based on estimating equations that imposes the nullity of the expectation of a vector function of the observations and the parameters of interest b 0. The basic idea is to choose a value for δ such that the empirical counterpart is as close as possible to zero. As close as possible to zero means that one minimizes the T distance between 1 T z t (y t x tb 0 ) and 0 p 1 using a weighting matrix (say, W T ). Two issues 1 Choice of instruments ; 2 Choice of the (optimal) weighting matrix. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
29 (Generalized) Method of moments Generalized method of moments Definition Consider the set of moment conditions E [h(z t, δ 0 )] = 0 p 1 where δ 0 R k, p k, and h is a p-dimensional (nonlinear) linear function of the parameters of interest and depends on (i.i.d.) observations. The generalized method of moments estimator of δ 0 associated with W T (symmetric positive definite matrix) is a solution to the problem min δ [ 1 T ] [ ] T 1 T h(z t ; δ) W T h(z t ; δ). T Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
30 (Generalized) Method of moments Generalized method of moments Definition The generalized method of moments estimator of δ 0 solves the first-order conditions [ 1 T T h(z t ; ˆδ ] [ gmm ) 1 W T δ T ] T h(z t ; ˆδ gmm ) = 0 k 1. This estimator is consistent and asymptotically normally distributed. Remark : The first-order conditions correspond to k linear combinations of the moments conditions. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
31 (Generalized) Method of moments Generalized method of moments This estimator depends on W t. There exists a best GMM estimator, i.e. an optimal weighting matrix. The optimal weighting matrix is given by (i.i.d. context) W T = V 1 [h(z t, δ 0 )] E 1 [ h(z t, δ 0 )h(z t, δ 0 ) ] and the sample analog is [ ] 1 1 T h(z t, δ 0 )h(z t, δ 0 ). T It does depend on δ 0! How can we proceed in practise? Two-step GMM or iterated GMM method ; One-step method (Continuously updating estimator, exponential tilting estimator, empirical likelihood estimator, etc). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
32 (Generalized) Method of moments Generalized method of moments Example : AR(1) estimation Let (X t ) be a covariance-stationary process defined by the fundamental representation ( φ < 1) : X t = φx t 1 + ɛ t. Since (ɛ t ) is the innovation process of (X t ), one has the following moment conditions E [x t k ɛ t ] = 0 for k > 0 Past values x t 1,, x t h can be used as instruments Let z t denote the set of instruments z t = (x t 1,, x t h ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
33 (Generalized) Method of moments Generalized method of moments The moment conditions write where E [h(z t, δ 0 )] = 0 h 1 h(z t, δ 0 ) = z t (x t φ 1 x t 1 ). The empirical counterpart is defined to be x 1 T t 1 T. (x t φ 1 x t 1 ) = 0 h 1. t=h+1 x t h Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
34 (Generalized) Method of moments Generalized method of moments The 2S-GMM estimator of φ is a solution to the problem [ T t=h+1 h(z t ; ˆδ gmm ) δ ] ˆΩ 1 T [ T t=h+1 h(z t ; ˆδ gmm ) ] = 0 h 1. where ˆΩ 1 T is a first-step consistent estimate of the inverse of the variance-covariance matrix of the moments conditions and it accounts for the dependence of the moment conditions. Remark : In a first step, one solves [ T t=h+1 ] [ T ] h(z t ; ˆδ 1,gmm ) I T h(z t ; δ ˆδ 1,gmm ) = 0 h 1. t=h+1 and then compute ˆΩ 1 T (ˆδ 1,gmm ) ˆΩ 1 T. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
35 (Generalized) Method of moments Generalized method of moments Discussion Choice of instruments (optimal instruments?) and of the weighting matrix ; Bias-efficiency trade-off ; Numerical procedures can be cumbersome! Other methods can be written as a GMM estimation. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
36 Maximum likelihood estimation Overview 5. Maximum likelihood estimation 5.1. Overview Consider the linear regression model (t = 1,, T ) : y t = x tb 0 + ɛ t where (y t, x t) are i.i.d., the regressors are stochastic, x t is the t th -row of the X matrix, and It follows that ɛ t x t i.i.d.n (0, σ 2 ɛ ). y t x t i.i.d.n (x tb 0, σ 2 ɛ ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
37 Maximum likelihood estimation Overview One observes a sample of (conditionally) i.i.d. observations (y t, x t). These observations are realizations of random variables with a known distribution but unknown parameters. The basic idea is to maximize the probability of observing these realizations (given a known parametric distribution with unknown parameters). The probability of observing these realizations is provided by the joint density (or pdf) of the observations (or evaluated at the observations). This joint density is the likelihood function, with the exception that it is a function of the unknown parameters (and not the observations!). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
38 5.2. Estimation Maximum likelihood estimation Estimation Definition The exact likelihood function of the linear regression model is defined to be L(y, x; δ 0 ) = T L t (y t, x t ; δ 0 ) where L t is the (exact) likelihood function for observation t L t (y t, x t ; δ 0 ) = f Yt X t (y t x t ; δ 0 ) f Xt (x t ). where f Yt Xt (respectively, f Xt ) is the (conditional) probability density function of Y t X t (respectively, of X t ). Accordingly, the exact log-likelihood function is defined to be l(y, x; δ 0 ) = T l t (y t, x t ; δ 0 ) T logl t (y t, x t ; δ 0 ). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
39 Maximum likelihood estimation Estimation The exact likelihood function is given by T T L(y, x; δ 0 ) = f Yt X t (y t x t ; δ 0 ) f Xt (x t ) The second right-hand side term is often neglected (especially if f X does not depend on δ 0 ) and one considers the conditional likelihood function T L(y x; δ 0 ) = f Yt X t (y t x t ; δ 0 ) T ( = σ 2 ɛ 2π ) 1/2 exp ( 1 ) 2σ 2 (y t x tb 0 ) 2 ɛ = ( σɛ2π 2 ) ( ) T /2 exp 1 T 2σɛ 2 (y t x tb 0 ) 2 and the conditional log-likelihood function is given by l(y x; δ 0 ) = T 2 log(σ2 ɛ) T 2 log(2π) 1 T 2σɛ 2 (y t x tb 0 ) 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
40 Maximum likelihood estimation Estimation Definition The maximum likelihood estimator of δ 0 in the linear regression model, which is the solution of the problem max δ L(y x; δ) maxl(y x; δ), is asymptotically unbiased and efficient, and normally distributed ) l T (ˆδml δ 0 N ( 0 k 1, I 1 ) F where I F is the Fisher information matrix. δ Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
41 Maximum likelihood estimation Example : AR(1) estimation Estimation Consider the following AR(1) representation ( φ < 1) X t = φx t 1 + ɛ t. where ɛ t or ɛ t X t 1 i.i.d.n ( 0, σɛ 2 ) Therefore Accordingly, X t X t 1 = x t 1 N ( φx t 1, σɛ 2 ) ) X 1 N ( 0, σ 2 ɛ 1 φ 2 f Xt X t 1 (x t x t 1 ; δ 0 ) = ( σ 2 ɛ 2π ) 1/2 exp ( 1 f X1 (x 1 ; δ 0 ) = 2σ 2 ɛ (x t φx t 1 ) 2 ) ( ) σ 2 1/2 ɛ 1 φ 2 2π exp ( 1 ) φ2 2σɛ 2 x1 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
42 Maximum likelihood estimation Exact (log-) likelihood function Estimation The exact likelihood function is given by T L(x; δ 0 ) = L 1 (x 1 ; δ 0 ) L τ (x τ x τ 1 ; δ 0 ) = f X1 (x 1 ; δ 0 ) = τ=2 T f Xτ X τ 1 (x τ x τ 1 ; δ 0 ) τ=2 ( ) σ 2 1/2 ɛ 1 φ 2 2π exp ( 1 φ2 ( σ 2 ɛ2π ) T 1 2 exp ( 1 2σ 2 ɛ 2σ 2 ɛ x 2 1 ) ) T (x t φx t 1 ) 2. The exact log-likelihood is l(x; δ 0 ) = T 2 log(2π) 1 ( ) σ 2 2 log ɛ 1 φ 2 1 φ2 T 1 log(σ 2 2 ɛ) 1 2σɛ 2 t=2 2σ 2 ɛ T (x t φx t 1 ) 2. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50 t=2 x 2 1
43 Maximum likelihood estimation Estimation Conditional (log-) likelihood function The conditional likelihood function is given by L(x x 1 ; δ 0 ) = = T L τ (x τ x τ 1 ; δ 0 ) τ=2 T f Xτ X τ 1 (x τ x τ 1 ; δ 0 ) τ=2 = ( σ 2 ɛ 2π ) T 1 2 exp The conditional log-likelihood is ( 1 2σ 2 ɛ ) T (x t φx t 1 ) 2. t=2 l(x x 1 ; δ 0 ) = T 1 log(2π) T 1 log(σɛ 2 ) T 2σɛ 2 (x t φx t 1 ) 2. t=2 Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
44 Maximum likelihood estimation Estimation The two methods lead to different maximum likelihood estimates ˆδ cml ˆδ eml The conditional maximum likelihood estimator of φ is also the ordinary least squares estimator of φ This also holds for any AR(p). ˆφ cml = ˆφ ols The Yule-Walker, conditional maximum likelihood and (ordinary) least squares estimator are asymptotically equivalent. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
45 Discussion Maximum likelihood estimation Estimation Generally the maximum likelihood estimates are built as if the observations are Gaussian, though it is not necessary that (X t ) is Gaussian when doing the estimation. If the model is misspecified, the only difference (to some extent...) is that estimates may be less efficient. In this case, one can use the pseudo- or quasi-maximum likelihood estimation. In the case of linear models, there might be not so much gain of using exact maximum likelihood against conditional likelihood method (or least squares method). Maximum likelihood estimation often requires nonlinear numerical optimization procedures. Estimation of the Fisher information matrix can be cumbersome (especially in finite samples). Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
46 Appendix Appendix : ML estimation of a MA(q) One has Step 1 : Writing ɛ = (ɛ 1,, ɛ T ) ɛ 1 q = ɛ 1 q. =. ɛ 0 = ɛ 0 ɛ 1 = x 1 θ 1 ɛ 1 θ q ɛ 1 q ɛ 1 = x 2 θ 1 ɛ 2 θ q ɛ 2 q. =. ɛ T = x T θ 1 ɛ T θ q ɛ T q Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
47 Appendix Therefore ɛ = NX + Zɛ = ( Z N ) ( ɛ X ) with X = (x 1,, x T ), ɛ = (ɛ 1 q, ɛ 2 q,, ɛ 1, ɛ 0 ), and N = ( 0q T A 1 (θ) ) ( Iq, Z = A 2 (θ) where A 1 (θ) is a lower triangular matrix (T T ), A 2 (θ) is a T q matrix, and I q is the identity matrix of order q. ) Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
48 Appendix Step 2 : Determination of the joint density of ɛ The joint probability density function of ɛ is defined to be [ ( ) 2πσ 2 T +q 2 ɛ exp 1 ] 2σɛ 2 (NX + Zɛ ) (NX + Zɛ ). Step 3 : Determination of the marginal density of X The marginal density of X is defined to be ( ) 2πσ 2 T ( 2 ɛ detx X ) [ 1 2 exp 1 ] 2σɛ 2 (NX + Z ˆɛ ) (NX + Z ˆɛ ) with ˆɛ = (Z Z) 1 Z NX. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
49 Appendix Step 4 : Determination of the log-likelihood funcion The log-likelihood function is defined to be l (x; δ) = T 2 log(2π) T 2 log(σ2 ɛ ) T 2 log ( detx X ) S(θ) 2σ 2 ɛ where δ = (θ, σ 2 ɛ ) and S(θ) = (NX + Z ˆɛ ) (NX + Z ˆɛ ). Step 5 : Determination of ˆσ ɛ,ml 2 Using the first-order condition w.r.t. σɛ 2, one has ) S (ˆθ ml ˆσ ɛ,ml 2 =. T Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
50 Appendix Step 6 : Determination of the concentrated log-likelihood funcion The concentrated log-likelihood function is defined to be l (x; θ) = T log [S(θ)] log [ detx X ]. Step 7 : Determination of ˆθ ml ˆθ ml solves Remarks : min θ { T log [S(θ)] + log [ detx X ]}. 1 Least squares methods only focus on the first right-hand side term of the log-likelihood function. Exact methods account for both right-hand side terms. 2 This method generalizes to ARMA(p,q) models. Florian Pelgrin (HEC) Univariate time series Sept Dec / 50
Lecture 2: ARMA(p,q) models (part 3)
Lecture 2: ARMA(p,q) models (part 3) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC) Univariate time series Sept.
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Identification of univariate time series models, cont.:
More informationTime Series Analysis III
Lecture 12: Time Series Analysis III MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis III 1 Outline Time Series Analysis III 1 Time Series Analysis III MIT 18.S096 Time Series Analysis
More informationUnivariate and Multivariate Methods PEARSON. Addison Wesley
Time Series Analysis Univariate and Multivariate Methods SECOND EDITION William W. S. Wei Department of Statistics The Fox School of Business and Management Temple University PEARSON Addison Wesley Boston
More informationTime Series Analysis 1. Lecture 8: Time Series Analysis. Time Series Analysis MIT 18.S096. Dr. Kempthorne. Fall 2013 MIT 18.S096
Lecture 8: Time Series Analysis MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis 1 Outline Time Series Analysis 1 Time Series Analysis MIT 18.S096 Time Series Analysis 2 A stochastic
More informationAnalysis and Computation for Finance Time Series - An Introduction
ECMM703 Analysis and Computation for Finance Time Series - An Introduction Alejandra González Harrison 161 Email: mag208@exeter.ac.uk Time Series - An Introduction A time series is a sequence of observations
More informationEstimating an ARMA Process
Statistics 910, #12 1 Overview Estimating an ARMA Process 1. Main ideas 2. Fitting autoregressions 3. Fitting with moving average components 4. Standard errors 5. Examples 6. Appendix: Simple estimators
More informationChapter 3: The Multiple Linear Regression Model
Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics
More informationUnivariate Time Series Analysis; ARIMA Models
Econometrics 2 Spring 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing
More informationUnivariate Time Series Analysis; ARIMA Models
Econometrics 2 Fall 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Univariate Time Series Analysis We consider a single time series, y,y 2,..., y T. We want to construct simple
More informationOverview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
More informationSome useful concepts in univariate time series analysis
Some useful concepts in univariate time series analysis Autoregressive moving average models Autocorrelation functions Model Estimation Diagnostic measure Model selection Forecasting Assumptions: 1. Non-seasonal
More information1 Short Introduction to Time Series
ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The
More information3.1 Stationary Processes and Mean Reversion
3. Univariate Time Series Models 3.1 Stationary Processes and Mean Reversion Definition 3.1: A time series y t, t = 1,..., T is called (covariance) stationary if (1) E[y t ] = µ, for all t Cov[y t, y t
More informationINDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its
More informationIntroduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
More informationChapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem
Chapter Vector autoregressions We begin by taking a look at the data of macroeconomics. A way to summarize the dynamics of macroeconomic data is to make use of vector autoregressions. VAR models have become
More informationChapter 4: Vector Autoregressive Models
Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...
More informationMultivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
More informationITSM-R Reference Manual
ITSM-R Reference Manual George Weigt June 5, 2015 1 Contents 1 Introduction 3 1.1 Time series analysis in a nutshell............................... 3 1.2 White Noise Variance.....................................
More informationSales forecasting # 2
Sales forecasting # 2 Arthur Charpentier arthur.charpentier@univ-rennes1.fr 1 Agenda Qualitative and quantitative methods, a very general introduction Series decomposition Short versus long term forecasting
More informationFinancial TIme Series Analysis: Part II
Department of Mathematics and Statistics, University of Vaasa, Finland January 29 February 13, 2015 Feb 14, 2015 1 Univariate linear stochastic models: further topics Unobserved component model Signal
More informationFactor analysis. Angela Montanari
Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number
More informationThe VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.
Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium
More informationTime Series Analysis
Time Series Analysis Autoregressive, MA and ARMA processes Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 212 Alonso and García-Martos
More informationTime Series Analysis in Economics. Klaus Neusser
Time Series Analysis in Economics Klaus Neusser May 26, 2015 Contents I Univariate Time Series Analysis 3 1 Introduction 1 1.1 Some examples.......................... 2 1.2 Formal definitions.........................
More informationPS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
More informationExamples. David Ruppert. April 25, 2009. Cornell University. Statistics for Financial Engineering: Some R. Examples. David Ruppert.
Cornell University April 25, 2009 Outline 1 2 3 4 A little about myself BA and MA in mathematics PhD in statistics in 1977 taught in the statistics department at North Carolina for 10 years have been in
More informationTIME SERIES ANALYSIS
TIME SERIES ANALYSIS L.M. BHAR AND V.K.SHARMA Indian Agricultural Statistics Research Institute Library Avenue, New Delhi-0 02 lmb@iasri.res.in. Introduction Time series (TS) data refers to observations
More information1 Teaching notes on GMM 1.
Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in
More informationSTATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More informationState Space Time Series Analysis
State Space Time Series Analysis p. 1 State Space Time Series Analysis Siem Jan Koopman http://staff.feweb.vu.nl/koopman Department of Econometrics VU University Amsterdam Tinbergen Institute 2011 State
More informationSYSTEMS OF REGRESSION EQUATIONS
SYSTEMS OF REGRESSION EQUATIONS 1. MULTIPLE EQUATIONS y nt = x nt n + u nt, n = 1,...,N, t = 1,...,T, x nt is 1 k, and n is k 1. This is a version of the standard regression model where the observations
More informationWooldridge, Introductory Econometrics, 3d ed. Chapter 12: Serial correlation and heteroskedasticity in time series regressions
Wooldridge, Introductory Econometrics, 3d ed. Chapter 12: Serial correlation and heteroskedasticity in time series regressions What will happen if we violate the assumption that the errors are not serially
More informationAdvanced Forecasting Techniques and Models: ARIMA
Advanced Forecasting Techniques and Models: ARIMA Short Examples Series using Risk Simulator For more information please visit: www.realoptionsvaluation.com or contact us at: admin@realoptionsvaluation.com
More informationEconometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
More informationWhat s New in Econometrics? Lecture 8 Cluster and Stratified Sampling
What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling Jeff Wooldridge NBER Summer Institute, 2007 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of Groups and
More informationTIME SERIES ANALYSIS
TIME SERIES ANALYSIS Ramasubramanian V. I.A.S.R.I., Library Avenue, New Delhi- 110 012 ram_stat@yahoo.co.in 1. Introduction A Time Series (TS) is a sequence of observations ordered in time. Mostly these
More informationChapter 2. Dynamic panel data models
Chapter 2. Dynamic panel data models Master of Science in Economics - University of Geneva Christophe Hurlin, Université d Orléans Université d Orléans April 2010 Introduction De nition We now consider
More informationJava Modules for Time Series Analysis
Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series
More informationFrom the help desk: Bootstrapped standard errors
The Stata Journal (2003) 3, Number 1, pp. 71 80 From the help desk: Bootstrapped standard errors Weihua Guan Stata Corporation Abstract. Bootstrapping is a nonparametric approach for evaluating the distribution
More informationTime Series Analysis
Time Series Analysis Andrea Beccarini Center for Quantitative Economics Winter 2013/2014 Andrea Beccarini (CQE) Time Series Analysis Winter 2013/2014 1 / 143 Introduction Objectives Time series are ubiquitous
More informationStatistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural
More informationAnalysis of algorithms of time series analysis for forecasting sales
SAINT-PETERSBURG STATE UNIVERSITY Mathematics & Mechanics Faculty Chair of Analytical Information Systems Garipov Emil Analysis of algorithms of time series analysis for forecasting sales Course Work Scientific
More informationTime Series Analysis and Forecasting
Time Series Analysis and Forecasting Math 667 Al Nosedal Department of Mathematics Indiana University of Pennsylvania Time Series Analysis and Forecasting p. 1/11 Introduction Many decision-making applications
More informationTime Series Analysis
JUNE 2012 Time Series Analysis CONTENT A time series is a chronological sequence of observations on a particular variable. Usually the observations are taken at regular intervals (days, months, years),
More informationChapter 6: Multivariate Cointegration Analysis
Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration
More informationOn the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information
Finance 400 A. Penati - G. Pennacchi Notes on On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information by Sanford Grossman This model shows how the heterogeneous information
More information2. Linear regression with multiple regressors
2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions
More informationTopic 5: Stochastic Growth and Real Business Cycles
Topic 5: Stochastic Growth and Real Business Cycles Yulei Luo SEF of HKU October 1, 2015 Luo, Y. (SEF of HKU) Macro Theory October 1, 2015 1 / 45 Lag Operators The lag operator (L) is de ned as Similar
More informationDiscrete Time Series Analysis with ARMA Models
Discrete Time Series Analysis with ARMA Models Veronica Sitsofe Ahiati (veronica@aims.ac.za) African Institute for Mathematical Sciences (AIMS) Supervised by Tina Marquardt Munich University of Technology,
More informationMultiple Linear Regression in Data Mining
Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple
More informationImpulse Response Functions
Impulse Response Functions Wouter J. Den Haan University of Amsterdam April 28, 2011 General definition IRFs The IRF gives the j th -period response when the system is shocked by a one-standard-deviation
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationRegression III: Advanced Methods
Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models
More informationMaximum likelihood estimation of mean reverting processes
Maximum likelihood estimation of mean reverting processes José Carlos García Franco Onward, Inc. jcpollo@onwardinc.com Abstract Mean reverting processes are frequently used models in real options. For
More informationDetekce změn v autoregresních posloupnostech
Nové Hrady 2012 Outline 1 Introduction 2 3 4 Change point problem (retrospective) The data Y 1,..., Y n follow a statistical model, which may change once or several times during the observation period
More informationChapter 7. Univariate Volatility Modeling. 7.1 Why does volatility change?
Chapter 7 Univariate Volatility Modeling Note: The primary references for these notes are chapters 1 and 11 in Taylor (5). Alternative, but less comprehensive, treatments can be found in chapter 1 of Hamilton
More informationAgenda. Managing Uncertainty in the Supply Chain. The Economic Order Quantity. Classic inventory theory
Agenda Managing Uncertainty in the Supply Chain TIØ485 Produkjons- og nettverksøkonomi Lecture 3 Classic Inventory models Economic Order Quantity (aka Economic Lot Size) The (s,s) Inventory Policy Managing
More informationStatistical machine learning, high dimension and big data
Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,
More informationCCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
More informationi=1 In practice, the natural logarithm of the likelihood function, called the log-likelihood function and denoted by
Statistics 580 Maximum Likelihood Estimation Introduction Let y (y 1, y 2,..., y n be a vector of iid, random variables from one of a family of distributions on R n and indexed by a p-dimensional parameter
More informationSoftware Review: ITSM 2000 Professional Version 6.0.
Lee, J. & Strazicich, M.C. (2002). Software Review: ITSM 2000 Professional Version 6.0. International Journal of Forecasting, 18(3): 455-459 (June 2002). Published by Elsevier (ISSN: 0169-2070). http://0-
More informationTime Series Analysis
Time Series Analysis Forecasting with ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos (UC3M-UPM)
More informationRob J Hyndman. Forecasting using. 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1
Rob J Hyndman Forecasting using 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1 Outline 1 Regression with ARIMA errors 2 Example: Japanese cars 3 Using Fourier terms for seasonality 4
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationMaximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
More information11. Time series and dynamic linear models
11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd
More informationTraffic Safety Facts. Research Note. Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen
Traffic Safety Facts Research Note March 2004 DOT HS 809 718 Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen Summary This research note uses
More informationLeast Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
More informationForecasting model of electricity demand in the Nordic countries. Tone Pedersen
Forecasting model of electricity demand in the Nordic countries Tone Pedersen 3/19/2014 Abstract A model implemented in order to describe the electricity demand on hourly basis for the Nordic countries.
More informationTime Series - ARIMA Models. Instructor: G. William Schwert
APS 425 Fall 25 Time Series : ARIMA Models Instructor: G. William Schwert 585-275-247 schwert@schwert.ssb.rochester.edu Topics Typical time series plot Pattern recognition in auto and partial autocorrelations
More informationParametric Statistical Modeling
Parametric Statistical Modeling ECE 275A Statistical Parameter Estimation Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 275A SPE Version 1.1 Fall 2012 1 / 12 Why
More informationCentre for Central Banking Studies
Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics
More informationIs the Basis of the Stock Index Futures Markets Nonlinear?
University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Engineering and Information Sciences 2011 Is the Basis of the Stock
More informationLecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
More informationTIME SERIES. Syllabus... Keywords...
TIME SERIES Contents Syllabus.................................. Books................................... Keywords................................. iii iii iv 1 Models for time series 1 1.1 Time series
More informationMonte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February 5, 2013 J. Olsson Monte Carlo-based
More informationSpatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
More informationPoisson Models for Count Data
Chapter 4 Poisson Models for Count Data In this chapter we study log-linear models for count data under the assumption of a Poisson error structure. These models have many applications, not only to the
More informationOnline Appendices to the Corporate Propensity to Save
Online Appendices to the Corporate Propensity to Save Appendix A: Monte Carlo Experiments In order to allay skepticism of empirical results that have been produced by unusual estimators on fairly small
More informationNonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
More informationARMA, GARCH and Related Option Pricing Method
ARMA, GARCH and Related Option Pricing Method Author: Yiyang Yang Advisor: Pr. Xiaolin Li, Pr. Zari Rachev Department of Applied Mathematics and Statistics State University of New York at Stony Brook September
More informationSections 2.11 and 5.8
Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and
More information7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
More informationClustering in the Linear Model
Short Guides to Microeconometrics Fall 2014 Kurt Schmidheiny Universität Basel Clustering in the Linear Model 2 1 Introduction Clustering in the Linear Model This handout extends the handout on The Multiple
More informationMaster s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
More informationOne-year reserve risk including a tail factor : closed formula and bootstrap approaches
One-year reserve risk including a tail factor : closed formula and bootstrap approaches Alexandre Boumezoued R&D Consultant Milliman Paris alexandre.boumezoued@milliman.com Yoboua Angoua Non-Life Consultant
More informationChapter 9: Univariate Time Series Analysis
Chapter 9: Univariate Time Series Analysis In the last chapter we discussed models with only lags of explanatory variables. These can be misleading if: 1. The dependent variable Y t depends on lags of
More informationAnalysis of Mean-Square Error and Transient Speed of the LMS Adaptive Algorithm
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 7, JULY 2002 1873 Analysis of Mean-Square Error Transient Speed of the LMS Adaptive Algorithm Onkar Dabeer, Student Member, IEEE, Elias Masry, Fellow,
More informationproblem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved
4 Data Issues 4.1 Truncated Regression population model y i = x i β + ε i, ε i N(0, σ 2 ) given a random sample, {y i, x i } N i=1, then OLS is consistent and efficient problem arises when only a non-random
More informationForecasting the US Dollar / Euro Exchange rate Using ARMA Models
Forecasting the US Dollar / Euro Exchange rate Using ARMA Models LIUWEI (9906360) - 1 - ABSTRACT...3 1. INTRODUCTION...4 2. DATA ANALYSIS...5 2.1 Stationary estimation...5 2.2 Dickey-Fuller Test...6 3.
More informationTrend and Seasonal Components
Chapter 2 Trend and Seasonal Components If the plot of a TS reveals an increase of the seasonal and noise fluctuations with the level of the process then some transformation may be necessary before doing
More informationChapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions
Chapter 5 Analysis of Multiple Time Series Note: The primary references for these notes are chapters 5 and 6 in Enders (2004). An alternative, but more technical treatment can be found in chapters 10-11
More informationAssessing the Relative Power of Structural Break Tests Using a Framework Based on the Approximate Bahadur Slope
Assessing the Relative Power of Structural Break Tests Using a Framework Based on the Approximate Bahadur Slope Dukpa Kim Boston University Pierre Perron Boston University December 4, 2006 THE TESTING
More informationLOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as
LOGISTIC REGRESSION Nitin R Patel Logistic regression extends the ideas of multiple linear regression to the situation where the dependent variable, y, is binary (for convenience we often code these values
More informationStatistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
More informationTime Series Analysis of Aviation Data
Time Series Analysis of Aviation Data Dr. Richard Xie February, 2012 What is a Time Series A time series is a sequence of observations in chorological order, such as Daily closing price of stock MSFT in
More information