Chapter 6: Multivariate Cointegration Analysis 1
Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration Analysis - Johansen Test... 3 VI.1 The Simpelst Case: p = 1, VAR(1)... 3 VI.2 VAR(p)-Model... 12 VI.3 Model Specification... 14 VI.4 Testing the Rank of Cointegration - An Example... 16 2
VI. Multivariate Cointegration Analysis - Johansen Test VI.1 The Simpelst Case: p = 1, VAR(1) For example, there is a three dimensional vector Y consisting of the three month interest rates for the US dollar, the Euro and the Yen. Within these three I(1) variables we can find up to two cointegrating relations due to the interest rate parity and stationary expected changes in the rate of exchange. Z 1 Z 0 1 0 Y 1 1 Y Y 3
As we seen before, we have a VAR(1) model for the M I(1) variables in levels. In this simple case, we can write: Y t = µ + ΓY t-1 + ε t where: Y, µ and ε are (Mx1) vectors and Γ is a (MxM) matrix. 4
By subtracting the lagged vectors Y from both sides of the equation we receive the following relation: Y t - Y t-1 = µ + ΓY t-1 - Y t-1 + ε t or Y t = µ + (A 1 - I)Y t-1 + ε t Y t = µ + (Γ - I)Y t-1 + ε t In this equation we have an I(0) vector on the left hand side. On the right side there is a vector of constants as well as another I(0) vector ε. Thus, the term (Γ - I)Y t-1 must be also I(0). If the variables are not cointegrated, then the matrix Γ must be a unit matrix I. On the other hand, if there exists r cointegrated relations (Z is a (rx1) vector), this term can be written as a I(0) variable: (Γ - I)Y t-1 = λγ Y t-1 = λz t-1 where γ is the (rxm) matrix of the cointegration coefficients and λ is a (Mxr) matrix. 5
When multiplying with the cointegration matrix the latter results in the (MxM) matrix (Γ - I). This term is I(0) and λ can be interpreted as the matrix of the M times r error correction coefficients: Y t = µ + λz t-1 + ε t This model is a generalization of the ECM in the previous section. In the case of a VAR(1) model there appears no lagged differences in the error correction model. If the initial model constitutes a VAR(p) model then the error correction representation contains additionally (p-1) difference terms. Since the matrix (Γ - I) can be represented by the product of a (rxm) and a (Mxr) matrix, it has the rank r. This means that the number of cointegrated relations is determined by the rank of the matrix. In the marginal case r = 0, i.e Γ = I, the model reduced to a VAR model in differences (M independent random walks). If r equals M we are concerned with M stationary level data, I(0). 6
The approach of Johansen is based on the maximum likelihood estimation of the matrix (Γ - I) under the assumption of normal distributed error variables. Following the estimation the hypotheses r = 0, r = 1,, r = M-1 are tested using likelihood ratio (LR) tests. In the formulation of a VAR(p) model we receive the equation: y t = A 0 + Πy t-1 + Γ i y t i + Bx t + ε t As all factors in this equation except Π y t-1 are clearly stationary if the variables are cointegrated, it means that also Π y t-1 must be stationary. Furthermore, every cointegration relationship has to appear in Π. Even more, their number is given by the rank of Π. Π can be decomposed as Π = αβ, where the relevant elements of the α matrix are adjustment coefficients and the β matrix contains the cointegrating vectors. As the interest lies in α and β, the system should be reduced to one containing only them. p-1 i=1 7
To do that, one should regress y t on y t-1,, y t-(p-1) and then Y t-1 on the same variables. The residuals are denoted respectively R 0t and R 1t. Now the regression equation is reduced to R 0t = αβ R 1t + e t This is a multivariate regression problem: S S 00 10 S S 01 11 is the matrix of sums of squares and sums of products of R 0t and R 1t. Johansen (1991) shows that the asymptotic variance of β R 1t is β Σ 11 β, the asymptotic variance of R 0t is Σ 11 and the asymptotic covariance matrix of β R 1t and R 0t is β Σ 10, where Σ 00, Σ 10, and Σ 11 are the population counterparts of S 00, S 10 and S 11. The procedure is to maximize the likelihood function first with respect to α holding β constant and then maximize with respect to β. For α the result is: α = (β S 11 β) -1 β S 10 8
The conditional maximum of the likelihood function with respect to β is (L(β)) -2/T = S 00 -S 01 β(β S 11 β) -1 β S 10 So maximization of the likelihood function with respect to β means minimization of this determinant. By further mathematical manipulations this is equivalent to the finding of the characteristic roots of the equation: S 00-1 -1 11 S10S00S -λi = 0 The roots of this equation are the r canonical correlations between R 0t and R 1t. It means that those linear combinations of Y t-1 will be selected that are highly correlated to linear combinations of Y t after conditioning on the lagged variables Y t-1,, Y t-(p-1). 9
Denoting with λ i the characteristic value, the maximum likelihood function will be (under the assumption of normal distributed error terms): L -2 / T max = S 00 n i= 1 (1-λˆ i ) Therefore, the estimation problem is a canonical correlation analysis of the current Y t and the lagged Y. 10
The trace statistic is Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie λ n = -T ln(1-λˆ ) trace i i= r+ 1 where λˆ r+ 1,, λˆ n are the smallest characteristic roots. If the statistic is bigger than the critical value, the null hypothesis of at most r cointegrating vectors is rejected. The maximum eigenvalue statistic is λ max = -Tln(1-λˆ r + 1) If the statistic is bigger than the critical value, the null hypothesis of exactly r cointegrated vectors is rejected. The critical values for both test are derived from the trace and maximum eigenvalue of the stochastic matrix and depend on whether we include a trend (either linear or quadratic) or a constant in the VAR model. Since we have not to deal with stationary variables, but with I(1) variables, the test values are not χ 2 and follow a different distribution that is tabulated by Johansen and Juselius. 11
VI.2 VAR(p)-Model Consider a VAR of order p with M I(1) variables in levels: y t = A 0 + A 1 y t-1 + A 2 y t-2 + + A p y t-p + Bx t + ε t y t = A 0 + (A 1 - I)y t-1 + A 2 y t-2 + A 3 y t-3 + + A p y t-p + Bx t + ε t y t = A 0 + (A 1 - I)y t-1 - (A 1 - I)y t-2 + (A 1 - I)y t-2 + A 2 y t-2 + A 3 y t-3 + + A p y t-p + Bx t + ε t y t = A 0 + (A 1 - I) y t-1 + (A 2 + A 1 - I)y t-2 + A 3 y t-3 + + A p y t-p + Bx t + ε t y t = A 0 + (A 1 - I) y t-1 + (A 2 + A 1 - I)y t-2 + (A 2 + A 1 - I)y t-3 + (A 2 + A 1 - I)y t-3 + A 3 )y t-3 + + A p y t-p + Bx t + ε t y t = A 0 + (A 1 - I) y t-1 + (A 2 + A 1 - I) y t-2 + (A 3 + A 2 + A 1 - I)y t-3 + + A p y t-p + Bx t + ε t with: Γ i = (A i + A i-1 + + A 1 ), I = unit vector where: y t-p is I(1) and Γ p y t-p is I(0) y t = A 0 + Γ 1 y t-1 + Γ 2 y t-2 + + Γ p-1 y t-p-1 + Γ p y t-p + Bx t + ε t 12
Γ p calculates stationary linear combinations of the non-stationary y and the rows of Γ p are the cointegrating vectors for the elements of y. z p := Γ p y t-p is I(0) or y t = A 0 + Πy t-1 + Γ y + Bx + ε t where y t is a k-vector of non-stationary I(1) variables, x t is a d-vector of deterministic variables, and ε t is a vector of innovations. We may rewrite the VAR as, p-1 i=1 i t i t with: Π = p A i - I and p Γ = - i i=1 j=i+ 1 Aj 13
VI.3 Model Specification Eviews considers the following five cases considered by Johansen (1995): 1. The level data y t have no deterministic trends and the cointegrating equations do not have intercepts: H(r): Πy t-1 + Bx t = αβ y t-1 2. The level data y t have no deterministic trends and the cointegrating equations have intercepts: H(r): Πy t-1 + Bx t = α(β y t-1 + ρ 0 ) 3. The level data y t have linear trends but the cointegrating equations have only intercepts: H(r): Πy t-1 + Bx t = α(β y t-1 + ρ 0 ) + α γ 0 14
4. The level data y t and the cointegrating equations have linear trends: H(r): Πy t-1 + Bx t = α(β y t-1 + ρ 0 + ρ 1 t) + α γ 0 5. The level data y t have quadratic trends and the cointegrating equations have linear trends: H(r): Πy t-1 + Bx t = α(β y t-1 + ρ 0 + ρ 1 t) + α (γ 0 + γ 1 t) The terms associated with α are the deterministic terms outside the cointegrating relations. When a deterministic term appears both inside and outside the cointegrating relation, the decomposition is not uniquely identified. Johansen (1995) identifies the part that belongs inside the error correction term by orthogonally projecting the exogenous terms on to the α space so that α is the null space of α such that α α = 0. EViews uses a different identification method so that the error correction term has a sample mean of zero. More specifically, we identify the part inside the error correction term by regressing the cointegration relations β y t on a constant (and linear trend). 15
VI.4 Testing the Rank of Cointegration - An Example a) The Choice of the optimal Lag Length Lag LogL LR FPE AIC SC HQ 0 354.2837 NA 6.74e-06-3.394046-3.345746-3.374514 * indicates lag order selected by the criterion 1 2472.603 4154.772 9.50e-15-23.77395-23.58075-23.69582 LR: sequential modified LR test statistic (each test at 5% level) 2 2659.508 361.1675 1.70e-15-25.49283-25.15473* -25.35610 FPE: Final prediction error 3 2678.005 35.20814 1.55e-15-25.58459-25.10159-25.38927 AIC: Akaike information criterion 4 2701.939 44.86089 1.35e-15-25.72888-25.10097-25.47496 SC: Schwarz information criterion 5 2717.762 29.20072 1.26e-15-25.79480-25.02200-25.48229* HQ: Hannan-Quinn information criterion 6 2727.733 18.11203* 1.25e-15* -25.80419* -24.88648-25.43308 7 2734.648 12.35907 1.28e-15-25.78404-24.72143-25.35433 8 2740.000 9.411987 1.32e-15-25.74880-24.54129-25.26049 9 2746.710 11.60371 1.35e-15-25.72666-24.37426-25.17976 10 2753.414 11.39994 1.39e-15-25.70448-24.20717-25.09898 16
b) Trace statistics Unrestricted Cointegration Rank Test (Trace) Hypothesize d Trace 0.05 No. of CE(s) Eigenvalue Statistic Critical Value Prob.** None * 0.142281 48.75529 29.79707 0.0001 At most 1 * 0.071604 15.91097 15.49471 0.0433 At most 2 5.30E-05 0.011335 3.841466 0.9150 Trace test indicates 2 cointegrating eqn(s) at the 0.05 level * denotes rejection of the hypothesis at the 0.05 level **MacKinnon-Haug-Michelis (1999) p-values 17
The portion of the output tells you whether there is cointegration and the number of cointegrated vectors. Here one cannot reject the null of two cointegrating vectors using the trace test. We saw in class the differences between the trace and maximal e igenvalue tests. The latter can be evaluated from the column of eigenvalues provided. The trace statistic reports in the first block tests the null hypothesis of r cointegrated relations against the alternative of k cointegrating relations, where k is the number of endogenous variables. We can see from the second column that the first two eigenvalues are much higher compared to the last eigenvalue, which lies near zero. This suggests that there exist two cointegrated relations. The null hypothesis r = 0 and r 1 can clearly be rejected. The calculated test value of 48,75 lies outside the interval between 0 and 29,79. Also the second test value of 15,91 is higher than 15,49. 18
c) Maximum eigenvalues statistics Unrestricted Cointegration Rank Test (Maximum Eigenvalue) Hypothesized Max-Eigen 0.05 No. of CE(s) Eigenvalue Statistic Critical Value Prob.** None * 0.142281 32.84433 21.13162 0.0007 At most 1 * 0.071604 15.89963 14.26460 0.0273 At most 2 5.30E-05 0.011335 3.841466 0.9150 Max-eigenvalue test indicates 2 cointegrating eqn(s) at the 0.05 level * denotes rejection of the hypothesis at the 0.05 level **MacKinnon-Haug-Michelis (1999) p-values 19