Vector autoregressions, VAR

Similar documents
Time Series Analysis

Univariate Time Series Analysis; ARIMA Models

Univariate Time Series Analysis; ARIMA Models

Sales forecasting # 2

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

Lecture 2: ARMA(p,q) models (part 3)

Chapter 4: Vector Autoregressive Models

Estimating an ARMA Process

Vector Time Series Model Representations and Analysis with XploRe

Topic 5: Stochastic Growth and Real Business Cycles

y t by left multiplication with 1 (L) as y t = 1 (L) t =ª(L) t 2.5 Variance decomposition and innovation accounting Consider the VAR(p) model where

Testing for Granger causality between stock prices and economic growth

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Multivariate time series analysis is used when one wants to model and explain the interactions and comovements among a group of time series variables:

Chapter 6: Multivariate Cointegration Analysis

Introduction to Matrix Algebra

The Engle-Granger representation theorem

Data Mining: Algorithms and Applications Matrix Math Review

1 Teaching notes on GMM 1.

Quadratic forms Cochran s theorem, degrees of freedom, and all that

Introduction to General and Generalized Linear Models

E 4101/5101 Lecture 8: Exogeneity

Basics of Statistical Machine Learning

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

SYSTEMS OF REGRESSION EQUATIONS

Orthogonal Diagonalization of Symmetric Matrices

Non-Stationary Time Series andunitroottests

A Log-Robust Optimization Approach to Portfolio Management

1 Short Introduction to Time Series

Least Squares Estimation

THE UNIVERSITY OF CHICAGO, Booth School of Business Business 41202, Spring Quarter 2014, Mr. Ruey S. Tsay. Solutions to Homework Assignment #2

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

3. Regression & Exponential Smoothing

What Drives Natural Gas Prices?

Additional sources Compilation of sources:

ITSM-R Reference Manual

Statistical Machine Learning

Exam Solutions. X t = µ + βt + A t,

Volatility Forecasting I: GARCH Models

State Space Time Series Analysis

7 Gaussian Elimination and LU Factorization

Introduction to Path Analysis

Linear Models for Continuous Data

Matrix Representations of Linear Transformations and Changes of Coordinates

Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

TURUN YLIOPISTO UNIVERSITY OF TURKU TALOUSTIEDE DEPARTMENT OF ECONOMICS RESEARCH REPORTS. A nonlinear moving average test as a robust test for ARCH

Similarity and Diagonalization. Similar Matrices

problem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved

Examining the Relationship between ETFS and Their Underlying Assets in Indian Capital Market

Time Series Analysis

Part 2: Analysis of Relationship Between Two Variables

Notes on Applied Linear Regression

Univariate and Multivariate Methods PEARSON. Addison Wesley

Autocovariance and Autocorrelation

Centre for Central Banking Studies

AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.

ADVANCED FORECASTING MODELS USING SAS SOFTWARE

Factor analysis. Angela Montanari

DATA ANALYSIS II. Matrix Algorithms

ANALYZING INVESTMENT RETURN OF ASSET PORTFOLIOS WITH MULTIVARIATE ORNSTEIN-UHLENBECK PROCESSES

Spatial Statistics Chapter 3 Basics of areal data and areal data modeling

Minimum LM Unit Root Test with One Structural Break. Junsoo Lee Department of Economics University of Alabama

Gaussian Conjugate Prior Cheat Sheet

Impulse Response Functions

Stock market integration: A multivariate GARCH analysis on Poland and Hungary. Hong Li* School of Economics, Kingston University, Surrey KT1 2EE, UK

Lecture L3 - Vectors, Matrices and Coordinate Transformations

Analysis and Computation for Finance Time Series - An Introduction

Internet Appendix to Stock Market Liquidity and the Business Cycle

TIME SERIES ANALYSIS

Advanced Forecasting Techniques and Models: ARIMA

Credit Risk Models: An Overview

Factor Analysis. Factor Analysis

ARMA, GARCH and Related Option Pricing Method

How To Model A Series With Sas

Java Modules for Time Series Analysis

Chapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions

Web-based Supplementary Materials for Bayesian Effect Estimation. Accounting for Adjustment Uncertainty by Chi Wang, Giovanni

Forecasting Chilean Industrial Production and Sales with Automated Procedures 1

Understanding and Applying Kalman Filtering

1 Solving LPs: The Simplex Algorithm of George Dantzig

Testing against a Change from Short to Long Memory

Continued Fractions and the Euclidean Algorithm

Time Series Analysis III

Rob J Hyndman. Forecasting using. 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1

Implied volatility transmissions between Thai and selected advanced stock markets

Multivariate normal distribution and testing for means (see MKB Ch 3)

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

Inner Product Spaces

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Financial TIme Series Analysis: Part II

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling

1 Portfolio mean and variance

11. Time series and dynamic linear models

An Introduction to Machine Learning

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

Econometric Methods for Panel Data

Wooldridge, Introductory Econometrics, 3d ed. Chapter 12: Serial correlation and heteroskedasticity in time series regressions

Some useful concepts in univariate time series analysis

4. Multivariate Time Series Models

Transcription:

1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS15/16

2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR, identification Estimation, model specification, forecasting Impulse responses Everybody should recapitulate the properties of univariate AR(1) models!

3 / 45 Notation Here we use the multiple equations notation. Vectors are of the form: y t = (y 1t,..., y kt ) a cloumn vector of length k. y t contains the values of k variables at time t. k... the number of variables (i.g. number of equations) T... the number of observations We do not distinguish in notation between data and random variables.

Motivation 4 / 45

5 / 45 Multivariate data From an empirical/atheoretical point of view observed time series movements are often related with each another. E.g. GDP growth and unemployment rate show an inverse pattern, oil prices might be a leading indicator for other energy prices, which on the other hand have an effect on oil.

6 / 45 Weak stationarity: multivariate We assume our series y t are weakly stationary. I.e. for k = 2 [ ] E(y E(y t ) = 1t ) = µ E(y 2t ) and Cov(y t, y t l ) = E(y t µ)(y t l µ) Cov(y t, y t l ) = [ Cov(y 1t, y 1,t l ) Cov(y 1t, y 2,t l ) Cov(y 2t, y 1,t l ) Cov(y 2t, y 2,t l ) are independent of time t. l =..., 1, 0, 1,.... µ is the mean vector, µ = (µ 1,..., µ k ). Γ 0 the contemporaneous/concurrent covariance matrix Γ 0 = [γ ij (0)] is a (k k) matrix. Γ l the cross-covariance matrix of order l, Γ l = [γ ij (l)] ] = Γ l

Cross-correlations 7 / 45

8 / 45 Cross-correlation matrices, CCMs Let D = diag(σ 1,..., σ k ) with σ i = γ ii (0), the diagonal matrix of the standard deviations of y i s. Then ρ l = D 1 Γ l D 1 with ρ l = [ρ ij (l)], is the cross-correlation matrix of order l. ρ ij (l) = γ ij (l) γii (0) γ jj (0) = Cov(y it, y j,t l ) σ i σ j ρ ij (l) is the correlation coefficient between y it and y j,t l.

9 / 45 Cross-correlation matrices: properties We say with l > 0, if with γ ij (l) = Cov(y it, y j,t l ) ρ ij (l) 0: y j leads y i. ρ ji (l) 0: y i leads y j. ρ ii (l) is the autocorrelation coefficient of order l of variable y i. Properties of Γ l and ρ l : Γ 0, the covariance matrix, is symmetric (and pos def). Γ l with l 0, is i.g. not symmetric. ρ ij (l) ρ ji (l) in general γ ij (l) = γ ji ( l) and so Γ l = Γ ( l) : Cov(y it, y j,t l ) = Cov(y j,t l, y it ) = Cov(y jt, y i,t+l ) = Cov(y jt, y i,t ( l) ) Therefore we consider only CCMs with l 0.

10 / 45 Types of relationships We distinguish several types of relationships between 2 series i and j (i j): no linear relationship: ρ ij (l) = ρ ji (l) = 0 for all l 0 concurrently correlated: ρ ij (0) 0 Some cross-correlations of higher order may be 0. coupled: ρ ij (l) = ρ ji (l) = 0 for all l > 0. There is no lead-lag but possibly a concurrent relationship. unidirectional relationship (wrt time) from i to j: ρ ij (l) = 0 for all l > 0 and ρ ji (ν) 0 for some ν > 0 feedback relationship (wrt time): ρ ij (l) 0 for some l > 0 and ρ ji (ν) 0 for some ν > 0

11 / 45 Sample cross-correlations, properties The sample values are straight forward as in the univariate case. ˆγ ij (l) = 1 T T (y it ȳ i )(y j,t l ȳ j ) t=l+1 The estimates of γ s and ρ s are consistent (but biased in small samples). Under the hypothesis of multivariate white noise, the cross-correlations may be tested individually with the (±1.96/ T )-rule. The distribution in small samples may deviate considerably from the expected, e.g. for stock returns due to heteroscedasticity and fat tails. Then bootstrap resampling methods are recommended.

12 / 45 Cross-correlations and autocorrelated y i s Cross-correlations can be interpreted in a nice way only, if at least one of both series is white noise. Otherwise the autocorrelation structure of one series may interact with that of the other and spurious cross effects may be observed. One way to resolve this problem is to use a prewhitening technique. Find a univariate model for one of the two series. Apply this estimated model also to the other. Interpret the cross-correlations between the residuals of the 1st model and the residuals of the 2nd model, instead.

13 / 45 Multivariate portmanteau test The univariate Ljung-Box statistic Q(m) has been generalized to (k k)-dimensional CCMs. The null hypothesis is tested against the alternative with Q k (m) = T 2 H 0 : ρ 1 =... = ρ m = 0 H A : ρ i 0 for some 1 i m m l=1 1 T l tr(ˆγ 1 l ˆΓ ˆΓ 0 l ˆΓ 1 0 ) χ2 (k 2 m) which is χ 2 distributed with (k 2 m) degrees of freedom.

14 / 45 Multivariate portmanteau test This test is suitable for the case where the interdependence of essentially (univariate) white noise returns is in question. If only a single series is not white noise, i.e. ρ ii (l) 0 for some l > 0, the test will reject as the univariate tests will do.

The VAR model 15 / 45

16 / 45 The VAR in standard form A model taking into account/approximating multivariate dynamic relationships is the VAR(p), vector autoregression of order p. y t = ϕ 0 + Φ 1 y t 1 +... + Φ p y t p + ϵ t y t is a vector of length k. There are k equations. p is the order of the VAR. {ϵ t } is a sequence of serially uncorrelated random vectors with concurrent full rank covariance matrix Σ (not diagonal i.g.). They are coupled. (Σ = Γ (ϵ) 0, Γ(ϵ) l = 0, l 0) ϕ 0 is a (k 1) vector of constants. Φ s are (k k) coefficient matrices. The model is called VARX, if additional explanatories are included.

17 / 45 Example of VAR(1), k = 2 [ y 1t y 2t ] = or equation by equation [ ϕ (0) 1 ϕ (0) 2 y t = ϕ 0 + Φ 1 y t 1 + ϵ t ] + [ ϕ (1) 11 ϕ (1) 12 ϕ (1) 21 ϕ (1) 22 ] [ y 1,t 1 y 2,t 1 ] + [ ϵ 1t ϵ 2t ] y 1t = ϕ (0) 1 + ϕ (1) 11 y 1,t 1 + ϕ (1) 12 y 2,t 1 + ϵ 1t y 2t = ϕ (0) 2 + ϕ (1) 21 y 1,t 1 + ϕ (1) 22 y 2,t 1 + ϵ 2t The concurrent relationship between y 1 and y 2 is measured by the off-diagonal elements of Σ. If σ 12 = 0 there is no concurrent relationship. σ 12 = Cov(ϵ 1t, ϵ 2t ) = Cov(y 1t (ϵ 1t ), y 2t (ϵ 2t ) y t 1 )

18 / 45 Example of VAR(1), k = 2 Φ 1 measures the dynamic dependence in y. ϕ 12 : ϕ 12 measures the linear dependence of y 1t on y 2,t 1 in the presence of y 1,t 1. If ϕ 12 = 0, y 1t does not depend on y 2,t 1. Then, y 1t depends only on its own past. Analogous for equation 2 and ϕ 21. If ϕ 12 = 0 and ϕ 21 0: There is a unidirectional relationship from y 1 to y 2. If ϕ 12 = 0 and ϕ 21 = 0: y 1 and y 2 are coupled (in the sense that possibly σ 12 0). If ϕ 12 0 and ϕ 21 0: There is a feedback relationship between both series.

Properties of the VAR 19 / 45

20 / 45 Properties of the VAR(1) We investigate the statistical properties of a VAR(1). y t = ϕ 0 + Φy t 1 + ϵ t Taking expectations we get µ = E(y t ) = ϕ 0 + ΦE(y t 1 ) or if (I Φ) 1 exists. µ = E(y t ) = (I Φ) 1 ϕ 0 We demean the series as in the univariate case and denote ỹ t = y t µ ỹ t = Φỹ t 1 + ϵ t Repeated substitutions gives the MA( ) representation ỹ t = ϵ t + Φϵ t 1 + Φ 2 ϵ t 2 + Φ 3 ϵ t 3 +...

Properties of the VAR(1) We see that y t l is predetermined wrt ϵ t. Cov(ϵ t, y t l ) = 0, l > 0. ϵ t is uncorrelated with all past y s. ϵ t may be interpreted as a shock or innovation at time t. It has possibly an effect on the future but not on the past. Cov(y t, ϵ t ) = Σ. (Multiply by ϵ t and take expectations.) y t depends on the innovation ϵ t j with the coeff matrix Φ j. This implies that Φ j has to vanish for stationarity when j. Thus, the eigenvalues of Φ have to be smaller 1 in modulus. This is necessary and sufficient for weak stationarity. Then also (I Φ) 1 exists. (Cp. AR(1) process: y t = αy t 1 + ϵ t with α < 1.) Γ l = ΦΓ l 1, l > 0. Many properties generalize from the univariate AR to the VAR(p). 21 / 45

22 / 45 Properties, weakly stationary processes We define the (matrix) polynomial (for a VAR(p)) Φ(L) = I Φ 1 L... Φ p L p For stability it is required that the roots of the characteristic equation Φ(z) = 0, z > 1 are outside the unit circle. E.g. VAR(1): Φ(L) = I Φ L. The VAR in standard form is well defined and can be used to approximate any weakly stationary process arbitrarily well by choosing a suitable order p.

23 / 45 Representation of a VAR(p) as a VAR(1) VAR(p) : y t = ϕ 0 + Φ 1 y t 1 +... + Φ p y t p + ϵ t Every VAR(p) can be written as a (k p)-dimensional VAR(1), x t = Φ x t 1 + ξ t x t = (ỹ t p+1,..., ỹ t ) and ξ t = (0,..., 0, ϵ t ) both (k p 1). Φ = 0 I 0... 0 0 0 I... 0 0 0 0... I Φ p Φ p 1 Φ p 2... Φ 1 The (k p k p) matrix Φ is called companion matrix to the matrix polynomial Φ(L) = I Φ 1 L... Φ p L p.

24 / 45 Representation of a VAR(p) as a VAR(1) The last component of x t is the mean corrected y t, ỹ t. The last row of Φ is essentially the VAR(p) recursion in reverse order.

Structural VAR and identification 25 / 45

26 / 45 Structural VAR The VAR in standard form is also called VAR in reduced form, as it does not contain the concurrent relationships in y explicitly. A VAR in structural form is Θy t = θ 0 + Θ 1 y t 1 +... + Θ p y t p + η t Θ is the coefficient matrix of the y t s. Its diagonal elements are all 1. η t is serially uncorrelated as ϵ t, but its concurrent covariance matrix, Ω, is diagonal. (The concurrent eff are captured in a possibly non diagonal Θ.) If Θ is invertible, multiplication with Θ 1 yields the VAR in standard form with Φ j = Θ 1 Θ j ϵ t = Θ 1 η t and Σ = Θ 1 Ω(Θ 1 )

27 / 45 Identification Comparing the number of coefficients including the error covariances, we find Standard form: 0 + k + p k 2 + [(k 2 k)/2 + k] Structural form: (k 2 k) + k + p k 2 + k The number of parameters of a generally specified structural model is always (except k = 1) larger than that of the reduced form.

28 / 45 Identification There is always (at least) one structural form corresponding to a standard form (e.g. via the Cholesky decomposition of Σ, see e.g. Tsay p.350) However, the representation in a structural form is not unique without putting the required number of restrictions on the parameters in the Θ matrices. From the point of view of the structural form this is called the identification problem. A disadvantage of the structural framework is that the concurrent relationships of the y s can be interpreted only together with some economic theory. Sims (1980) conjecture was that the dynamic relationships can be interpreted well, even without economic theory. Unfortunately, this does not hold. (See the discussion of orthogonalized shocks η t below.)

29 / 45 Estimation of a VAR in standard form We assume the true relationship is a VAR(p), the model is correctly specified, and the error is normal or close to normal. LS: If all variables are inluded in each equation (no restrictions on the Φ j -parameters, j > 0) simple single equation LS is consistent and asy normal. Clearly, single equation LS cannot be efficient, since the cross-covariances in Σ are ignored. GLS: Using the estimated covariances of ϵ t improves efficiency. Interpretation of the t-statistics of the single equation estimates is straight forward.

30 / 45 Curse of dimensionality A VAR(p) in standard form has p k 2 parameters not counting the constant terms and the error variance-covariances. The number of the parameters is linear in the order of the VAR, and increases quadratically with the dimension k. The number of observations (k T ) increases only linear with k. Remark: Emperically, it is difficult to interpret e.g. a significant ˆΦ (9) 7,23.

31 / 45 Model selection Like in the univariate ARMA modeling, single coefficients in Φ j, j > 0, are not set to zero a priori. The model selection procedure starts with a maximal plausible order p max. All models with p = 0, 1,..., p max are estimated. The models can be ranked according to an information criterion. The (conditional) ML estimate of the concurrent error covariance matrix of the model with order p is ˆΣ(p) = 1 T ˆϵ t (p) ˆϵ T t(p) t=p+1 The model selection criteria are [log( ˆΣ(p) ) (2/T ) ll(p)] AIC(p) = log( ˆΣ(p) ) + 2 p k 2 /T SBC(p) = log( ˆΣ(p) ) + log(t ) p k 2 /T

32 / 45 Model validity As in the univariate case the residuals of the chosen model should be (multivariate) white noise. Here the ρ (ϵ) l for l > 0 should be zero. The multivariate Ljung-Box test may be applied with number of freedoms as the number of cross-correlation coefficients under test reduced for the number of estimated parameters: k 2 (m p)

Forecasting 33 / 45

34 / 45 Forecasting Existing concurrent structural relationships between the endogenous variables are demanding to interpret. However, due to the unique representation of the reduced/standard form the model is suitable for forecasting. The 1-step ahead forecast (conditional on data up to T ) is y T (1) = ϕ 0 + Φ 1 y T +... + Φ p y T p+1 The 1-step ahead forecast error, e T (1) = y T +1 y T (1), is e T (1) = ϵ T +1

35 / 45 Forecasting The 2-step ahead forecast is y T (2) = ϕ 0 + Φ 1 y T (1) + Φ 2 y T +... + Φ p y T p+2 The 2-step ahead forecast error is e T (2) = y T +2 y T (2) = ϵ T +2 + Φ 1 [y T +1 y T (1)] = ϵ T +2 + Φ 1 ϵ T +1 The l-step ahead forecast y T (l) converges to the mean vector µ as the forecast horizon l increases. The covariance matrices are for e T (1) Σ, for e T (2) Σ + Φ 1 ΣΦ 1. The covariance matrix of the l-step ahead forecast error converges to the covariance matrix of y t as l increases.

Impulse responses 36 / 45

37 / 45 Impulse responses Analogous to the Wold representation of univariate weakly stationary process the VAR(p) can be written in terms of concurrent and past innovations. y t = µ + ϵ t + Ψ 1 ϵ t 1 + Ψ 2 ϵ t 2 +... This moving average representation makes the impact of an innovation ϵ t i at (t i) on y t explicit. It is Ψ i. This impact is the same as of ϵ t on y t+i. {Ψ i } is called impulse response function, IRF.

38 / 45 Impulse responses Since the components of ϵ t are correlated the separating out of the effects of single innovations is difficult. Thus we consider y t = µ + η t + Ψ 1 η t 1 + Ψ 2 η t 2 +... where the components of η t are uncorrelated. Ψ j η t j = [Ψ j L]η t j = Ψ j [Lη t j ] = Ψ j ϵ t j L is the Cholesky factor of Σ: Σ = LDL and ϵ t = Lη t The diagonal matrix D is the covariance matrix of η t. Cov(η t, η t ) = D and Cov(ϵ t, ϵ t ) = Σ. Then the covariances of Lη t j are those of ϵ t j. E(ϵ t ϵ t) = E(Lη t )(Lη t ) = E(L(η t η t)l ) = LDL = Σ

39 / 45 Impulse responses Element ψ ij (l) of Ψ l is the impact of innovation η j,t of size 1 on y i,t+l. In practice instead of size 1, the standard deviation of the shock is used. Then the diagonal elements of D are transformed to 1: and Ψ = Ψ L is used instead of Ψ. LDL = LI L Transformation from the representation of y in ϵ to the representation in η is a trial to infer from the errors of the reduced form to the structural errors. However, they are not unique in general. The impulse responses will change when the Cholesky algorithm does not start with element (1,1) of Σ - as is commonly done - but e.g. with (k, k).

Exercise and References 40 / 45

41 / 45 Exercises Choose at least 2 of (1, 2, 3, 4) and 2 of (5G, 6G, 7G, 8G). 1 The univariate AR(2) process (a) y t = 0.1y t 1 + 0.2y t 2 + ϵ t (b) y t = 1 + 0.1y t 1 + 0.2y t 2 + ϵ t is given. Write the process as a VAR(1) using the companion matrix. Show that both univariate and the multivariate representations are equivalent. 2 Series of length T = 100 are available. (a) k = 5, (b) k = 20 variables should be modeled as VAR(p). What is the maximal order p, which can be used in this VAR?

42 / 45 Exercises 3 Give (a) the 5-step ahead forecast of a VAR(1), y T (5), and (b) its covariance matrix. 4 Show that the roots of I Φz = 0 are (1/λ Φ ), the inverse eigenvalues of Φ, for λ 0. 5G Choose a set of returns and test for (a) univariate, (b) multivariate dynamic dependencies with the Ljung-Box test. Use Ex2_5_LjungBox_R.txt.

43 / 45 Exercises 6G Consider returns of various stock indices before and after 9/11 2001 and interpret the dynamic dependencies in low order VARs in standard form. Choose 2 stock indices first and use p = 1, 2. Then extend the analysis to three dimensions. Interpret the results wrt unidirectional, feedback and coupled relationships. Use Ex2_6_DynDep_R.txt.

44 / 45 Exercises 7G Find the best fitting model for the dynamic dependencies of a set of stock returns wrt the AIC or SBC. Test the residuals for dynamic dependencies. See 6G and Ex2_7_VARselection_R.txt. 8G Choose a data set and a suitable VAR. Calculate (a) 1- to 10-step ahead forecasts with forecast intervals, (b) try to interpret the impulse response function, (c) interpret the composition of the forecast error variances. Use Ex2_8_VARfc_R.txt.

45 / 45 References Tsay 8.1-2 Geyer 3.1 Christopher A. Sims(1980), Macroeconomics and Reality, Econometrica 48