USING OUT-OF-SAMPLE MEAN SQUARED PREDICTION ERRORS TO TEST THE MARTINGALE DIFFERENCE HYPOTHESIS. Todd E. Clark Federal Reserve Bank of Kansas City

Similar documents
Approximately normal tests for equal predictive accuracy in nested models

Statistical Tests for Multiple Forecast Comparison

Online Appendices to the Corporate Propensity to Save

Out-of-Sample Forecast Tests Robust to the Window Size Choice

Part 2: Analysis of Relationship Between Two Variables

Lesson19: Comparing Predictive Accuracy of two Forecasts: Th. Diebold-Mariano Test

Some useful concepts in univariate time series analysis

Predictability of Non-Linear Trading Rules in the US Stock Market Chong & Lam 2010

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

The Changing Relationship Between Commodity Prices and Prices of Other Assets with Global Market Integration

Journal of International Money and Finance

Financial TIme Series Analysis: Part II

Time Series Analysis

SYSTEMS OF REGRESSION EQUATIONS

2. Linear regression with multiple regressors

Least Squares Estimation

Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics

PITFALLS IN TIME SERIES ANALYSIS. Cliff Hurvich Stern School, NYU

CHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression

Regression Analysis: A Complete Example

A Century of PPP: Supportive Results from non-linear Unit Root Tests

The Canadian dollar as a commodity currency: Has the relationship changed?* Monica Mow Supervisor: Dr. Graham Voss

Time Series Analysis

INFLATION, INTEREST RATE, AND EXCHANGE RATE: WHAT IS THE RELATIONSHIP?

Is the Forward Exchange Rate a Useful Indicator of the Future Exchange Rate?

Please follow the directions once you locate the Stata software in your computer. Room 114 (Business Lab) has computers with Stata software

THE UNIVERSITY OF CHICAGO, Booth School of Business Business 41202, Spring Quarter 2014, Mr. Ruey S. Tsay. Solutions to Homework Assignment #2

Descriptive Statistics

How To Understand The Theory Of Probability

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Vector Time Series Model Representations and Analysis with XploRe

WORKING PAPER NO OUT-OF-SAMPLE FORECAST TESTS ROBUST TO THE CHOICE OF WINDOW SIZE

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.

Statistics Graduate Courses

Non-Stationary Time Series andunitroottests

Simple Linear Regression Inference

Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools

Corporate Defaults and Large Macroeconomic Shocks

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

Chapter 6: Multivariate Cointegration Analysis

Are Macroeconomic Variables Useful for Forecasting the Distribution of U.S. Inflation?

Practical. I conometrics. data collection, analysis, and application. Christiana E. Hilmer. Michael J. Hilmer San Diego State University

1 Short Introduction to Time Series

The Phillips Curve in an Era. of Well-Anchored Inflation Expectations

Factors affecting online sales

Empirical Properties of the Indonesian Rupiah: Testing for Structural Breaks, Unit Roots, and White Noise

1) Write the following as an algebraic expression using x as the variable: Triple a number subtracted from the number

Is the Basis of the Stock Index Futures Markets Nonlinear?

Chapter 9: Univariate Time Series Analysis

Week TSX Index

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

TURUN YLIOPISTO UNIVERSITY OF TURKU TALOUSTIEDE DEPARTMENT OF ECONOMICS RESEARCH REPORTS. A nonlinear moving average test as a robust test for ARCH

Testing for Granger causality between stock prices and economic growth

Chapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions

Sales forecasting # 2

Time Series Analysis in Economics. Klaus Neusser

Standard errors of marginal effects in the heteroskedastic probit model

Out-of-Sample Forecast Tests Robust to the Choice of Window Size

Performing Unit Root Tests in EViews. Unit Root Testing

MULTIPLE REGRESSIONS ON SOME SELECTED MACROECONOMIC VARIABLES ON STOCK MARKET RETURNS FROM

Monitoring Structural Change in Dynamic Econometric Models

From the help desk: Bootstrapped standard errors

Forecasting Chilean Industrial Production and Sales with Automated Procedures 1

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Relations Between Time Domain and Frequency Domain Prediction Error Methods - Tomas McKelvey

Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.

Final Exam Practice Problem Answers

Online Appendix. Supplemental Material for Insider Trading, Stochastic Liquidity and. Equilibrium Prices. by Pierre Collin-Dufresne and Vyacheslav Fos

Daily Changes in Fed Funds Futures Prices

Lean Six Sigma Analyze Phase Introduction. TECH QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY

Univariate and Multivariate Methods PEARSON. Addison Wesley

Department of Economics

Technical Efficiency Accounting for Environmental Influence in the Japanese Gas Market

FULLY MODIFIED OLS FOR HETEROGENEOUS COINTEGRATED PANELS

Appendices with Supplementary Materials for CAPM for Estimating Cost of Equity Capital: Interpreting the Empirical Evidence

Finance and Economics Discussion Series Divisions of Research & Statistics and Monetary Affairs Federal Reserve Board, Washington, D.C.

On Marginal Effects in Semiparametric Censored Regression Models

business statistics using Excel OXFORD UNIVERSITY PRESS Glyn Davis & Branko Pecar

B.3. Robustness: alternative betas estimation

ITSM-R Reference Manual

Statistics I for QBIC. Contents and Objectives. Chapters 1 7. Revised: August 2013

Non Parametric Inference

Internet Appendix to Stock Market Liquidity and the Business Cycle

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling

Time Series Analysis

ECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2

Robust Inferences from Random Clustered Samples: Applications Using Data from the Panel Survey of Income Dynamics

II. DISTRIBUTIONS distribution normal distribution. standard scores

Centre for Central Banking Studies

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

FACTOR MODEL FORECASTS OF EXCHANGE RATES. Charles Engel University of Wisconsin. Nelson C. Mark University of Notre Dame

Testing against a Change from Short to Long Memory

VICENTIU COVRIG* BUEN SIN LOW

Association Between Variables

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Introduction to General and Generalized Linear Models

Master programme in Statistics

Section Format Day Begin End Building Rm# Instructor. 001 Lecture Tue 6:45 PM 8:40 PM Silver 401 Ballerini

Introduction to Regression and Data Analysis

Pearson s Correlation

Transcription:

USING OUT-OF-SAMPLE MEAN SQUARED PREDICTION ERRORS TO TEST THE MARTINGALE DIFFERENCE HYPOTHESIS Todd E. Clark Federal Reserve Bank of Kansas City Kenneth D. West University of Wisconsin

Example: predicting monthly changes in Swiss franc/dollar exchange rate Null model: random walk (predicts value of 0 for next month s change in the exchange rate) Alternative model: use interest differential to predict y t+1 = % change in franc/dollar rate in month t+1 i t -i * t = U.S. interest rate - Swiss interest rate, monthly Predict using OLS estimates of y t+1 = b 0 +b(i t -i * t ) + e t+1

To estimate y t+1 = b 0 +b(i t -i * t ) + e t+1, we use 10 year rolling regressions ( R = 120 months), with the first sample beginning in 1975: estimate 1975:1-1984:12, predict 1985:1 estimate 1975:2-1985:1, predict 1985:2... estimate 1993:10-2003:9, predict 2003:10 So predictions run from 1985:1-2003:10 (number of predictions P = 226) Figure (plot of ^b t, t=1984:12-2003:9)

Mean squared prediction error (MSPE) from random walk model 3 2 003: 9 t = 1985:1 y2 t+1 /226 = 12.27 MSPE from interest parity model = G 2 t 003: 9 = 198 5:1 [y t+1 - ^b 0t - ^b t (i t -i * t )]2 /226 = 12.33 As in many earlier studies, random walk model has smaller MSPE (i.e., 12.27<12.33). We argue: one can nonetheless reject at the 5% level the null the random walk model predicts as well as the interest parity model by a MSPE criterion, against the alternative that the interest parity model predicts better.

Intuition: If the random walk model is right, the alternative model (the interest parity model) introduces noise into the forecasting process: the alternative model attempts to estimate parameters b 0 and b that are zero in population. In finite samples, use of the noisy estimate of the parameter will raise the estimated MSPE of the interest parity model relative to the random walk model. So if the random walk model is right, the random walk MSPE should not only be smaller but be smaller by the amount of estimation noise.

That noise is estimable (details below), and one can and should adjust MSPE comparisons for the estimated noise. In this example, the estimate happens to be 0.96. So ^F 1 2 = MSPE for random walk = 12.27 ^ F2 2 = MSPE for interest parity = 12.33 adj. = adjustment for estimation noise = 0.96 ^ F2 2 -adj. = MSPE for interest parity, adjusted for estimation noise = 12.33-0.96 = 11.37 Point estimate: 11.37<12.27, i.e., adjusted MSPE for interest parity model is less than that of random walk model One can construct in familiar fashion a standard error for a test of H 0 : After adjustment for estimation noise, MSPE for interest parity = MSPE for random walk (Diebold and Mariano (1995), West (1996))

Previous literature on MSPE comparisons for nested models: West (1996) notes specifically that his procedures maintain a rank condition that is not satisfied when models are nested; a similar rank condition is implicit in Diebold and Mariano (1995) McCracken (2000) and Clark and McCracken (2001, 2003) present analytical and simulation evidence to argue vigorously that standard critical values should not be used when models are nested Some applied papers therefore use simulations to get critical values (e.g., Mark (1995), Kilian and Taylor (2003)) Other applied papers use standard critical values, sometimes with apology (Clarida et al. (2003), Cheung et al. (2003)) by construction, would not reject null in Swiss franc example, since MSPE for random walk < (unadjusted) MSPE for interest parity

Some theoretical papers have proposed alternative tests for nested models Chao, Corradi and Swanson (2003) propose an encompassing test Giacomini and White (2003) propose a certain conditional test Our procedure - adjust the alternative model s MSPE for bias from estimation, then compare to the null model using familiar procedures - implicitly transforms the MSPE comparison to an encompassing test, though not the test proposed by Chao et al. (2001)

I. Introduction and background II. This paper s procedure III. Simulation evidence IV. Empirical example V Conclusions

II. This paper s procedure Specifically, (2.1) y t = e t (model 1: null model), (2.2) y t = X t N$+e t (model 2: alternative model). (2.4) (2.5) ^ F 2 1 / P-1 G T t=t-p+1 y2 t+1 = MSPE from model 1, ^ F 2 2 / P-1 G T t=t-p+1 (y t+1 -X t+1 N ^$ t ) 2 = MSPE from model 2. Since we have y 2 t+1 - (y t+1 -X t+1 N ^$ t ) 2 = 2y t+1 X t+1 N ^$ t - (X t+1 N ^$ t ) 2 (2.6) ^ F 2 1 - ^F 2 2 = 2(P-1 G T t=t-p+1 y t+1 X t+1 N ^$ t ) - [P -1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ].

(2.6) ^ F 2 1 - ^F 2 2 = 2(P-1 G T t=t-p+1 y t+1 X t+1 N ^$ t ) - [P -1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ]. ==== 2(P -1 G T t=t-p+1 y t+1 X t+1 N ^$ t ). 0, - [P -1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ] < 0 so under the MDS null we expect ^ F 2 1 - ^F 2 2 <0 or: we expect the MSPE from the MDS model to be less than that from the alternative model. We propose looking not at ^F 1 2 - ^F 2 2 but at ^F 1 2 - ^F 2 2 -adj., where ^ F 2 2 -adj. = ^F 2 2 - [P-1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ].

^ F 2 2 -adj. = ^F 2 2 - [P-1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ] === In the Swiss franc data, ^ F 2 2 = 12.33 [P -1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ] = 0.96 ^ F 2 2 -adj. = 11.37 < 12.27 = ^F 2 1 = MSPE from random walk model

Figures 1 and 2

Inference (3.3) ^ f t+1 / y 2 t+1 - [(y t+1 -X t+1 N ^$ t ) 2 - (X t+1 N ^$ t ) 2 ]. Then one can compute (3.4) b f = P -1 G T t=t-p+1 Then ^ f t = ^F 2 1 - ^F 2 2 -adj. = ^F 2 1 - ^F 2 2 - [P-1 G T t=t-p+1 (X t+1 N ^$ t ) 2 ]. qp b f - A N(0,V), V consistently estimated by ^V = P -1 G T t=t-p+1 Perform t-test with qp b f / q ^V ^ f t 2.

Asymptotics rolling samples used to generate regressions (rolling sample size R=120 in the exchange rate example) P = number of predictions (P=168 in the exchange rate example) Total sample size T+1=R+P As T64, P64 but R stays fixed (Giacomini and White (2003)) versus West and McCracken (1998), McCracken (2000) and Clark and McCracken (2001, 2003): T64, P64, R64. Various mixing and moment conditions (Giacomini and White (2003)); conditional heteroskedasticity allowed; stationarity; parametric linear alternative extensions to allow moment drift and parametric nonlinear alternative are straightforward

III. Simulation evidence We examine MSPE-adjusted (our statistic) MSPE: normal MSPE: McCracken (same statistic as MSPE: normal, but different critical values) CCS: a certain encompassing test of Chao, Corradi and Swanson (2001)

Two DGPs Both assume alternative model is univariate (X includes a constant and a single stochastic variable x) x is highly persistent (AR(1) with parameter 0.95) y (=presumed martingale difference sequence) much more variable than x

DGP 1: intended to capture some central characteristics of monthly exchange rate data little contemporaneous correlation between innovation in x (= interest differential) and innovation in y (=exchange rate) when simulating under alternative, y t+1 = -2x t + shock DGP 2: intended to capture some central characteristics of monthly stock price data high negative correlation between innovation in x (=dividend/price ratio) and innovation in y (=stock return) conditional heteroskedasticity of y in some parameterizations when simulating under alternative, y t+1 =.365x t + shock

Summary of simulation results for nominal one-tailed10% tests, one step ahead forecasts MSPE-adjusted has size of 5-11%; across DGPs and various sample sizes, median size is 8% MSPE-normal has size of 0-7%; median size is 0% MSPE-McCracken has size of 2-21%; median size is 9% Chao et al. has size of 10-25%; median size is 13% Size-adjusted power comparable across tests

Tables 1-6, Figure 3

Unadjusted power (DGP 1, b=-2) (4.6) R=120, P=144 R=120, P=240 MPSE-adjusted 0.477 0.608 MSPE: normal 0.087 0.086 MSPE: McCracken 0.458 0.589 CCS 0.515 0.686

IV. Empirical example Monthly bilateral exchange rates and interest differentials, U.S. vs: Canada, Japan, Switzerland, U.K. Rolling samples of size R=120 months First regression sample is 1975:1-1984:12 for Switzerland and U.K., 1980:1-1989:12 for Canada and Japan Last prediction for all four countries is 2003:12; number of monthly predictions P=166 for Canada and Japan, P = 226 for Switzerland and the U.K.

(Unadjusted) MSPE interest differential model is more than that for random walk in Japan, Switzerland, U.K.; interest differential model does beat random walk for Canada Adjusted MSPE for interest differential model is less than that for random walk in all four countries MSPE-adjusted test statistic rejects null in Switzerland and Canada at the 5% level.

MSPE- MSPE: adjusted normal ^ F1 2 ^ F2 2 ^ adj. F2 2 -adj ^F2 1 -( ^F 2 2 -adj.) ^F2 1 - ^F 2 2 Canada 2.36 2.32 0.09 2.22 0.13 0.04 CCS (0.08) 1.78 ** 0.54 3.67 Japan 11.32 11.55 0.75 10.80 0.53-0.23 (0.43) 1.23-0.52 5.23 * Swiss 12.27 12.33 0.96 11.37 0.90-0.060 (0.48) 1.88 ** -0.13 2.43 U.K. 9.73 10.16 0.44 9.72 0.01-0.43 (0.33) 0.03-1.27 0.78

V. Conclusions When testing the MDS hypothesis, MSPE comparisons should be adjusted for noise from parameter estimation Our procedure is straightforward and works reasonably well

Important extensions null model that uses estimated parameter vector to predict recursive (instead of rolling) regressions to generate predictions

1 Rolling Coefficient Estimates, Switzerland 0-1 -2-3 -4-5 1985 1987 1989 1991 1993 1995 1997 1999 2001 2003

5.0 Rolling Coefficient Estimates, Switzerland (with 90% confidence band) 2.5 0.0-2.5-5.0-7.5 1985 1987 1989 1991 1993 1995 1997 1999 2001 2003

90 80 70 60 50 40 30 20 10 0 Figure 1 Density of Simulation MSPEs Under the Null, R=120, P Varying, DGP 1 A. MSPE(1) - MSPE(2) P = 48 P = 144 P = 240 P = 1200-0.12-0.08-0.04 0.00 0.04 0.08 0.12 84 72 60 48 36 24 12 B. Average of (X*betahat)^2 P = 48 P = 144 P = 240 P = 1200 0 0.000 0.016 0.032 0.048 0.064 0.080 0.096 0.112 56 48 40 32 24 16 8 C. MSPE(1) - [MSPE(2) - average of (X*betahat)^2] P = 48 P = 144 P = 240 P = 1200 0-0.12-0.08-0.04 0.00 0.04 0.08 0.12

45 40 35 30 25 20 15 10 5 0 Figure 2 Density of Simulation MSPEs Under the Null, R Varying, P=144, DGP 1 R = 60 R = 120 R = 240 A. MSPE(1) - MSPE(2) -0.12-0.08-0.04 0.00 0.04 0.08 0.12 100 B. Average of (X*betahat)^2 R = 60 R = 120 R = 240 75 50 25 0 0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14 40 35 C. MSPE(1) - [MSPE(2) - average of (X*betahat)^2] R = 60 R = 120 R = 240 30 25 20 15 10 5 0-0.12-0.08-0.04 0.00 0.04 0.08 0.12

Figure 3 Density of Simulation MSPE-Adjusted Test Statistic Under the Null, DGP 1 0.45 A. R Varying, P=144 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00-3.2-2.4-1.6-0.8-0.0 0.8 1.6 2.4 R = 60 R = 120 R = 240 0.45 B. R=120, P Varying 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00-5.0-2.5 0.0 2.5 P = 48 P = 144 P = 240 P = 1200

Figure 4 Density of Simulation MSPE Test Statistic (Unadjusted) Under the Null, DGP 1 0.56 A. R Varying, P=144 0.48 0.40 0.32 0.24 0.16 0.08 0.00-3.2-2.4-1.6-0.8-0.0 0.8 1.6 2.4 R = 60 R = 120 R = 240 0.6 B. R=120, P Varying 0.5 0.4 0.3 0.2 0.1 0.0-5.0-2.5 0.0 2.5 P = 48 P = 144 P = 240 P = 1200

90 80 70 60 50 40 30 20 10 0 Figure 1 Density of Simulation MSPEs Under the Null, R=120, P Varying, DGP 1 A. MSPE(1) - MSPE(2) P = 48 P = 144 P = 240 P = 1200-0.12-0.08-0.04 0.00 0.04 0.08 0.12 84 72 60 48 36 24 12 B. Average of (X*betahat)^2 P = 48 P = 144 P = 240 P = 1200 0 0.000 0.016 0.032 0.048 0.064 0.080 0.096 0.112 56 48 40 32 24 16 8 C. MSPE(1) - [MSPE(2) - average of (X*betahat)^2] P = 48 P = 144 P = 240 P = 1200 0-0.12-0.08-0.04 0.00 0.04 0.08 0.12

45 40 35 30 25 20 15 10 5 0 Figure 2 Density of Simulation MSPEs Under the Null, R Varying, P=144, DGP 1 R = 60 R = 120 R = 240 A. MSPE(1) - MSPE(2) -0.12-0.08-0.04 0.00 0.04 0.08 0.12 100 B. Average of (X*betahat)^2 R = 60 R = 120 R = 240 75 50 25 0 0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14 40 35 C. MSPE(1) - [MSPE(2) - average of (X*betahat)^2] R = 60 R = 120 R = 240 30 25 20 15 10 5 0-0.12-0.08-0.04 0.00 0.04 0.08 0.12

Figure 3 Density of Simulation MSPE-Adjusted Test Statistic Under the Null, DGP 1 0.45 A. R Varying, P=144 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00-5.0-2.5 0.0 2.5 R = 60 R = 120 R = 240 0.45 B. R=120, P Varying 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00-5.0-2.5 0.0 2.5 P = 48 P = 144 P = 240 P = 1200

Figure 4 Density of Simulation MSPE Test Statistic (Unadjusted) Under the Null, DGP 1 0.56 A. R Varying, P=144 0.48 0.40 0.32 0.24 0.16 0.08 0.00-5.0-2.5 0.0 2.5 R = 60 R = 120 R = 240 0.6 B. R=120, P Varying 0.5 0.4 0.3 0.2 0.1 0.0-5.0-2.5 0.0 2.5 P = 48 P = 144 P = 240 P = 1200

Table 1 Empirical Size: DGP 1 Nominal Size = 10% A. R=60 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.074 0.072 0.072 0.075 0.080 0.092 MSPE-normal 0.009 0.002 0.000 0.000 0.000 0.000 MSPE-McCracken 0.085 0.072 0.048 0.052 0.037 0.025 CCS 0.141 0.121 0.108 0.114 0.106 0.101 B. R=120 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.069 0.068 0.063 0.065 0.069 0.081 MSPE-normal 0.020 0.009 0.003 0.001 0.000 0.000 MSPE-McCracken 0.088 0.090 0.076 0.070 0.062 0.050 CCS 0.142 0.119 0.116 0.109 0.105 0.096 C. R=240 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.082 0.074 0.071 0.070 0.066 0.076 MSPE-normal 0.040 0.022 0.014 0.006 0.001 0.000 MSPE-McCracken 0.106 0.099 0.095 0.100 0.081 0.074 CCS 0.145 0.130 0.125 0.114 0.102 0.100 Notes: 1. The predictand y t+1 is i.i.d. N(0,1); the alternative model s predictor x t follows an AR(1) with parameter 0.95; data are conditionally homoskedastic. In each simulation, one step ahead forecasts of y t+1 are formed from the martingale difference null and from rolling estimates of a regression of y t on X t = (1,x t-1 )N. 2. R is the size of the rolling regression sample. P is the number of out-of-sample predictions. 3. Our MSPE adjusted statistic, defined in (3.1) and (4.4), uses standard normal critical values. MSPE normal, defined in (4.5), refers to the usual (unadjusted) t-test for equal MSPE, and also uses standard normal critical values. MSPE-McCracken relies on the MSPE-normal statistic but uses the asymptotic critical values of McCracken (2000). 4. The number of simulations is 10,000. The table reports the fraction of simulations in which each test rejected the null using a one-sided test at the 10% level. For example, the figure of.078 in panel A, P=48, MSPE-adjusted, indicates that in 780 of the 10,000 simulations the MSPE-adjusted statistic was greater than 1.65.

Table 2 Empirical Size: DGP 2 Nominal Size = 10% A. R=60 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.094 0.081 0.079 0.083 0.084 0.089 MSPE-normal 0.019 0.005 0.001 0.000 0.000 0.000 MSPE-McCracken 0.131 0.097 0.060 0.056 0.037 0.018 CCS 0.239 0.183 0.153 0.132 0.119 0.111 B. R=120 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.098 0.085 0.080 0.074 0.077 0.086 MSPE-normal 0.040 0.018 0.008 0.002 0.000 0.000 MSPE-McCracken 0.140 0.117 0.104 0.083 0.065 0.043 CCS 0.249 0.179 0.163 0.137 0.120 0.110 C. R=240 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.100 0.085 0.082 0.075 0.074 0.078 MSPE-normal 0.056 0.035 0.023 0.010 0.002 0.000 MSPE-McCracken 0.137 0.123 0.116 0.115 0.092 0.075 CCS 0.245 0.177 0.157 0.131 0.120 0.110 Notes: 1. See the notes to Table 1.

Table 3 Empirical Size: DGP 2 with Conditional Heteroskedasticity Nominal Size = 10% A. GARCH P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.096 0.091 0.078 0.073 0.078 0.087 MSPE-normal 0.037 0.016 0.009 0.002 0.000 0.000 MSPE-McCracken 0.134 0.124 0.099 0.084 0.071 0.056 CCS 0.246 0.187 0.159 0.139 0.129 0.111 B. Multiplicative conditional heteroskedasticity P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.105 0.096 0.085 0.072 0.057 0.049 MSPE-normal 0.032 0.014 0.007 0.002 0.000 0.000 MSPE-McCracken 0.127 0.126 0.114 0.109 0.112 0.210 CCS 0.237 0.203 0.181 0.161 0.140 0.122 Notes: 1. See the notes to Table 1. The regression sample size R is 120. 2. In the upper panel of results, the predictand y t+1 is a GARCH process, with the parameterization given in equation (4.4). In the lower panel, the predictand y t+1 has conditional heteroskedasticity of the form given in equation (4.5), in which the conditional variance at t is a function of x 2 t-1.

Table 4 Size-Adjusted Power: DGP 1 Empirical Size = 10% A. R=60 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.280 0.376 0.441 0.556 0.735 0.943 MSPE 0.257 0.363 0.439 0.554 0.738 0.944 CCS 0.234 0.371 0.503 0.676 0.919 1.000 B. R=120 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.356 0.477 0.561 0.678 0.848 0.986 MSPE 0.290 0.407 0.511 0.652 0.837 0.983 CCS 0.232 0.380 0.484 0.674 0.914 1.000 C. R=240 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.383 0.527 0.635 0.764 0.918 0.997 MSPE 0.292 0.411 0.516 0.678 0.888 0.994 CCS 0.224 0.353 0.470 0.668 0.914 0.999 Notes: 1. The DGP is defined in equation 4.3, with: b=-2; e t+1 ~ i.i.d. N(0,1); x t ~ AR(1) with parameter 0.95; data are conditionally homoskedastic. In each simulation, one step ahead forecasts of y t+1 are formed from the martingale difference null and from rolling estimates of a regression of y t on X t = (1,x t-1 )N. 2. In each experiment, power is calculated by comparing the test statistics against simulation critical values, calculated as the 90th percentile of the distributions of the statistics in the corresponding size experiment reported in Table 1. 3. The number of simulations is 10,000. 4. See the notes to Table 1.

Table 5 Size-Adjusted Power: DGP 2 Empirical Size = 10% A. R=60 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.104 0.106 0.105 0.108 0.106 0.107 MSPE 0.112 0.114 0.118 0.122 0.121 0.134 CCS 0.105 0.112 0.117 0.143 0.191 0.338 B. R=120 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.119 0.120 0.125 0.136 0.147 0.162 MSPE 0.123 0.124 0.135 0.142 0.157 0.174 CCS 0.108 0.113 0.121 0.145 0.197 0.348 C. R=240 P=48 P=96 P=144 P=240 P=480 P=1200 MSPE-adjusted 0.130 0.139 0.145 0.155 0.179 0.236 MSPE 0.133 0.142 0.142 0.156 0.180 0.235 CCS 0.107 0.114 0.122 0.147 0.191 0.342 Notes: 1. The coefficient b is set to 0.365. Remaining details are as described in the notes to Table 4.

Table 6 Forecasts of Monthly Changes in U.S. Dollar Exchange Rates (1) (2) (3) (4) (5) (6) (7) (8) (9) MSPE- MSPEcountry prediction adjusted normal ^ ^ ^ sample adj. F 2 2 -adj ^F2 1 -(^F 2 2 -adj.) ^F2 1 -^F 2 2 CCS F 2 1 F 2 2 Canada 1990:1-2.36 2.32 0.09 2.22 0.13 0.04 2003:10 (0.08) 1.78 ** 0.54 3.67 Japan 1990:1-11.32 11.55 0.75 10.80 0.53-0.23 2003:10 (0.43) 1.24-0.52 5.23 * Switzerland 1985:1-12.27 12.33 0.96 11.37 0.90-0.06 2003:10 (0.48) 1.88 ** -0.13 2.43 U.K. 1985:1-9.73 10.16 0.44 9.72 0.01-0.43 2003:10 (0.33) 0.03-1.27 0.78 Notes: 1. In column (3), ^F 2 1 is the out of sample MSPE of the no change or random walk model, which forecasts a value of zero for the one month ahead change in the exchange rate. 2. In column (4), ^F 2 2 is the out of sample MSPE of a model that regresses the exchange rate on a constant and the previous month s cross-country interest differentials. The estimated regression vector and the current month s interest differential are then used to predict next month s exchange rate. Rolling regressions are used, with a sample size R of 120 months. 3. In column (5), adj. is the adjustment term P -1 G T t=t-p+1 (X t+1 N ^$ t ) 2, where: P is the number of predictions, P = 166 for Canada and Japan, P=226 for Switzerland and the U.K.; T=2003:10; X t+1 =(constant, interest differential at end of month t)n; ^$ t is the estimated regression vector. 4. In columns (7)-(9), standard errors are in parentheses and t-statistics (columns (7) and (8)) or a P 2 (2) statistic (column (9)) are in italics. Standard errors are computed as described in the text. * and ** denote test statistics significant at the 10 and 5 percent level, respectively, based on one sided tests using critical values from a standard normal (columns (7) and (8)) or chi-squared (CCS) distribution. In columns (7)and (8), and denote statistics significant at the 10 and 5 percent level based on McCracken s (2000) asymptotic critical values. 5. Data are described in the text. See notes to earlier tables for additional definitions.