Introduction to Time Series Analysis. Lecture 22.

Similar documents
Local classification and local likelihoods

Time Series Laboratory

Time Series Analysis

Definition 8.1 Two inequalities are equivalent if they have the same solution set. Add or Subtract the same value on both sides of the inequality.

The degree of a polynomial function is equal to the highest exponent found on the independent variables.

Basics of Statistical Machine Learning

Math 120 Final Exam Practice Problems, Form: A

Introduction to Time Series Analysis. Lecture 6.

Introduction to nonparametric regression: Least squares vs. Nearest neighbors

Method To Solve Linear, Polynomial, or Absolute Value Inequalities:

Forecasting model of electricity demand in the Nordic countries. Tone Pedersen

ITSM-R Reference Manual

Short-time FFT, Multi-taper analysis & Filtering in SPM12

Univariate and Multivariate Methods PEARSON. Addison Wesley

The Calculation of G rms

Dithering in Analog-to-digital Conversion

Lecture 18: The Time-Bandwidth Product

L9: Cepstral analysis

Lecture 2: ARMA(p,q) models (part 3)

LOGIT AND PROBIT ANALYSIS

Financial TIme Series Analysis: Part II

AN Application Note: FCC Regulations for ISM Band Devices: MHz. FCC Regulations for ISM Band Devices: MHz

Advanced Signal Processing and Digital Noise Reduction

Time Series Analysis

ANALYZER BASICS WHAT IS AN FFT SPECTRUM ANALYZER? 2-1

Performing Unit Root Tests in EViews. Unit Root Testing

Simple linear regression

Signal to Noise Instrumental Excel Assignment

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Application Note: Spread Spectrum Oscillators Reduce EMI for High Speed Digital Systems

Uses of Derivative Spectroscopy

SWISS ARMY KNIFE INDICATOR John F. Ehlers

Analysis/resynthesis with the short time Fourier transform

Statistical Confidence Calculations

RANDOM VIBRATION AN OVERVIEW by Barry Controls, Hopkinton, MA

Indiana State Core Curriculum Standards updated 2009 Algebra I

Probability and Random Variables. Generation of random variables (r.v.)

Software Review: ITSM 2000 Professional Version 6.0.

Time Series Analysis

S-DOMAIN ANALYSIS: POLES, ZEROS, AND BODE PLOTS

This is a square root. The number under the radical is 9. (An asterisk * means multiply.)

AR(p) + MA(q) = ARMA(p, q)

March 29, S4.4 Theorems about Zeros of Polynomial Functions

Design of FIR Filters

T = 1 f. Phase. Measure of relative position in time within a single period of a signal For a periodic signal f(t), phase is fractional part t p

Zero: If P is a polynomial and if c is a number such that P (c) = 0 then c is a zero of P.

Analysis and Computation for Finance Time Series - An Introduction

Time Series Analysis

7 Time series analysis

Spectrum Level and Band Level

Chapter 7 - Roots, Radicals, and Complex Numbers

Recall this chart that showed how most of our course would be organized:

Some useful concepts in univariate time series analysis

STATISTICA Formula Guide: Logistic Regression. Table of Contents

1 Norms and Vector Spaces

Zeros of Polynomial Functions

Measuring Line Edge Roughness: Fluctuations in Uncertainty

Moving averages. Rob J Hyndman. November 8, 2009

Summary Nonstationary Time Series Multitude of Representations Possibilities from Applied Computational Harmonic Analysis Tests of Stationarity

Regression III: Advanced Methods

Chapter 10 Introduction to Time Series Analysis

Zeros of Polynomial Functions

Chapter 8 - Power Density Spectrum

Zeros of a Polynomial Function

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

This unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.

Statistical Analysis of Arterial Blood Pressure (ABP), Central Venous. Pressure (CVP), and Intracranial Pressure (ICP)

Do Electricity Prices Reflect Economic Fundamentals?: Evidence from the California ISO

Zeros of Polynomial Functions

B3. Short Time Fourier Transform (STFT)

PeakVue Analysis for Antifriction Bearing Fault Detection

Nonparametric statistics and model selection

Polynomial and Rational Functions

Electronic Communications Committee (ECC) within the European Conference of Postal and Telecommunications Administrations (CEPT)

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

MATH 21. College Algebra 1 Lecture Notes

VCO Phase noise. Characterizing Phase Noise

Possibilities of Automation of the Caterpillar -SSA Method for Time Series Analysis and Forecast. Th.Alexandrov, N.Golyandina

TTT4110 Information and Signal Theory Solution to exam

3.1. RATIONAL EXPRESSIONS

ALGEBRA 2 CRA 2 REVIEW - Chapters 1-6 Answer Section

Lecture 8: Gamma regression

MATRIX TECHNICAL NOTES

Binomial Sampling and the Binomial Distribution

Stat 5102 Notes: Nonparametric Tests and. confidence interval

TIME SERIES ANALYSIS

Students Currently in Algebra 2 Maine East Math Placement Exam Review Problems

Extended Resolution TOA Measurement in an IFM Receiver

Lecture 1-6: Noise and Filters

THE UNIVERSITY OF CHICAGO, Booth School of Business Business 41202, Spring Quarter 2014, Mr. Ruey S. Tsay. Solutions to Homework Assignment #2

Jeff Thomas Tom Holmes Terri Hightower. Learn RF Spectrum Analysis Basics

SR2000 FREQUENCY MONITOR

Multiple-Comparison Procedures

Domain of a Composition

Time Series Analysis in WinIDAMS

Syntax Menu Description Options Remarks and examples Stored results Methods and formulas References Also see. Description

A course in Time Series Analysis. Suhasini Subba Rao

Least Squares Estimation

Transcription:

Introduction to Time Series Analysis. Lecture 22. 1. Review: The smoothed periodogram. 2. Examples. 3. Parametric spectral density estimation. 1

Review: Periodogram The periodogram is defined as I(ν) = X(ν) 2 = X 2 c (ν) + X 2 s(ν). X c (ν) = 1 n n t=1 X s (ν) = 1 n n t=1 cos(2πtν)x t, sin(2πtν j )x t. Under general conditions, X c (ν j ), X s (ν j ) are asymptotically independent and N(0, f(ν j )/2). Thus, EI(ˆν (n) ) f(ν), but Var(I(ˆν (n) )) f(ν) 2. 2

Review: Smoothed spectral estimators ˆf(ν) = j L n W n (j)i(ˆν (n) j/n), where the spectral window function satisfies L n, L n /n 0, W n (j) 0, W n (j) = W n ( j), W n (j) = 1, and W 2 n(j) 0. Then ˆf(ν) f(ν) (in the mean square sense), and asymptotically where d = 2/ W 2 n(j). ˆf(ν k ) f(ν k ) χ2 d d, 3

Introduction to Time Series Analysis. Lecture 22. 1. Review: The smoothed periodogram. 2. Examples. 3. Parametric spectral density estimation. 4

Example: Southern Oscillation Index Figure 4.4 in the text shows the periodogram of the SOI time series. The SOI is the scaled, standardized, mean-adjusted, difference between monthly average air pressure at sea level in Tahiti and Darwin: SOI = 10 P Tahiti P Darwin σ For the time series in the text, n = 453 months. x. The periodogram has a large peak at ν = 0.084 cycles/sample. This corresponds to 0.084 cycles per month, or a period of 1/0.084 = 11.9 months. There are smaller peaks at ν 0.02: I(0.02) 1.0. The frequency ν = 0.02 corresponds to a period of 50 months, or 4.2 years. 5

Example: Southern Oscillation Index Consider the hypothesized El Niño effect, at a period of around four years. The approximate 95% confidence interval at this frequency, ν = 1/(4 12), is 2I(ν) f(ν) 2I(ν) (0.025) χ 2 2 (0.975) 2 0.64 2 0.64 f(ν) 7.3778 0.0506 0.17 f(ν) 25.5. χ 2 2 The lower extreme of this confidence interval is around the noise baseline, so it is difficult to conclude much about the hypothesized El Niño effect. 6

Example: Southern Oscillation Index Figure 4.5 in the text shows the smoothed periodogram, with L = 9. Again, there is a large peak at ν = 0.08 cycles/month (period 12 months). There are smaller peaks at integer multiples of this frequency (harmonics). There is also a peak at ν = 0.0215 (period 46.5 months): ˆf(0.0215) 0.62. 7

Example: Southern Oscillation Index The approximate 95% confidence interval at the hypothesized El Niño frequency is 2L ˆf(ν) 2L ˆf(ν) f(ν) (0.025) χ 2 2L (0.975) 18 0.62 18 0.62 f(ν) 31.526 8.231 0.354 f(ν) 1.36. χ 2 2L The lower extreme of this confidence interval is well above the noise baseline (the level of the spectral density if the signal were white and the energy were uniformly spread across frequencies). The text modifies the number of degrees of freedom slightly, to account for the fact that the signal is padded with zeros to make n a highly composite number, which simplifies the computation of the periodogram. 8

Choosing the bandwidth A common approach is to start with a large bandwidth, and look at the effect on the spectral estimates as it is reduced ( closing the window ). As the bandwidth becomes too small, the variance gets large and the spectral estimate becomes more jagged, with spurious peaks introduced. But if it is too small, the spectral estimate is excessively smoothed, and details of the shape of the spectrum are lost. The value of L = 9 chosen in the text for Figure 4.5 corresponds to a bandwidth of B = L/n = 9/480 = 0.01875 cycles per month. This means we are averaging over frequencies in a band of this width, so we are treating the spectral density as approximately constant over this bandwidth. Equivalently, we are not hoping to resolve frequencies more finely than about half of this bandwidth. 9

Simultaneous confidence intervals We derived the confidence intervals for f(ν) assuming that ν was fixed. But in examining peaks, we might wish to choose ν after we ve seen the data. If we want to make statements about the probability that k unlikely events E 1,...,E k occur, we can use the Bonferroni inequality (also called the union bound): { k } k Pr E i Pr{E i }, i=1 i=1 and this probability is no more than kα if Pr{E i } = α. For example, if E i represents the event that f(ν i ) falls outside some confidence interval at level α, then we can bound the probability that the spectral density is far from our estimates at any of the frequencies ν 1,...,ν k. 10

Introduction to Time Series Analysis. Lecture 22. 1. Review: The smoothed periodogram. 2. Examples. 3. Parametric spectral density estimation. 11

Parametric versus nonparametric estimation Parametric estimation = estimate a model that is specified by a fixed number of parameters. Nonparametric estimation = estimate a model that is specified by a number of parameters that can grow as the sample grows. Thus, the smoothed periodogram estimates we have considered are nonparametric: the estimates of the spectral density can be parameterized by estimated values at each of the Fourier frequencies. As the sample size grows, the number of distinct frequency values increases. The time domain models we considered (linear processes) are parametric. For example, and ARMA(p,q) process can be completely specified with p + q + 1 parameters. 12

Parametric spectral estimation In parametric spectral estimation, we consider the class of spectral densities corresponding to ARMA models. Recall that, for a linear process Y t = ψ(b)w t, f y (ν) = ψ ( e 2πiν) 2 σ 2 w. For an AR model, ψ(b) = 1/φ(B), so {Y t } has the rational spectrum f y (ν) = σ 2 w φ (e 2πiν ) 2 = φ 2 p σ 2 w p j=1 e 2πiν p j 2, where p j are the poles, or roots of the polynomial φ. 13

Parametric spectral estimation The typical approach to parametric spectral estimation is to use the maximum likelihood parameter estimates (ˆφ 1,..., ˆφ p, ˆσ 2 w) for the parameters of an AR(p) model for the process, and then compute the spectral density for this estimated AR model: ˆf y (ν) = ˆσ 2 w ˆφ(e 2πiν ) 2. 14

Parametric spectral estimation For large n, Var( ˆf(ν)) 2p n f2 (ν). (There are results for the asymptotic distribution, but they are rather weak.) Notice the bias-variance trade-off in the parametric case: as we increase the number of parameters, p: The bias decreases; we can model more complex spectra. For example, with an AR(p), we cannot have more than p/2 spectral peaks in the interval (0, 1). (This is because each pair of complex conjugate poles contributes one factor and hence peak to the product.) The variance increases linearly with p. 15

ARMA spectral estimation Sometimes ARMA models are used instead: estimate the parameters of an ARMA(p,q) model and compute its spectral density (recall that ψ(b) = θ(b)/φ(b)): ˆf(ν) = ˆσ w 2 ˆθ(e 2πiν 2 ). ˆφ(e 2πiν ) However, it is more common to use large AR models, rather than ARMA models. 16

Parametric versus nonparametric spectral estimation The main advantage of parametric spectral estimation over nonparametric is that it often gives better frequency resolution of a small number of peaks: To keep the variance down with a parametric estimate, we need to make sure that we do not try to estimate too many parameters. While this may affect the bias, even p = 2 allows a sharp peak at one frequency. In contrast, to keep the variance down with a nonparametric estimate, we need to make sure that the bandwidth is not too small. This corresponds to having a smooth spectral density estimate, so the frequency resolution is limited. This is especially important if there is more than one peak at nearby frequencies. The disadvantage is the inflexibility (bias) due to the use of the restricted class of ARMA models. 17

Parametric spectral estimation: Summary Given data x 1, x 2,...,x n, 1. Estimate the AR parameters φ 1,...,φ p, σ 2 w (for example, using Yule-Walker/least squares or maximum likelihood), and choose a suitable model order p (for example, using AIC c = (n + p)/(n p 2) or BIC = p log n/n). 2. Use the estimates ˆφ 1,..., ˆφ p, ˆσ w 2 to compute the estimated spectral density: ˆf y (ν) = ˆσ 2 w ˆφ(e 2πiν ) 2. 18

Introduction to Time Series Analysis. Lecture 22. 1. Review: The smoothed periodogram. 2. Examples. 3. Parametric spectral density estimation. 19