TIME SERIES. Syllabus... Keywords...

Save this PDF as:

Size: px
Start display at page:

Transcription

1 TIME SERIES Contents Syllabus Books Keywords iii iii iv 1 Models for time series Time series data Trend, seasonality, cycles and residuals Stationary processes Autoregressive processes Moving average processes White noise The turning point test Models of stationary processes Purely indeterministic processes ARMA processes ARIMA processes Estimation of the autocovariance function Identifying a MA(q) process Identifying an AR(p) process Distributions of the ACF and PACF Spectral methods The discrete Fourier transform The spectral density Analysing the effects of smoothing Estimation of the spectrum The periodogram Distribution of spectral estimates The fast Fourier transform Linear filters The Filter Theorem Application to autoregressive processes Application to moving average processes The general linear process i

2 5.5 Filters and ARMA processes Calculating autocovariances in ARMA models Estimation of trend and seasonality Moving averages Centred moving averages The Slutzky-Yule effect Exponential smoothing Calculation of seasonal indices Fitting ARIMA models The Box-Jenkins procedure Identification Estimation Verification Tests for white noise Forecasting with ARMA models State space models Models with unobserved states The Kalman filter Prediction Parameter estimation revisited ii

3 Syllabus Time series analysis refers to problems in which observations are collected at regular time intervals and there are correlations among successive observations. Applications cover virtually all areas of Statistics but some of the most important include economic and financial time series, and many areas of environmental or ecological data. In this course, I shall cover some of the most important methods for dealing with these problems. In the case of time series, these include the basic definitions of autocorrelations etc., then time-domain model fitting including autoregressive and moving average processes, spectral methods, and some discussion of the effect of time series correlations on other kinds of statistical inference, such as the estimation of means and regression coefficients. Books 1. P.J. Brockwell and R.A. Davis, Time Series: Theory and Methods, Springer Series in Statistics (1986). 2. C. Chatfield, The Analysis of Time Series: Theory and Practice, Chapman and Hall (1975). Good general introduction, especially for those completely new to time series. 3. P.J. Diggle, Time Series: A Biostatistical Introduction, Oxford University Press (1990). 4. M. Kendall, Time Series, Charles Griffin (1976). iii

4 Keywords ACF, 2 AR(p), 2 ARIMA(p,d,q), 6 ARMA(p,q), 5 autocorrelation function, 2 autocovariance function, 2, 5 autoregressive moving average process, 5 autoregressive process, 2 Box-Jenkins, 18 classical decomposition, 1 estimation, 18 filter generating function, 12 Gaussian process, 5 identifiability, 14 identification, 18 integrated autoregressive moving average process, 6 invertible process, 4 MA(q), 3 moving average process, 3 nondeterministic, 5 nonnegative definite sequence, 6 PACF, 11 periodogram, 15 sample partial autocorrelation coefficient, 11 second order stationary, 2 spectral density function, 8 spectral distribution function, 8 strictly stationary, 1 strongly stationary, 1 turning point test, 4 verification, 20 weakly stationary, 2 white noise, 4 Yule-Walker equations, 3 iv

5 1 Models for time series 1.1 Time series data A time series is a set of statistics, usually collected at regular intervals. Time series data occur naturally in many application areas. economics - e.g., monthly data for unemployment, hospital admissions, etc. finance - e.g., daily exchange rate, a share price, etc. environmental - e.g., daily rainfall, air quality readings. medicine - e.g., ECG brain wave activity every 2 8 secs. The methods of time series analysis pre-date those for general stochastic processes and Markov Chains. The aims of time series analysis are to describe and summarise time series data, fit low-dimensional models, and make forecasts. We write our real-valued series of observations as..., X 2, X 1, X 0, X 1, X 2,..., a doubly infinite sequence of real-valued random variables indexed by Z. 1.2 Trend, seasonality, cycles and residuals One simple method of describing a series is that of classical decomposition. The notion is that the series can be decomposed into four elements: Trend (T t ) long term movements in the mean; Seasonal effects (I t ) cyclical fluctuations related to the calendar; Cycles (C t ) other cyclical fluctuations (such as a business cycles); Residuals (E t ) other random or systematic fluctuations. The idea is to create separate models for these four elements and then combine them, either additively X t = T t + I t + C t + E t or multiplicatively 1.3 Stationary processes X t = T t I t C t E t. 1. A sequence {X t, t Z} is strongly stationary or strictly stationary if (X t1,..., X tk ) D =(X t1 +h,..., X tk +h) for all sets of time points t 1,..., t k and integer h. 2. A sequence is weakly stationary, or second order stationary if 1

6 (a) E(X t ) = µ, and (b) cov(x t, X t+k ) = γ k, where µ is constant and γ k is independent of t. 3. The sequence {γ k, k Z} is called the autocovariance function. 4. We also define Remarks. ρ k = γ k /γ 0 = corr(x t, X t+k ) and call {ρ k, k Z} the autocorrelation function (ACF). 1. A strictly stationary process is weakly stationary. 2. If the process is Gaussian, that is (X t1,..., X tk ) is multivariate normal, for all t 1,..., t k, then weak stationarity implies strong stationarity. 3. γ 0 = var(x t ) > 0, assuming X t is genuinely random. 4. By symmetry, γ k = γ k, for all k. 1.4 Autoregressive processes The autoregressive process of order p is denoted AR(p), and defined by p X t = φ r X t r + ǫ t (1.1) r=1 where φ 1,..., φ r are fixed constants and {ǫ t } is a sequence of independent (or uncorrelated) random variables with mean 0 and variance σ 2. The AR(1) process is defined by X t = φ 1 X t 1 + ǫ t. (1.2) To find its autocovariance function we make successive substitutions, to get X t = ǫ t + φ 1 (ǫ t 1 + φ 1 (ǫ t 2 + )) = ǫ t + φ 1 ǫ t 1 + φ 2 1ǫ t 2 + The fact that {X t } is second order stationary follows from the observation that E(X t ) = 0 and that the autocovariance function can be calculated as follows: γ 0 = E ( ǫ t + φ 1 ǫ t 1 + φ 2 1 ǫ t 2 + )2 ( = 1 + φ φ ) σ 2 = ( ) γ k = E φ r 1ǫ t r φ s 1ǫ t+k s = σ2 φ k 1 1 φ 2. 1 r=0 s=0 σ2 1 φ 2 1 2

7 There is an easier way to obtain these results. Multiply equation (1.2) by X t k and take the expected value, to give E(X t X t k ) = E(φ 1 X t 1 X t k ) + E(ǫ t X t k ). Thus γ k = φ 1 γ k 1, k = 1, 2,... Similarly, squaring (1.2) and taking the expected value gives E(X 2 t ) = φ 1 E(X 2 t 1) + 2φ 1 E(X t 1 ǫ t ) + E(ǫ 2 t) = φ 2 1E(X 2 t 1) σ 2 and so γ 0 = σ 2 /(1 φ 2 1 ). More generally, the AR(p) process is defined as X t = φ 1 X t 1 + φ 2 X t φ p X t p + ǫ t. (1.3) Again, the autocorrelation function can be found by multiplying (1.3) by X t k, taking the expected value and dividing by γ 0, thus producing the Yule-Walker equations ρ k = φ 1 ρ k 1 + φ 2 ρ k φ p ρ k p, k = 1, 2,... These are linear recurrence relations, with general solution of the form where ω 1,..., ω p are the roots of ρ k = C 1 ω k C p ω k p, ω p φ 1 ω p 1 φ 2 ω p 2 φ p = 0 and C 1,..., C p are determined by ρ 0 = 1 and the equations for k = 1,..., p 1. It is natural to require γ k 0 as k, in which case the roots must lie inside the unit circle, that is, ω i < 1. Thus there is a restriction on the values of φ 1,..., φ p that can be chosen. 1.5 Moving average processes The moving average process of order q is denoted MA(q) and defined by X t = q θ s ǫ t s (1.4) s=0 where θ 1,..., θ q are fixed constants, θ 0 = 1, and {ǫ t } is a sequence of independent (or uncorrelated) random variables with mean 0 and variance σ 2. It is clear from the definition that this is second order stationary and that { 0, k > q γ k = σ 2 q k s=0 θ sθ s+k, k q 3

8 We remark that two moving average processes can have the same autocorrelation function. For example, X t = ǫ t + θǫ t 1 and X t = ǫ t + (1/θ)ǫ t 1 both have ρ 1 = θ/(1 + θ 2 ), ρ k = 0, k > 1. However, the first gives ǫ t = X t θǫ t 1 = X t θ(x t 1 θǫ t 2 ) = X t θx t 1 + θ 2 X t 2 This is only valid for θ < 1, a so-called invertible process. No two invertible processes have the same autocorrelation function. 1.6 White noise The sequence {ǫ t }, consisting of independent (or uncorrelated) random variables with mean 0 and variance σ 2 is called white noise (for reasons that will become clear later.) It is a second order stationary series with γ 0 = σ 2 and γ k = 0, k The turning point test We may wish to test whether a series can be considered to be white noise, or whether a more complicated model is required. In later chapters we shall consider various ways to do this, for example, we might estimate the autocovariance function, say {ˆγ k }, and observe whether or not ˆγ k is near zero for all k > 0. However, a very simple diagnostic is the turning point test, which examines a series {X t } to test whether it is purely random. The idea is that if {X t } is purely random then three successive values are equally likely to occur in any of the six possible orders. In four cases there is a turning point in the middle. Thus in a series of n points we might expect (2/3)(n 2) turning points. In fact, it can be shown that for large n, the number of turning points should be distributed as about N(2n/3, 8n/45). We reject (at the 5% level) the hypothesis that the series is unsystematic if the number of turning points lies outside the range 2n/3 ± n/45. 4

9 2 Models of stationary processes 2.1 Purely indeterministic processes Suppose {X t } is a second order stationary process, with mean 0. Its autocovariance function is γ k = E(X t X t+k ) = cov(x t, X t+k ), k Z. 1. As {X t } is stationary, γ k does not depend on t. 2. A process is said to be purely-indeterministic if the regression of X t on X t q, X t q 1,... has explanatory power tending to 0 as q. That is, the residual variance tends to var(x t ). An important theorem due to Wold (1938) states that every purelyindeterministic second order stationary process {X t } can be written in the form X t = µ + θ 0 Z t + θ 1 Z t 1 + θ 2 Z t 2 + where {Z t } is a sequence of uncorrelated random variables. 3. A Gaussian process is one for which X t1,..., X tn has a joint normal distribution for all t 1,..., t n. No two distinct Gaussian processes have the same autocovariance function. 2.2 ARMA processes The autoregressive moving average process, ARMA(p, q), is defined by p q X t φ r X t r = θ s ǫ t s r=1 where again {ǫ t } is white noise. This process is stationary for appropriate φ, θ. Example 2.1 Consider the state space model s=0 X t = φx t 1 + ǫ t, Y t = X t + η t. Suppose {X t } is unobserved, {Y t } is observed and {ǫ t } and {η t } are independent white noise sequences. Note that {X t } is AR(1). We can write ξ t = Y t φy t 1 = (X t + η t ) φ(x t 1 + η t 1 ) = (X t φx t 1 ) + (η t φη t 1 ) = ǫ t + η t φη t 1 5

10 Now ξ t is stationary and cov(ξ t, ξ t+k ) = 0, k 2. As such, ξ t can be modelled as a MA(1) process and {Y t } as ARMA(1, 1). 2.3 ARIMA processes If the original process {Y t } is not stationary, we can look at the first order difference process X t = Y t = Y t Y t 1 or the second order differences X t = 2 Y t = ( Y ) t = Y t 2Y t 1 + Y t 2 and so on. If we ever find that the differenced process is a stationary process we can look for a ARMA model of that. The process {Y t } is said to be an autoregressive integrated moving average process, ARIMA(p, d, q), if X t = d Y t is an ARMA(p, q) process. AR, MA, ARMA and ARIMA processes can be used to model many time series. A key tool in identifying a model is an estimate of the autocovariance function. 2.4 Estimation of the autocovariance function Suppose we have data (X 1,..., X T ) from a stationary time series. We can estimate the mean by X = (1/T) T 1 X t, the autocovariance by c k = ˆγ k = (1/T) T t=k+1 (X t X)(X t k X), and the autocorrelation by r k = ˆρ k = ˆγ k /ˆγ 0. The plot of r k against k is known as the correlogram. If it is known that µ is 0 there is no need to correct for the mean and γ k can be estimated by ˆγ k = (1/T) T t=k+1 X tx t k. Notice that in defining ˆγ k we divide by T rather than by (T k). When T is large relative to k it does not much matter which divisor we use. However, for mathematical simplicity and other reasons there are advantages in dividing by T. Suppose the stationary process {X t } has autocovariance function {γ k }. Then ( ) var a t X t = a t a s cov(x t, X s ) = a t a s γ t s 0. s=1 A sequence {γ k } for which this holds for every T 1 and set of constants (a 1,..., a T ) is called a nonnegative definite sequence. The following theorem states that {γ k } is a valid autocovariance function if and only if it is nonnegative definite. 6 s=1

11 Theorem 2.2 (Blochner) The following are equivalent. 1. There exists a stationary sequence with autocovariance function {γ k }. 2. {γ k } is nonnegative definite. 3. The spectral density function, f(ω) = 1 γ k e ikω = 1 π π γ π is positive if it exists. k= γ k cos(ωk), Dividing by T rather than by (T k) in the definition of ˆγ k ensures that {ˆγ k } is nonnegative definite (and thus that it could be the autocovariance function of a stationary process), and can reduce the L 2 -error of r k. 2.5 Identifying a MA(q) process In a later lecture we consider the problem of identifying an ARMA or ARIMA model for a given time series. A key tool in doing this is the correlogram. The MA(q) process X t has ρ k = 0 for all k, k > q. So a diagnostic for MA(q) is that r k drops to near zero beyond some threshold. 2.6 Identifying an AR(p) process The AR(p) process has ρ k decaying exponentially. This can be difficult to recognise in the correlogram. Suppose we have a process X t which we believe is AR(k) with k X t = φ j,k X t j + ǫ t j=1 with ǫ t independent of X 1,..., X t 1. Given the data X 1,..., X T, the least squares estimates of (φ 1,k,..., φ k,k ) are obtained by minimizing 1 T t=k+1 ( X t k=1 ) 2 k φ j,k X t j. j=1 This is approximately equivalent to solving equations similar to the Yule-Walker equations, k ˆγ j = ˆφ l,kˆγ j l, j = 1,..., k These can be solved by the Levinson-Durbin recursion: l=1 7

12 Step 0. σ 2 0 := ˆγ 0, ˆφ1,1 = ˆγ 1 /ˆγ 0, k := 0 Step 1. Repeat until ˆφ k,k near 0: ˆφ k,k := ( ˆγ k k := k + 1 k 1 j=1 ˆφ j,k 1ˆγ k j )/ σ 2 k 1 ˆφ j,k := ˆφ j,k 1 ˆφ k,k ˆφk j,k 1, for j = 1,..., k 1 σ 2 k := σ2 k 1 (1 ˆφ 2 k,k ) We test whether the order k fit is an improvement over the order k 1 fit by looking to see if ˆφ k,k is far from zero. The statistic ˆφ k,k is called the kth sample partial autocorrelation coefficient (PACF). If the process X t is genuinely AR(p) then the population PACF, φ k,k, is exactly zero for all k > p. Thus a diagnostic for AR(p) is that the sample PACFs are close to zero for k > p. 2.7 Distributions of the ACF and PACF Both the sample ACF and PACF are approximately normally distributed about their population values, and have standard deviation of about 1/ T, where T is the length of the series. A rule of thumb it that ρ k is negligible (and similarly φ k,k ) if r k (similarly ˆφ k,k ) lies between ±2/ T. (2 is an approximation to Recall that if Z 1,..., Z n N(µ, 1), a test of size 0.05 of the hypothesis H 0 : µ = 0 against H 1 : µ 0 rejects H 0 if and only if Z lies outside ±1.96/ n). Care is needed in applying this rule of thumb. It is important to realize that the sample autocorrelations, r 1, r 2,..., (and sample partial autocorrelations, ˆφ 1,1, ˆφ 2,2,...) are not independently distributed. The probability that any one r k should lie outside ±2/ T depends on the values of the other r k. A portmanteau test of white noise (due to Box & Pierce and Ljung & Box) can be based on the fact that approximately Q m = T(T + 2) m (T k) 1 rk 2 χ 2 m. k=1 The sensitivity of the test to departure from white noise depends on the choice of m. If the true model is ARMA(p, q) then greatest power is obtained (rejection of the white noise hypothesis is most probable) when m is about p + q. 8

13 3 Spectral methods 3.1 The discrete Fourier transform If h(t) is defined for integers t, the discrete Fourier transform of h is The inverse transform is H(ω) = t= h(t)e iωt, π ω π h(t) = 1 π e iωt H(ω) dω. 2π π If h(t) is real-valued, and an even function such that h(t) = h( t), then and 3.2 The spectral density H(ω) = h(0) + 2 h(t) = 1 π π 0 h(t) cos(ωt) cos(ωt)h(ω) dω. The Wiener-Khintchine theorem states that for any real-valued stationary process there exists a spectral distribution function, F( ), which is nondecreasing and right continuous on [0, π] such that F(0) = 0, F(π) = γ 0 and γ k = π 0 cos(ωk) df(ω). The integral is a Lebesgue-Stieltges integral and is defined even if F has discontinuities. Informally, F(ω 2 ) F(ω 1 ) is the contribution to the variance of the series made by frequencies in the range (ω 1, ω 2 ). F( ) can have jump discontinuities, but always can be decomposed as F(ω) = F 1 (ω) + F 2 (ω) where F 1 ( ) is a nondecreasing continuous function and F 2 ( ) is a nondecreasing step function. This is a decomposition of the series into a purely indeterministic component and a deterministic component. Suppose the process is purely indeterministic, (which happens if and only if k γ k < ). In this case F( ) is a nondecreasing continuous function, and differentiable at all points (except possibly on a set of measure zero). Its derivative f(ω) = F (ω) exists, and is called the spectral density function. Apart from a 9

14 multiplication by 1/π it is simply the discrete Fourier transform of the autocovariance function and is given by with inverse f(ω) = 1 π k= γ k = γ k e ikω = 1 π γ π π 0 γ k cos(ωk), k=1 cos(ωk)f(ω) dω. Note. Some authors define the spectral distribution function on [ π, π]; the use of negative frequencies makes the interpretation of the spectral distribution less intuitive and leads to a difference of a factor of 2 in the definition of the spectra density. Notice, however, that if f is defined as above and extended to negative frequencies, f( ω) = f(ω), then we can write γ k = π π 1 2 eiωk f(ω) dω. Example 3.1 (a) Suppose {X t } is i.i.d., γ 0 = var(x t ) = σ 2 > 0 and γ k = 0, k 1. Then f(ω) = σ 2 /π. The fact that the spectral density is flat means that all frequencies are equally present accounts for our calling this sequence white noise. (b) As an example of a process which is not purely indeterministic, consider X t = cos(ω 0 t + U) where ω 0 is a value in [0, π] and U U[ π, π]. The process has zero mean, since E(X t ) = 1 π cos(ω 0 t + u) du = 0 2π π and autocovariance γ k = E(X t, X t+k ) = 1 π 2π = 1 2π π π π cos(ω 0 t + u) cos(ω 0 t + ω 0 k + u)du 1 2 [cos(ω 0k) + cos(2ω 0 t + ω 0 k + 2u)]du = 1 1 2π 2 [2π cos(ω 0k) + 0] = 1 2 cos(ω 0k). Hence X t is second order stationary and we have γ k = 1 2 cos(ω 0k), F(ω) = 1 2 I [ω ω 0 ] and f(ω) = 1 2 δ ω 0 (ω). 10

15 Note that F is a nondecreasing step function. More generally, the spectral density f(ω) = n j=1 1 2 a jδ ωj (ω) corresponds to the process X t = n j=1 a j cos(ω j t + U j ) where ω j [0, π] and U 1,..., U n are i.i.d. U[ π, π]. (c) The MA(1) process, X t = θ 1 ǫ t 1 + ǫ t, where {ǫ t } is white noise. Recall γ 0 = (1 + θ 2 1)σ 2, γ 1 = θ 1 σ 2, and γ k = 0, k > 1. Thus f(ω) = σ2 (1 + 2θ 1 cosω + θ 2 1) π (d) The AR(1) process, X t = φ 1 X t 1 + ǫ t, where {ǫ t } is white noise. Recall var(x t ) = φ 2 1 var(x t 1) + σ 2 = γ 0 = φ 2 1 γ 0 + σ 2 = γ 0 = σ 2 /(1 φ 2 1 ) where we need φ 1 < 1 for X t stationary. Also, γ k = cov(x t, X t k ) = cov(φ 1 X t 1 + ǫ t, X t k ) = φ 1 γ k 1. So γ k = φ k 1 γ 0, k Z. Thus f(ω) = γ 0 π + 2 π = γ 0 π = k=1 φ k 1γ 0 cos(kω) = γ 0 π { {1 + φ 1e iω 1 φ 1 e iω + φ 1e iω 1 φ 1 e iω σ 2 π(1 2φ 1 cosω + φ 2 1 ). 1 + } k=1 = γ 0 π. φ k 1 [ e iωk + e iωk]} 1 φ φ 1 cosω + φ 2 1 Note that φ > 0 has power at low frequency, whereas φ < 0 has power at high frequency. 4 φ 1 = 1 2 φ 1 = f(ω) f(ω) ω ω 11

16 Plots above are the spectral densities for AR(1) processes in which {ǫ t } is Gaussian white noise, with σ 2 /π = 1. Samples for 200 data points are shown below. φ 1 = φ 1 = Analysing the effects of smoothing Let {a s } be a sequence of real numbers. A linear filter of {X t } is Y t = a s X t s. s= In Chapter 5 we show that the spectral density of {Y t } is given by where a(z) is the transfer function f Y (ω) = a(ω) 2 f X (ω), a(ω) = s= a s e iωs. This result can be used to explore the effect of smoothing a series. Example 3.2 Suppose the AR(1) series above, with φ 1 = 0.5, is smoothed by a moving average on three points, so that smoothed series is Y t = 1 3 [X t+1 + X t + X t 1 ]. Then a(ω) 2 = 1 3 e iω eiω 2 = 1 9 (1 + 2 cosω)2. Notice that γ X (0) = 4π/3, γ Y (0) = 2π/9, so {Y t } has 1/6 the variance of {X t }. Moreover, all components of frequency ω = 2π/3 (i.e., period 3) are eliminated in the smoothed series. a(ω) 2 1 f Y (ω) = a(ω) 2 f X (ω) ω ω

17 4 Estimation of the spectrum 4.1 The periodogram Suppose we have T = 2m + 1 observations of a time series, y 1,..., y T. Define the Fourier frequencies, ω j = 2πj/T, j = 1,..., m, and consider the regression model m m y t = α 0 + α j cos(ω j t) + β j sin(ω j t), j=1 which can be written as a general linear model, Y = Xθ + ǫ, where y 1 1 c 11 s 11 c m1 s m1 Y =., X =....., θ = y T 1 c 1T s 1T c mt s mt j=1 α 0 α 1 β 1. α m, ǫ = ǫ 1. ǫ T c jt = cos(ω j t), s jt = sin(ω j t). The least squares estimates in this model are given by ˆθ = (X X) 1 X Y. β m Note that and = c jt + i e iω jt = eiω j (1 eiω jt ) 1 e iω j s jt = 0 = c jt s jt = 1 2 c 2 jt = 1 2 s 2 jt = 1 2 c jt s kt = = 0 c jt = sin(2ω j t) = 0, {1 + cos(2ω j t)} = T/2, {1 cos(2ω j t)} = T/2, c jt c kt = s jt = 0 s jt s kt = 0, j k. 13

18 Using these, we have ˆα 0 T 0 0 ˆα 1 ˆθ =. = 0 T/ ˆβ m 0 0 T/2 1 t y t t c 1ty t. t s mty t ȳ = (2/T) t c 1ty t. (2/T) t s mty t and the regression sum of squares is Ŷ Ŷ = Y X(X X) 1 X Y = Tȳ 2 + m j=1 2 T { } 2 { c jt y t + } 2 s jt y t. Since we are fitting T unknown parameters to T data points, the model fits with no residual error, i.e., Ŷ = Y. Hence { m } 2 { } 2 (y t ȳ) 2 2 = c jt y t + s jt y t. T j=1 This motivates definition of the periodogram as { 2 { } 2 I(ω) = 1 y t cos(ωt)} + y t sin(ωt). πt A factor of (1/2π) has been introduced into this definition so that the sample variance, ˆγ 0 = (1/T) T (y t ȳ) 2, equates to the sum of the areas of m rectangles, whose heights are I(ω 1 ),..., I(ω m ), whose widths are 2π/T, and whose bases are centred at ω 1,..., ω m. I.e., ˆγ 0 = (2π/T) m j=1 I(ω j). These rectangles approximate the area under the curve I(ω), 0 ω π. I(ω) I(ω 5 ) 0 2π/T ω 5 π 14

19 Using the fact that T c jt = T s jt = 0, we can write { 2 { πti(ω j ) = y t cos(ω j t)} + y t sin(ω j t) Hence } 2 { 2 { = (y t ȳ) cos(ω j t)} + (y t ȳ) sin(ω j t) = = = (y t ȳ)e iω jt (y t ȳ)e iω jt 2 (y s ȳ)e iω js s=1 T 1 (y t ȳ) I(ω j ) = 1 πˆγ π k=1 t=k+1 } 2 (y t ȳ)(y t k ȳ) cos(ω j k). T 1 k=1 ˆγ k cos(ω j k). I(ω) is therefore a sample version of the spectral density f(ω). 4.2 Distribution of spectral estimates If the process is stationary and the spectral density exists then I(ω) is an almost unbiased estimator of f(ω), but it is a rather poor estimator without some smoothing. Suppose {y t } is Gaussian white noise, i.e., y 1,..., y T are iid N(0, σ 2 ). Then for any Fourier frequency ω = 2πj/T, where A(ω) = I(ω) = 1 πt [ A(ω) 2 + B(ω) 2], (4.1) y t cos(ωt), B(ω) = Clearly A(ω) and B(ω) have zero means, and y t sin(ωt). (4.2) var[a(ω)] = σ 2 cos 2 (ωt) = Tσ 2 /2, var[b(ω)] = σ 2 sin 2 (ωt) = Tσ 2 /2, 15

20 cov[a(ω), B(ω)] = E [ ] y t y s cos(ωt) sin(ωs) = σ 2 cos(ωt) sin(ωt) = 0. s=1 Hence A(ω) 2/Tσ 2 and B(ω) 2/Tσ 2 are independently distributed as N(0, 1), and 2 [ A(ω) 2 + B(ω) 2] /(Tσ 2 ) is distributed as χ 2 2. This gives I(ω) (σ 2 /π)χ 2 2/2. Thus we see that I(w) is an unbiased estimator of the spectrum, f(ω) = σ 2 /π, but it is not consistent, since var[i(ω)] = σ 4 /π 2 does not tend to 0 as T. This is perhaps surprising, but is explained by the fact that as T increases we are attempting to estimate I(ω) for an increasing number of Fourier frequencies, with the consequence that the precision of each estimate does not change. By a similar argument, we can show that for any two Fourier frequencies, ω j and ω k the estimates I(ω j ) and I(ω k ) are statistically independent. These conclusions hold more generally. Theorem 4.1 Let {Y t } be a stationary Gaussian process with spectrum f(ω). Let I( ) be the periodogram based on samples Y 1,..., Y T, and let ω j = 2πj/T, j < T/2, be a Fourier frequency. Then in the limit as T, (a) I(ω j ) f(ω j )χ 2 2/2. (b) I(ω j ) and I(ω k ) are independent for j k. Assuming that the underlying spectrum is smooth, f(ω) is nearly constant over a small range of ω. This motivates use of an estimator for the spectrum of ˆf(ω j ) = 1 2p + 1 p I(ω j+l ). Then ˆf(ω j ) f(ω j )χ 2 2(2p+1) /[2(2p+1)], which has variance f(ω)2 /(2p+1). The idea is to let p as T. 4.3 The fast Fourier transform l= p I(ω j ) can be calculated from (4.1) (4.2), or from I(ω j ) = 1 πt y t e iω jt Either way, this requires of order T multiplications. Hence to calculate the complete periodogram, i.e., I(ω 1 ),..., I(ω m ), requires of order T 2 multiplications. Computation effort can be reduced significantly by use of the fast Fourier transform, which computes I(ω 1 ),..., I(ω m ) using only order T log 2 T multiplications

21 5 Linear filters 5.1 The Filter Theorem A linear filter of one random sequence {X t } into another sequence {Y t } is Y t = s= a s X t s. (5.1) Theorem 5.1 (the filter theorem) Suppose X t is a stationary time series with spectral density f X (ω). Let {a t } be a sequence of real numbers such that t= a t <. Then the process Y t = s= a sx t s is a stationary time series with spectral density function f Y (ω) = A(e iω ) 2 f X (ω) = a(ω) 2 f X (ω), where A(z) is the filter generating function A(z) = s= a s z s, z 1. and a(ω) = A(e iω ) is the transfer function of the linear filter. Proof. cov(y t, Y t+k ) = r Z a r a s cov(x t r, X t+k s ) s Z = r,s Za r a s γ k+r s = r,s Z = = = π π π π π π a r a s π π 1 2 eiω(k+r s) f X (ω)dω A(e iω )A(e iω ) 1 2 eiωk f X (ω)dω 1 2 eiωk A(e iω ) 2 f X (ω)dω 1 2 eiωk f Y (ω)dω. Thus f Y (ω) is the spectral density for Y and Y is stationary. 5.2 Application to autoregressive processes Let us use the notation B for the backshift operator B 0 = I, (B 0 X) t = X t, (BX) t = X t 1, (B 2 X) t = X t 2,... 17

22 Then the AR(p) process can be written as or φ(b)x = ǫ, where φ is the function (I p r=1 φ rb r ) X = ǫ φ(z) = 1 p r=1 φ rz r. By the filter theorem, f ǫ (ω) = φ ( e iω ) 2 f X (ω ), so since f ǫ (ω) = σ 2 /π, f X (ω) = σ 2 π φ(e iω ) 2. (5.2) As f X (ω) = (1/π) k= γ ke iωk, we can calculate the autocovariances by expanding f X (ω) as a power series in e iω. For this to work, the zeros of φ(z) must lie outside the unit circle in C. This is the stationarity condition for the AR(p) process. Example 5.2 For the AR(1) process, X t φ 1 X t 1 = ǫ t, we have φ(z) = 1 φ 1 z, with its zero at z = 1/φ 1. The stationarity condition is φ 1 < 1. Using (5.2) we find f X (ω) = σ 2 π 1 φe iω 2 = σ 2 π(1 2φ cosω + φ 2 ), which is what we found by other another method in Example 3.1(c). To find the autocovariances we can write, taking z = e iω, 1 φ 1 (z) = 1 2 φ 1 (z)φ 1 (1/z) = 1 (1 φ 1 z)(1 φ 1 /z) = = k= z k (φ k 1 (1 + φ2 1 + φ )) = = f X (ω) = 1 π k= and so γ k = σ 2 φ k 1 /(1 φ2 1 ) as we saw before. σ 2 φ k 1 1 φ 2 e iωk 1 k= r=0 φ r 1 zr z k φ k 1 1 φ 2 1 s=0 φ s 1 z s In general, it is often easier to calculate the spectral density function first, using filters, and then deduce the autocovariance function from it. 5.3 Application to moving average processes The MA(q) process X t = ǫ t + q s=1 θ sǫ t s can be written as X = θ(b)ǫ where θ(z) = q s=0 θ sb s. By the filter theorem, f X (ω) = θ(e iω ) 2 (σ 2 /π). 18

23 Example 5.3 For the MA(1), X t = ǫ t + θ 1 ǫ t 1, θ(z) = 1 + θ 1 z and f X (ω) = σ2 ( ) 1 + 2θ1 cosω + θ1 2. π As above, we can obtain the autocovariance function by expressing f X (ω) as a power series in e iω. We have f X (ω) = σ2 ( θ1 e iω + (1 + θ 2 π 1) + θ 1 e iω) = σ2 π θ(eiω )θ(e iω ) So γ 0 = σ 2 (1 + θ 2 1 ), γ 1 = θ 1 σ 2, γ 2 = 0, k > 1. As we remarked in Section 1.5, the autocovariance function of a MA(1) process with parameters (σ 2, θ 1 ) is identical to one with parameters (θ1σ 2 2, θ1 1 ). That is, γ0 = θ1σ 2 2 (1 + 1/θ1) 2 = σ 2 (1 + θ1) 2 = γ 0 ρ 1 = θ 1 1 /(1 + θ 2 1 ) = θ 1/(1 + θ1 2 ) = ρ 1. In general, the MA(q) process can be written as X = θ(b)ǫ, where q q θ(z) = θ k z k = (ω k z). k=0 So the autocovariance generating function is q q g(z) = γ k z k = σ 2 θ(z)θ(z 1 ) = σ 2 (ω k z)(ω k z 1 ). (5.3) k= q Note that (ω k z)(ω k z 1 ) = ωk 2(ω 1 k z)(ω 1 k z 1 ). So g(z) is unchanged in (5.3) if (for any k such that ω k is real) we replace ω k by ω 1 k and multiply σ 2 by ωk 2. Thus (if all roots of θ(z) = 0 are real) there can be 2 q different MA(q) processes with the same autocovariance function. For identifiability, we assume that all the roots of θ(z) lie outside the unit circle in C. This is equivalent to the invertibility condition, that ǫ t can be written as a convergent power series in {X t, X t 1,...}. 5.4 The general linear process k=1 k=1 A special case of (5.1) is the general linear process, Y t = a s X t s, where {X t } is white noise. This has s=0 cov(y t, Y t+k ) = σ 2 a s a s+k σ 2 a 2 s, s=0 19 s=0

Analysis and Computation for Finance Time Series - An Introduction

ECMM703 Analysis and Computation for Finance Time Series - An Introduction Alejandra González Harrison 161 Email: mag208@exeter.ac.uk Time Series - An Introduction A time series is a sequence of observations

Time Series HILARY TERM 2010 PROF. GESINE REINERT http://www.stats.ox.ac.uk/~reinert

Time Series HILARY TERM 2010 PROF. GESINE REINERT http://www.stats.ox.ac.uk/~reinert Overview Chapter 1: What are time series? Types of data, examples, objectives. Definitions, stationarity and autocovariances.

A Multiplicative Seasonal Box-Jenkins Model to Nigerian Stock Prices

A Multiplicative Seasonal Box-Jenkins Model to Nigerian Stock Prices Ette Harrison Etuk Department of Mathematics/Computer Science, Rivers State University of Science and Technology, Nigeria Email: ettetuk@yahoo.com

ITSM-R Reference Manual

ITSM-R Reference Manual George Weigt June 5, 2015 1 Contents 1 Introduction 3 1.1 Time series analysis in a nutshell............................... 3 1.2 White Noise Variance.....................................

TIME SERIES ANALYSIS

TIME SERIES ANALYSIS L.M. BHAR AND V.K.SHARMA Indian Agricultural Statistics Research Institute Library Avenue, New Delhi-0 02 lmb@iasri.res.in. Introduction Time series (TS) data refers to observations

Discrete Time Series Analysis with ARMA Models

Discrete Time Series Analysis with ARMA Models Veronica Sitsofe Ahiati (veronica@aims.ac.za) African Institute for Mathematical Sciences (AIMS) Supervised by Tina Marquardt Munich University of Technology,

1 Short Introduction to Time Series

ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The

Financial TIme Series Analysis: Part II

Department of Mathematics and Statistics, University of Vaasa, Finland January 29 February 13, 2015 Feb 14, 2015 1 Univariate linear stochastic models: further topics Unobserved component model Signal

Univariate and Multivariate Methods PEARSON. Addison Wesley

Time Series Analysis Univariate and Multivariate Methods SECOND EDITION William W. S. Wei Department of Statistics The Fox School of Business and Management Temple University PEARSON Addison Wesley Boston

Introduction to ARMA Models

Statistics 910, #8 1 Overview Introduction to ARMA Models 1. Modeling paradigm 2. Review stationary linear processes 3. ARMA processes 4. Stationarity of ARMA processes 5. Identifiability of ARMA processes

Time Series Analysis

JUNE 2012 Time Series Analysis CONTENT A time series is a chronological sequence of observations on a particular variable. Usually the observations are taken at regular intervals (days, months, years),

Time Series Analysis 1. Lecture 8: Time Series Analysis. Time Series Analysis MIT 18.S096. Dr. Kempthorne. Fall 2013 MIT 18.S096

Lecture 8: Time Series Analysis MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis 1 Outline Time Series Analysis 1 Time Series Analysis MIT 18.S096 Time Series Analysis 2 A stochastic

Time Series - ARIMA Models. Instructor: G. William Schwert

APS 425 Fall 25 Time Series : ARIMA Models Instructor: G. William Schwert 585-275-247 schwert@schwert.ssb.rochester.edu Topics Typical time series plot Pattern recognition in auto and partial autocorrelations

Lecture 4: Seasonal Time Series, Trend Analysis & Component Model Bus 41910, Time Series Analysis, Mr. R. Tsay

Lecture 4: Seasonal Time Series, Trend Analysis & Component Model Bus 41910, Time Series Analysis, Mr. R. Tsay Business cycle plays an important role in economics. In time series analysis, business cycle

Time Series Analysis

Time Series Analysis Identifying possible ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos

Introduction to Time Series Analysis. Lecture 6.

Introduction to Time Series Analysis. Lecture 6. Peter Bartlett www.stat.berkeley.edu/ bartlett/courses/153-fall2010 Last lecture: 1. Causality 2. Invertibility 3. AR(p) models 4. ARMA(p,q) models 1 Introduction

Sales forecasting # 2

Sales forecasting # 2 Arthur Charpentier arthur.charpentier@univ-rennes1.fr 1 Agenda Qualitative and quantitative methods, a very general introduction Series decomposition Short versus long term forecasting

Time Series Analysis

Time Series Analysis Forecasting with ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos (UC3M-UPM)

Analysis of Financial Time Series with EViews

Analysis of Financial Time Series with EViews Enrico Foscolo Contents 1 Asset Returns 2 1.1 Empirical Properties of Returns................. 2 2 Heteroskedasticity and Autocorrelation 4 2.1 Testing for

TIME SERIES ANALYSIS

TIME SERIES ANALYSIS Ramasubramanian V. I.A.S.R.I., Library Avenue, New Delhi- 110 012 ram_stat@yahoo.co.in 1. Introduction A Time Series (TS) is a sequence of observations ordered in time. Mostly these

Traffic Safety Facts. Research Note. Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen

Traffic Safety Facts Research Note March 2004 DOT HS 809 718 Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen Summary This research note uses

3.1 Stationary Processes and Mean Reversion

3. Univariate Time Series Models 3.1 Stationary Processes and Mean Reversion Definition 3.1: A time series y t, t = 1,..., T is called (covariance) stationary if (1) E[y t ] = µ, for all t Cov[y t, y t

Projects in time series analysis MAP565

Projects in time series analysis MAP565 Programming : Any software can be used: Scilab, Matlab, R, Octave, Python,... The scripts must be archived in a single file (zip, tar or tgz) and should not be printed

Time Series Analysis

Time Series Analysis Autoregressive, MA and ARMA processes Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 212 Alonso and García-Martos

Univariate Time Series Analysis; ARIMA Models

Econometrics 2 Spring 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing

Time Series Analysis

Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Identification of univariate time series models, cont.:

Time Series Analysis in Economics. Klaus Neusser

Time Series Analysis in Economics Klaus Neusser May 26, 2015 Contents I Univariate Time Series Analysis 3 1 Introduction 1 1.1 Some examples.......................... 2 1.2 Formal definitions.........................

Time Series Graphs. Model ACF PACF. White Noise All zeros All zeros. AR(p) Exponential Decay P significant lags before dropping to zero

Time Series Graphs Model ACF PACF White Noise All zeros All zeros AR(p) Exponential Decay P significant lags before dropping to zero MA(q) q significant lags before dropping to zero Exponential Decay ARMA(p,q)

Univariate Time Series Analysis; ARIMA Models

Econometrics 2 Fall 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Univariate Time Series Analysis We consider a single time series, y,y 2,..., y T. We want to construct simple

State Space Time Series Analysis

State Space Time Series Analysis p. 1 State Space Time Series Analysis Siem Jan Koopman http://staff.feweb.vu.nl/koopman Department of Econometrics VU University Amsterdam Tinbergen Institute 2011 State

Some useful concepts in univariate time series analysis

Some useful concepts in univariate time series analysis Autoregressive moving average models Autocorrelation functions Model Estimation Diagnostic measure Model selection Forecasting Assumptions: 1. Non-seasonal

Trend and Seasonal Components

Chapter 2 Trend and Seasonal Components If the plot of a TS reveals an increase of the seasonal and noise fluctuations with the level of the process then some transformation may be necessary before doing

9th Russian Summer School in Information Retrieval Big Data Analytics with R

9th Russian Summer School in Information Retrieval Big Data Analytics with R Introduction to Time Series with R A. Karakitsiou A. Migdalas Industrial Logistics, ETS Institute Luleå University of Technology

Time Series Analysis in WinIDAMS

Time Series Analysis in WinIDAMS P.S. Nagpaul, New Delhi, India April 2005 1 Introduction A time series is a sequence of observations, which are ordered in time (or space). In electrical engineering literature,

Time Series Analysis

Time Series 1 April 9, 2013 Time Series Analysis This chapter presents an introduction to the branch of statistics known as time series analysis. Often the data we collect in environmental studies is collected

Introduction to Time Series Analysis. Lecture 1.

Introduction to Time Series Analysis. Lecture 1. Peter Bartlett 1. Organizational issues. 2. Objectives of time series analysis. Examples. 3. Overview of the course. 4. Time series models. 5. Time series

Introduction to time series analysis

Introduction to time series analysis Margherita Gerolimetto November 3, 2010 1 What is a time series? A time series is a collection of observations ordered following a parameter that for us is time. Examples

Rob J Hyndman. Forecasting using. 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1

Rob J Hyndman Forecasting using 11. Dynamic regression OTexts.com/fpp/9/1/ Forecasting using R 1 Outline 1 Regression with ARIMA errors 2 Example: Japanese cars 3 Using Fourier terms for seasonality 4

Estimating an ARMA Process

Statistics 910, #12 1 Overview Estimating an ARMA Process 1. Main ideas 2. Fitting autoregressions 3. Fitting with moving average components 4. Standard errors 5. Examples 6. Appendix: Simple estimators

Lecture 1: Univariate Time Series B41910: Autumn Quarter, 2008, by Mr. Ruey S. Tsay

Lecture 1: Univariate Time Series B41910: Autumn Quarter, 2008, by Mr. Ruey S. Tsay 1 Some Basic Concepts 1. Time Series: A sequence of random variables measuring certain quantity of interest over time.

Filtering Data using Frequency Domain Filters

Filtering Data using Frequency Domain Filters Wouter J. Den Haan London School of Economics c Wouter J. Den Haan August 27, 2014 Overview Intro lag operator Why frequency domain? Fourier transform Data

Ronald Christensen Advanced Linear Modeling Multivariate, Time Series, and Spatial Data; Nonparametric Regression and Response Surface Maximization Second Edition Springer Preface to the Second Edition

Analysis of algorithms of time series analysis for forecasting sales

SAINT-PETERSBURG STATE UNIVERSITY Mathematics & Mechanics Faculty Chair of Analytical Information Systems Garipov Emil Analysis of algorithms of time series analysis for forecasting sales Course Work Scientific

Calculating Interval Forecasts

Calculating Chapter 7 (Chatfield) Monika Turyna & Thomas Hrdina Department of Economics, University of Vienna Summer Term 2009 Terminology An interval forecast consists of an upper and a lower limit between

Time Series Analysis III

Lecture 12: Time Series Analysis III MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis III 1 Outline Time Series Analysis III 1 Time Series Analysis III MIT 18.S096 Time Series Analysis

Time Series Analysis of Aviation Data

Time Series Analysis of Aviation Data Dr. Richard Xie February, 2012 What is a Time Series A time series is a sequence of observations in chorological order, such as Daily closing price of stock MSFT in

Time Series in Mathematical Finance

Instituto Superior Técnico (IST, Portugal) and CEMAT cnunes@math.ist.utl.pt European Summer School in Industrial Mathematics Universidad Carlos III de Madrid July 2013 Outline The objective of this short

Time Series Analysis and Forecasting Methods for Temporal Mining of Interlinked Documents

Time Series Analysis and Forecasting Methods for Temporal Mining of Interlinked Documents Prasanna Desikan and Jaideep Srivastava Department of Computer Science University of Minnesota. @cs.umn.edu

A course in Time Series Analysis. Suhasini Subba Rao Email: suhasini.subbarao@stat.tamu.edu

A course in Time Series Analysis Suhasini Subba Rao Email: suhasini.subbarao@stat.tamu.edu August 5, 205 Contents Introduction 8. Time Series data................................. 8.. R code....................................2

C: LEVEL 800 {MASTERS OF ECONOMICS( ECONOMETRICS)}

C: LEVEL 800 {MASTERS OF ECONOMICS( ECONOMETRICS)} 1. EES 800: Econometrics I Simple linear regression and correlation analysis. Specification and estimation of a regression model. Interpretation of regression

Software Review: ITSM 2000 Professional Version 6.0.

Lee, J. & Strazicich, M.C. (2002). Software Review: ITSM 2000 Professional Version 6.0. International Journal of Forecasting, 18(3): 455-459 (June 2002). Published by Elsevier (ISSN: 0169-2070). http://0-

Graphical Tools for Exploring and Analyzing Data From ARIMA Time Series Models

Graphical Tools for Exploring and Analyzing Data From ARIMA Time Series Models William Q. Meeker Department of Statistics Iowa State University Ames, IA 50011 January 13, 2001 Abstract S-plus is a highly

Lecture 2: ARMA(p,q) models (part 3)

Lecture 2: ARMA(p,q) models (part 3) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC) Univariate time series Sept.

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.

Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium

Study & Development of Short Term Load Forecasting Models Using Stochastic Time Series Analysis

International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 9, Issue 11 (February 2014), PP. 31-36 Study & Development of Short Term Load Forecasting

Time Series Analysis and Forecasting

Time Series Analysis and Forecasting Math 667 Al Nosedal Department of Mathematics Indiana University of Pennsylvania Time Series Analysis and Forecasting p. 1/11 Introduction Many decision-making applications

15.062 Data Mining: Algorithms and Applications Matrix Math Review

.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

Chapter 7 Chapter Table of Contents OVERVIEW...193 GETTING STARTED...194 TheThreeStagesofARIMAModeling...194 IdentificationStage...194 Estimation and Diagnostic Checking Stage...... 200 Forecasting Stage...205

3. Regression & Exponential Smoothing

3. Regression & Exponential Smoothing 3.1 Forecasting a Single Time Series Two main approaches are traditionally used to model a single time series z 1, z 2,..., z n 1. Models the observation z t as a

Time Series Analysis

Time Series Analysis Time series and stochastic processes Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos

11. Time series and dynamic linear models

11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd

Time Series Analysis

Time Series Analysis Lecture Notes for 475.726 Ross Ihaka Statistics Department University of Auckland April 14, 2005 ii Contents 1 Introduction 1 1.1 Time Series.............................. 1 1.2 Stationarity

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

Chapter Vector autoregressions We begin by taking a look at the data of macroeconomics. A way to summarize the dynamics of macroeconomic data is to make use of vector autoregressions. VAR models have become

LOGNORMAL MODEL FOR STOCK PRICES

LOGNORMAL MODEL FOR STOCK PRICES MICHAEL J. SHARPE MATHEMATICS DEPARTMENT, UCSD 1. INTRODUCTION What follows is a simple but important model that will be the basis for a later study of stock prices as

Introduction to Time Series and Forecasting, Second Edition

Introduction to Time Series and Forecasting, Second Edition Peter J. Brockwell Richard A. Davis Springer Springer Texts in Statistics Advisors: George Casella Stephen Fienberg Ingram Olkin Springer New

Chapter 4: Vector Autoregressive Models

Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...

Time Series Analysis by Higher Order Crossings. Benjamin Kedem Department of Mathematics & ISR University of Maryland College Park, MD

Time Series Analysis by Higher Order Crossings Benjamin Kedem Department of Mathematics & ISR University of Maryland College Park, MD 1 Stochastic Process A stochastic or random process {Z t },, 1,0,1,,

The University of Chicago, Booth School of Business Business 41202, Spring Quarter 2013, Mr. Ruey S. Tsay Solutions to Final Exam

The University of Chicago, Booth School of Business Business 41202, Spring Quarter 2013, Mr. Ruey S. Tsay Solutions to Final Exam Problem A: (40 points) Answer briefly the following questions. 1. Describe

Maximum Likelihood Estimation of an ARMA(p,q) Model

Maximum Likelihood Estimation of an ARMA(p,q) Model Constantino Hevia The World Bank. DECRG. October 8 This note describes the Matlab function arma_mle.m that computes the maximum likelihood estimates

COMP6053 lecture: Time series analysis, autocorrelation. jn2@ecs.soton.ac.uk

COMP6053 lecture: Time series analysis, autocorrelation jn2@ecs.soton.ac.uk Time series analysis The basic idea of time series analysis is simple: given an observed sequence, how can we build a model that

4.3 Moving Average Process MA(q)

66 CHAPTER 4. STATIONARY TS MODELS 4.3 Moving Average Process MA(q) Definition 4.5. {X t } is a moving-average process of order q if X t = Z t + θ 1 Z t 1 +... + θ q Z t q, (4.9) where and θ 1,...,θ q

Forecasting of Paddy Production in Sri Lanka: A Time Series Analysis using ARIMA Model

Tropical Agricultural Research Vol. 24 (): 2-3 (22) Forecasting of Paddy Production in Sri Lanka: A Time Series Analysis using ARIMA Model V. Sivapathasundaram * and C. Bogahawatte Postgraduate Institute

2.2 Elimination of Trend and Seasonality

26 CHAPTER 2. TREND AND SEASONAL COMPONENTS 2.2 Elimination of Trend and Seasonality Here we assume that the TS model is additive and there exist both trend and seasonal components, that is X t = m t +

U.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, 2009. Notes on Algebra

U.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, 2009 Notes on Algebra These notes contain as little theory as possible, and most results are stated without proof. Any introductory

Statistics in Retail Finance. Chapter 6: Behavioural models

Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural

Senior Secondary Australian Curriculum

Senior Secondary Australian Curriculum Mathematical Methods Glossary Unit 1 Functions and graphs Asymptote A line is an asymptote to a curve if the distance between the line and the curve approaches zero

Luciano Rispoli Department of Economics, Mathematics and Statistics Birkbeck College (University of London)

Luciano Rispoli Department of Economics, Mathematics and Statistics Birkbeck College (University of London) 1 Forecasting: definition Forecasting is the process of making statements about events whose

Probability and Random Variables. Generation of random variables (r.v.)

Probability and Random Variables Method for generating random variables with a specified probability distribution function. Gaussian And Markov Processes Characterization of Stationary Random Process Linearly

Hypothesis Testing in the Classical Regression Model

LECTURE 5 Hypothesis Testing in the Classical Regression Model The Normal Distribution and the Sampling Distributions It is often appropriate to assume that the elements of the disturbance vector ε within

Probability and Statistics

CHAPTER 2: RANDOM VARIABLES AND ASSOCIATED FUNCTIONS 2b - 0 Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be

Chapter 6: Multivariate Cointegration Analysis

Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration

FOURIER SERIES BASED PHASE MODEL FOR STATISTICAL SIGNAL PROCESSING

EE 602 TERM PAPER FOURIER SERIES BASED PHASE MODEL FOR STATISTICAL SIGNAL PROCESSING - Chong yung chi Instructor: Dr. Rajesh Hegde Presented by: Mounika Bopuddi (Y8104037) Richa Tripathi (Y3288) ABSTRACT

Statistics and Probability Letters

Statistics and Probability Letters 79 (2009) 1884 1889 Contents lists available at ScienceDirect Statistics and Probability Letters journal homepage: www.elsevier.com/locate/stapro Discrete-valued ARMA

Forecasting in supply chains

1 Forecasting in supply chains Role of demand forecasting Effective transportation system or supply chain design is predicated on the availability of accurate inputs to the modeling process. One of the

1 Limiting distribution for a Markov chain

Copyright c 2009 by Karl Sigman Limiting distribution for a Markov chain In these Lecture Notes, we shall study the limiting behavior of Markov chains as time n In particular, under suitable easy-to-check

Time series Forecasting using Holt-Winters Exponential Smoothing

Time series Forecasting using Holt-Winters Exponential Smoothing Prajakta S. Kalekar(04329008) Kanwal Rekhi School of Information Technology Under the guidance of Prof. Bernard December 6, 2004 Abstract

Advanced Signal Processing and Digital Noise Reduction

Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New

Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

Is the Basis of the Stock Index Futures Markets Nonlinear?

University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Engineering and Information Sciences 2011 Is the Basis of the Stock

Matrices and Polynomials

APPENDIX 9 Matrices and Polynomials he Multiplication of Polynomials Let α(z) =α 0 +α 1 z+α 2 z 2 + α p z p and y(z) =y 0 +y 1 z+y 2 z 2 + y n z n be two polynomials of degrees p and n respectively. hen,

Chapter 10 Introduction to Time Series Analysis

Chapter 1 Introduction to Time Series Analysis A time series is a collection of observations made sequentially in time. Examples are daily mortality counts, particulate air pollution measurements, and

Factor analysis. Angela Montanari

Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number

UCL DEPARTMENT OF SECURITY AND CRIME SCIENCE Advanced time-series analysis Lisa Tompson Research Associate UCL Jill Dando Institute of Crime Science l.tompson@ucl.ac.uk Overview Fundamental principles

Monte Carlo Methods in Finance

Author: Yiyang Yang Advisor: Pr. Xiaolin Li, Pr. Zari Rachev Department of Applied Mathematics and Statistics State University of New York at Stony Brook October 2, 2012 Outline Introduction 1 Introduction

Time Series Analysis

Time Series Analysis Andrea Beccarini Center for Quantitative Economics Winter 2013/2014 Andrea Beccarini (CQE) Time Series Analysis Winter 2013/2014 1 / 143 Introduction Objectives Time series are ubiquitous

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur Module No. #01 Lecture No. #15 Special Distributions-VI Today, I am going to introduce

Stochastic Inventory Control

Chapter 3 Stochastic Inventory Control 1 In this chapter, we consider in much greater details certain dynamic inventory control problems of the type already encountered in section 1.3. In addition to the