Integrated Wavelet Denoising Method for High-Frequency Financial Data Forecasting

Size: px
Start display at page:

Download "Integrated Wavelet Denoising Method for High-Frequency Financial Data Forecasting"

Transcription

1 Integrated Wavelet Denoising Method for High-Frequency Financial Data Forecasting Edward W. Sun KEDGE Business School, France Yi-Ting Chen School of Computer Science National Chiao Tung University, Taiwan Min-Teh Yu National Chiao Tung University, Taiwan Abstract Intelligent pattern recognition imposes new challenges in high-frequency financial data mining due to its irregularities and roughness. Based on the wavelet transform for decomposing systematic patterns and noise, in this paper we propose a new integrated wavelet denoising method, named smoothness-oriented wavelet denoising algorithm (SOWDA), that optimally determines the wavelet function, maximal level of decomposition, and the threshold rule by using a smoothness score function that simultaneously detects the global and local extrema. We discuss the properties of our method and propose a new evaluation procedure to show its robustness. In addition, we apply this method both in simulation and empirical investigation. Both the simulation results based on three typical stylized features of financial data and the empirical results in analyzing high-frequency financial data from Frankfurt Stock Exchange confirm that SOWDA significantly (based on the RMSE comparison) improves the performance of classical econometric models after denoising the data with the discrete wavelet transform (DWT) and maximal overlap discrete wavelet transform (MODWT) methods. Keywords: Data denoising, DWT, High-frequency data, MODWT, Wavelet Corresponding author: Edward W. Sun, KEDGE Business School, 680 Cours de la Libèration, Talance Cedex, France. edward.sun@bem.edu. 1

2 1 Introduction The financial service sector is typically an expert at churning enormous amounts of data, gathered at the tick-by-tick level from customer interactions and transactions. Such ultra-high frequency data have a complex structure of irregularities and roughness that have been described as exhibiting multifractal phenomena - that is, different fragments of the data have different fractal properties (see Mandelbrot (1982)). The heterogeneity characterized by multifractal phenomena is caused by a large number of instantaneous changes in the markets and trading noises (see Sun et al. (2007) and references therein). Mining high-frequency data turns out to be more fundamental for financial informatics and stimulates interest in intelligently extracting information conveyed in such data. For example, McCulloch and Tsay (2004) discuss the non-linear behavior of high-frequency data and propose a price change duration model for analyzing price changes, Kelly and Steigerwald (2004) apply the stochastic volatility model to replicate features of high-frequency data in their market microstructure analysis, and Sun et al. (2009) present a high-frequency Value-at-Risk measure based on Lèvy processes. The wavelet method is one of the multifractal spectrum computing methods and is proven to be a reliable tool in econometric analysis (see, for example, Fan and Gençay (2010), Fan and Wang (2007), and Hong and Kao (2004)). Particularly, it is suitable for time series analysis, like in smoothing, denoising, and jump detection (see, for example, Gençay et al. (2010), In et al. (2011), Donoho and Johnstone (1998), and Sun and Meinl (2012), among others). Ramsey (2002) highlights some research areas where wavelet analysis might be applied in economics, and Crowley (2007) provides a survey about how wavelet methods have been used in the economics and finance literature. The advantage of the wavelet method is that it performs a multiresolution analysis - that is, it allows us to analyze the data at different scales (each one associated with a particular frequency passband) at the same time. In this way, wavelets can identify single events truncated in one frequency range as well as coherent structures across different scales. Several studies have applied wavelet methods in mining financial data. For example, Ramsey and Lampart (1998) and Kim and In (2008) apply wavelets to analyze relationships and dependencies among key macroeconomic and financial variables. Gençay et al. (2003) propose a method based on a wavelet multiscaling approach for decomposing time series data. Gençay et al. (2005) introduce a method to estimate the systematic risk of an asset based on a wavelet multiscaling approach of decomposing the underlying time series on a scale-by-scale basis. Lada and Wilson (2006) develop a wavelet-based spectral method for steady-state simulation analysis. Esteban-Bravo and Vidal-Sanz (2007) propose a wavelet-based method for solving boundary value problems in growth models. Jensen (2007) implements compactly supported wavelets to develop an estimator of the long memory process. Laukaitis (2008) applies wavelet transforms for high-frequency data denoising in the study of credit card intraday cash flow and intensity of transactions. Gençay and Gradojevic (2011) introduce a wavelet approach to estimate the parameters of a linear regression 2

3 model. Sun et al. (2011) propose a wavelet method for analyzing the currency market with high-frequency data. Aguiar-Conraria et al. (2012) apply wavelet tools in macroeconomics data analysis. Haven et al. (2012) show the efficiency of the wavelet method in denoising option price data. A classic assumption for data mining is that the data are generated by certain systematic patterns plus random noise. Denoising high-frequency data provides a fundamental tool to extract the systematic patterns conveyed in the data. As Sun and Meinl (2012) point out a specific problem arises when the trend component exhibits occasional jumps that are in contrast to the slow evolving long-term trend. These occasional jumps are often caused by, for example, unexpected large transactions or extreme prices and should not be contributed to the normal short-term variations (since jumps are often considered as noises), but indeed to the long-run trend. Traditional linear denoising methods (e.g., moving average) usually fail to capture this information accurately as these linear methods tend to blur out jumps, while non-linear filters are not appropriate to smooth out high-frequency fluctuations sufficiently since the trends extracted by these methods are not smooth enough (i.e., usually with kicks) to present long-run dynamic information (see Sun and Meinl (2012) and references therein). Several works have been done to overcome the above-mentioned problem. For example, Connor and Rossiter (2005) estimate the wavelet variance by using non-decimated wavelet transforms. Studies of applying wavelets in denoising data and coefficient construction can be found in Gençay et al. (2002), Keinert (2004), Mallat and Hwang (1992), and Percival and Walden (2006), among others. Among these methods, both DWT and MODWT need to decide the wavelet function, level of decomposition, and thresholding rule. A common approach in choosing the wavelet function is to use the shortest wavelet filter that can provide reasonable results (see Percival and Walden (2006)). The level of decomposition leads to the choice that considers the higher the better" in general. The thresholding rule, which is a function identifying the wavelet coefficients to be deleted, has also been investigated in academic research (see, for example, Gençay et al. (2002)). Meinl and Sun (2012) propose the local linear scaling approximation (LLSA) method in denoising high-frequency data and show its robustness in empirical application under statistical goodness-of-fit tests. However, LLSA focuses on the linear characteristics around jumps for the re-constructed wavelet coefficients, but only takes advantage of it when the wavelet function is pre-determined. The remaining challenge is how to determine the combination of the wavelet function, level of decomposition, and thresholding rule to reach an optimal smoothness that generally improves the performance of classic models after denoising the data. The algorithm (named SOWDA) proposed in this paper can optimally determine the wavelet function, level of decomposition, and thresholding rule by using smoothness as a regularization variable. The goal of our method is to denoise the data and obtain the trend that: (1) contains as much information as possible, (2) exhibits a certain degree of smoothness that can utilize the classic model, and (3) preserves as few artifacts (i.e., undesired structures, like oscillating nature, 3

4 generated through the denoising process) as possible. In our method, we define the measures for smoothness. Intuitively, these measures for smoothness describe the characteristics of the denoised data with wavelet transform that optimally provide output for further analysis with the classical model. We show that the resulting difference sequence between the denoised data and the original signal must converge in probability at a predetermined confidence level. This requires that: (1) the structural change (e.g., jumps) of the denoised data and the original signal should be synchronous, (2) there are no outliers in the denoised data, and (3) the local extremum in the denoised data should be bounded. We show the analytical properties of SOWDA that confirm the proposed method can lead to an optimal solution to satisfy these requirements. Therefore, SOWDA can be used to improve the performance of econometric models that are parametrically built on the i.i.d. white noise assumption by denoising the unexpected outliers. In addition, we propose a new performance evaluation method based on the jump detection test suggested by Xue et al. (2014). Through this procedure, we verify the robustness of SOWDA in its performance. We investigate the performance of SOWDA with numerical simulations that consider some typical patterns often observed in high-frequency financial data, e.g., excessive volatility and regime switching. With a comparison, our method results in a better performance than the alternative methods, the numerical results confirm the analytical properties of SOWDA that show the proposed algorithm maintains the original wavelet transform s computational complexity, and its approximation errors are bounded. In order to confirm the computational reliability and consistency that we have shown in the simulations, we further perform an empirical investigation by applying our algorithm with high-frequency data (5-minute data) of DAX 30 stocks. In the empirical study, we work both on the in-sample model fitting and out-of-sample (one-step ahead and two-step ahead) forecasting. The results we obtain from such a large sample investigation coincide with the previous simulation results. When using the denoised data generated by our algorithm for forecasting with the high-frequency data, we find that the performances (i.e., accuracy of forecasting) of the classic models e.g., AR, ARMA, and ARMA-GARCH, significantly improve (based on the RMSE comparison) and confirm the efficiency of the proposed algorithm. We organized the paper as follows. We describe the methodology in details and summarize it with an algorithm chart in Section 2. In Section 3, we show the implementation of SOWDA and its analytical properties with respect to the jumps in high frequency data and proposes a new performance measure procedure based on the jump detection test. Section 4 investigates the performance of SOWDA by conducting simulations. The simulation results confirm the superior performance of our method. In Section 5 we execute an empirical study by applying our method to analyze the high-frequency data collected from Frankfurt Stock Exchange (i.e., the DAX 30 stocks), and illustrate both in-sample modeling and out-of-sample forecasting results. We summarize our conclusions in Section 6. 4

5 2 The Methodology 2.1 Wavelets and wavelet transforms for denoising Wavelets are bases of L 2 (R) first developed to analyze geophysical signals (see Morlet (1983)). In contrast to the Fourier transformation they enable a localized time-frequency analysis with wavelet functions that usually have either compact support or decay exponentially fast to zero. In addition, wavelets provide a multiresolution of a signal that allows for analyzing the signal simultaneously on different (usually dyadic) scales. To achieve these desirable properties, the wavelet transformation is subject to the following conditions, i.e., the admissibility condition and orthonormality: ψ(ω) 2 dω <, ω ψ 2 (u) du = 1 and l=0 l=0 l=0 ψ(u) du = 0. The latter condition signifies that ψ oscillates around zero. An analog discrete formulation is developed by introducing discrete wavelet filter banks that correspond to a linear moving weighted average high-pass filter (see Mallat (1989)). For this, equivalent discrete oscillation and orthonormality conditions must hold: L 1 L 1 L 1 h l = 0, h 2 l = 1 and h l h l+2n = 0 for all n N, where h l = (h 0, h 1,, h L 1 ) is a finite length discrete wavelet filter. We note that these conditions are only necessary, but not sufficient to construct reasonable high-pass wavelet filters, requiring additional regularity conditions. Daubechies (1992) provides a thorough treatment of this topic. In our paper we focus on the application of wavelets and only consider the most commonly used wavelet functions in financial data analysis, e.g., Daubechies (D4) and the least asymmetric (LA) wavelet, which differ in their constructional regularity conditions and are the ones in their class with the smallest support (see Percival and Walden (2006) and Sun and Meinl (2012)). A discrete wavelet transform (DWT) is an orthogonal transform of a vector (discrete signal or time series data) X of length N (which must be a multiple of 2 J ) into J wavelet coefficient vectors W j R N/2j, 1 j J and one scaling coefficient vector V J R N/2J : [W 1,..., W J, V J ] = WX. (1) With the transformation matrix W determined by the wavelet filter banks h, it is more convenient to use the pyramid algorithm developed by Mallat (1989), with a computational complexity of O(N). Applying the inverse transform on these vectors yields: [W 1,..., W J, V J ]W = S J + 5 J D j = X, (2) j=1

6 that is, an additive decomposition of the original signal. When S J 1 = S J + J D j (3) holds, S j, approximations (i. e., moving weighted averages) of X at different dyadic scales, and D j, the detail vectors that are the details we lose at each approximation level, result in a multiscale decomposition. For each scale j we can separate the signal into high and low frequencies by a wavelet filter with the bandwidth determined by j. Maximal overlap discrete wavelet transform (MODWT) is an extension of the traditional DWT, which differs from DWT in the sense that all vectors are in R N and have a higher computational complexity of O(N log 2 N) (see, for example, Percival and Walden (2006) and references therein). For all discrete wavelet transforms in this paper we utilize circular filtering extensions at the borders, that is, for data points X 0, X 1,... and X N+1, X N+2,..., required for the moving weighted average, we substitute them by X N, X N 1,... with X 1, X 2,..., respectively. Percival and Walden (2006) provide a discussion on boundary extensions and distortions, as well as methods to deal with them. Wavelet denoising might result in different outcomes due to different choice of input variables. As we have pointed out, wavelets are oscillating functions and transform the signal orthogonally in the Hilbert space. Consequently, we do not have any trouble from reconstruction based on the wavelet transform. However, there are many wavelet functions that meet the requirement for the transform. In practice, different wavelets are used for different reasons. For example, the Haar wavelet has very small support, and wavelets of higher orders such as Daubechies (D4) and least asymmetric (LA) have bigger support. Bigger support can ensure a smoother shape of S j, 1 j J with each wavelet, and the scaling coefficient can carry more information due to the increased filter width. The optimal band-pass filter, which dictates the capacity to isolate features to specific frequency intervals is determined by the length of a wavelet function in approximation. In addition, the wavelet function must be able to mimic the features contained in the signal of interest in order to optimally represent the conveyed information. Therefore, the choice of the wavelet basis function turns out to be important when analyzing a given signal. After choosing the wavelet, we then decompose the signal into several levels. In general, we use the pyramid algorithm to accomplish this (see Gençay et al. (2002)). This algorithm decomposes the signal into detail and approximation coefficients in its first iteration. Each following iteration applies the same procedure on the approximation coefficients from the iteration one step ahead. 1 We can do this at a maximum log 2 (N) times (N is the number of observations). The quality of the denoising process varies with the number of iterations. The resulting outcome also critically depends on the thresholding rule that decides all wavelet coefficients less than a fixed constant in magnitude are zero. The coefficients obtained in each iteration are subject to the thresholding 1 For DWT, we down sample the coefficients, i. e., in each iteration we halve the number of detail coefficients. j= J 6

7 rule before we reconstruct the denoised signal. This rule identifies the coefficients that represent noise. We thus have three factors that influence the quality of denoising based on the wavelet transform: wavelet function (or mother wavelet), number of maximal iterations (or level of decomposition), and the thresholding rule. However, there is no straightforward method to determine these three factors simultaneously. In this paper we propose a new method that aims to optimally determine the choice of wavelet function, number of maximal iterations, and thresholding (we refer to the choice of them as a combination of denoising factors) simultaneously based on the smoothness-oriented criterion. 2.2 Smoothness-oriented wavelet denoising algorithm (SOWDA) In this section we introduce the method that helps us to decide the combination of denoising factors. Assume that the observed data X can be decomposed as follows: X t = S t + N t, where S t is the true signal and S t is its estimation. Here, N t is the additive noise sampled at time t. In order to evaluate the denoising performance, i.e., to see how close S t is toward S t, we define the smoothness properties as follows. Definition 1. Let x n be a random variable and x n = S t S t. If there exist constants c and ε, then the smoothness is, when ε > 0: lim Pr ( x n c > ε ) = 0. n This definition of smoothness intuitively states that the sequence x n of the difference between S t and S t becomes close to a controllable constant c. In other words, the resulting difference sequence between S t and S t converges in probability to c. This requires (i) the structural change (e.g., jumps) between S t and S t is synchronous, (ii) there are no outliers in x n, and (iii) the local extremum in x n is bounded and leads to the following measures. Definition 2. Let (Y 1, Y 2 ) T be a vector of continuous random variables with marginal distribution functions F 1, F 2, then the coefficient η H is: ( ) η H (u) = lim P Y 2 > F2 1 (u) Y 1 > F1 1 (u), (4) u 1 and the coefficient η L is ( ) η L (u) = lim P Y 2 < F2 1 (u) Y 1 < F1 1 (u). (5) u 0 When ε > 0, u 0, u 0 > u we have: η H (u) η L (u) 1 < ε. We then say Y 1 and Y 2 are synchronous, that is, η H (u) η L (u). 7

8 When η H > 0, there exists upper tail synchronicity and the positive extreme values in Y 1 and Y 2 can be observed simultaneously. When η L > 0, there exists lower tail synchronicity, and the negative extreme values can be observed simultaneously. We further require the smoothness measure to be able to detect artifacts and jumps. We suggest two different measures here: one considers artifacts (τ 1, based on an outliers test) and the other considers jumps (τ 2, based on local extrema). In other words, we use τ 1 to detect the global extrema and τ 2 the local extrema. Both of them have the ability to detect boundary problems, that is, inefficient approximation at the beginning and end of the signal. We suggest to apply Grubbs test for identifying artifacts, which is an iterative test for outliers based on an approximately normal distributed sample (see Grubbs (1969)). Let µ = 1 T T 1 X t be the sample mean of vector X t and s 2 = 1 T 1 T 1 (X t µ) 2 its sample variance. The test statistics are then given by G = max X i µ. s Here, G can be assumed to be t-distributed, and a test for outliers with significant level α (e.g. α = 0.05) can easily be performed by rejecting the null hypothesis of no outliers if: G > z α = T 1 t2 α,t 2 2 T T T 2 + t 2 α 2 T,T 2. When an outlier (i.e., the global extremum) is detected, it is removed from the data and the test then proceeds. As a measure of the amount of artifacts (or jumps of high magnitude), we can identify the number of iterations to run the test until it confirms there is no outlier. We apply the test until g(x) = 0 and count the number of outliers as a measure of structure. Definition 3. Let C(x) be a function determining whether there is one outlier in vector X: { 1, if G > z α C(x) = 0, if otherwise We define τ 1 as: where n is the sample size. τ 1 = n 1 1 C(x)=1, (6) i=1 In order to control all structural changes to be bounded, our proposed method investigates the local extrema (maxima or minima, respectively) at a certain magnitude. In order to avoid redundant computation (since τ 1 controls the outlier detection), we only run the test procedure for the output data after the wavelet transform. The local extrema referred to here are the largest and smallest values that a function takes at a point within a given neighborhood. 8

9 Definition 4. If there exists a Λ and Λ R, so that lim sup x n = Λ, n N and there exists a subsequence x kn of x n for which we have that x kn < Λ, n, then Λ is the local maxima. If there exists a λ and λ R, so that lim inf n N x n = λ, and there exists a subsequence x kn of x n for which we have that x kn > λ, n, then λ is the local minima. Definition 5. Let D(x) be a function that detects local maxima { 1, if x i Λ D(x) = 0, if otherwise, and D (x) detects local minima. D (x) = { 1, if x i λ 0, if otherwise. We define τ 2 as: where n is the sample size. τ 2 = n 1 1 D(x)=1 + i=1 n 1 1 D (x)=1, (7) i=1 The algorithm in this paper applies a linear score function T ( ) of τ 1 and τ 2 (i.e., a linear combination of measures for global and local extrema) to compute the overall score: T (τ 1, τ 2 ) = α τ 1 + β τ 2, (8) where α and β are the weights assigned to the score function, e.g., α + β = 1, which can be determined based on the data characteristics. The combination of denoising factors with the lowest value of T (τ 1, τ 2 ) is preferred and identified as the optimal set for the wavelet-based denoising process Summary of SOWDA algorithm Given a set of wavelet functions (f ), maximal levels of decomposition (l ), and thresholding rules (s ) as denoising factors, and an input data vector X t as input variables, SOWDA determines the output of the wavelet transform. We define a score function T (τ 1, τ 2 ) as the criterion to determine the combination of denoising factors. The function min{t (τ 1, τ 2 )} will lead to the optimal combination of denoising factors (f, l and s ) when its score value is minimized. The pseudo code for SOWDA is summarized in the following Algorithm Chart. 9

10 Algorithm 1 Smoothness-oriented wavelet denoising algorithm (SOWDA) Require: X: input data vector; F : a set of wavelet functions; L: a set of maximal levels of decomposition; S: a set of thresholding rules; W: wavelet transform {DWT, MODWT}; C : F L S W; Ensure: optimal f, l, s ; ˆX = W(f, l, s ) 1: for C c do 2: Conduct wavelet transform. 3: Apply threshold rule to wavelet coefficients. 4: Conduct wavelet inverse transform. 5: Compute evaluation function and return T (c). 6: end for 7: (W, f, l, s ) = c = argmin T (c). 8: Apply W and using (f, l, s ) to extract trend of X, ˆX. 3 Implementation of SOWDA 3.1 Jumps in high-frequency data A substantial amount of significant discontinuity has been observed in financial data and these discontinuities are commonly called jumps. Several empirical and theoretical studies have discussed the existence of jumps and their substantial impact on asset pricing, portfolio and risk management, and hedging (see Lee and Mykland (2008)). Jumps come to markets irregularly and their arrivals and amplitudes depend on market information that leads to the roughness of the financial time series data, particularly for high-frequency data (see Sun et al. (2007)). We need to detect jumps with robust tools, such as the wavelet method (see Sun and Meinl (2012) and Xue et al. (2014)), and then clarify which information is dynamically related to jumps in order to discover market phenomena and improve modeling, forecasting, pricing, and hedging. In finance, a one-dimensional asset return process with a fixed complete probability space (Ω, F t, P), where {F t : t [0, T ]} is a right-continuous information filtration for market participants, and P is a data-generating measure. Letting S(t) be the continuously compounded return of the asset at t under P, S(t) can be expressed as the Itô drift-diffusion process that satisfies the stochastic differential equation: ds(t) = µ(t)dt + σ(t)db(t), where B(t), µ(t), and σ(t) are the F t -adapted standard Brownian motion, the drift, and the spot volatility, respectively. When there are jumps: ds(t) = µ(t)dt + σ(t)db(t) + λ(t)dj(t), where J(t) is a counting process that is independent of W (t), and λ(t) is the jump size that is independent and identically distributed. 10

11 Lee and Mykland (2008) impose two necessary assumptions on price processes given as follows. For any ɛ > 0, 0 = t 0 < t 1,..., < t n = T : and sup t sup µ(u) µ(t i ) = O p ( t 1 2 ɛ ), t i u t i+1 sup t sup σ(u) σ(t i ) = O p ( t 1 2 ɛ ). t i u t i+1 The notation O p is used for random vector {X n } and non-negative random variable {x n }, X n = O p (x n ), whereby if for any ɛ > 0 there exists a finite constant C ɛ such that P ( X n > C ɛ x n ) < ɛ. These two assumptions show the drift and diffusion coefficients do not change dramatically over a finer time interval and the maximum changes in mean and spot volatility in a given interval is bounded (see Lee and Mykland (2008)). Several jump detecting methods based on wavelet transform have been proposed (see, for example, Sun and Meinl (2012) and Xue et al. (2014)). 3.2 Property of SOWDA The SOWDA algorithm we propose in this paper is exactly trying to remove the jumps without jeopardizing the information possessed in the original data. Let (F, L) be fixed, and we can directly obtain S SOWDA J = S DWT J and S SOWDA J = S MODWT J. For any other choice of (F, L, S), more details will be added to the estimator SJ SOWDA, according to T (τ 1, τ 2 ). After the reconstruction (see Equation (3)), these additional details correspond to the estimators S DWT J and SJ MODWT. For the refinement process, several wavelet coefficients are is available, as this procedure causes the set to zero by S, and no exact description of SJ SOWDA information carried by the wavelet coefficients at different levels to be intermixed. However, noting that either the DWT or MODWT approximation of any level is bounded by the signal itself, we can set up an ε-tube around the initially estimated trend by ε := max X t min X t. (9) t t We then estimate an upper bound for the error by a given constant c and c = N ε, which suffices the following equation: N t=1 ϑ t (X) S GOWDA J, t < c, (10) where X is the signal of length N and ϑ(x) is its denoised approximation. As we have: min X t SJ, SOWDA t max X t, and min X t SJ, MODWT t max X t, t t t t for all 1 j J, it follows that: ϑ t (X) S DWT J, t ε, and ϑ t (X) S MODWT J, t ε. 11

12 This must also hold for SOWDA, such that: min X t S SOWDA t J, t max t X t, and ϑ t (X) S SOWDA J, t ε. Assuming for fixed F and L, the wavelet coefficients Ω W j, k are ordered in time, that is, ω j, k+1 ω j, k, for all 1 j J and 1 k K 1. With t K := max t, j {t ΩS j, K} for all t > t K and the same ε as in Equation (9), we then obtain: N t=1 µ ( ϑ t (X) ) µ ( S SOWDA J, t ) = t K t=1 µ ( ϑ t (X) ) µ ( ) SJ, SOWDA t tk ε, and N t=1 σ ( ϑ t (X) ) σ ( S SOWDA J, t ) = t K t=1 σ ( ϑ t (X) ) σ ( ) SJ, SOWDA t tk ε. 3.3 Performance measure with jump detection Wavelet transforms have the ability to decompose the data into high-frequency components and low-frequency components. Since jumps are unexpected and instantaneous, Xue et al. (2014) indicate that jump dynamics should be contained in the high-frequency components of prices. They propose a non-parametric jump detection method, which estimates the high-frequency components of instantaneous volatility. Let the logarithm of an asset s market price be P t = log S t where S t is the asset price at time t. P 1,t are the wavelet coefficients of P t = log(s t ) for the first scale MODWT. A test statistic defined as J w, which detects jump occurrences at time t i for i = 3,..., T, is given as follows: J w (i) = P 1,t i ˆ σ ti 1 where ˆσ t 2 i 1 = 1 i 1 i 2 k=2 P i,t k P i,tk 1. When the test statistics exceeds the threshold, the null hypothesis of jumps in this interval is rejected. In this paper, we setup the threshold level as 1% of the null distribution. We define a jump as being detected when the test statistics exceeds the threshold level. We have learned that SOWDA is capable of identifying jumps and reconstructing them to preserve information, i.e., classify the spurious jumps and actual jumps. We propose a performance evaluation procedure for SOWDA based on the jump detection test (XGF test in short) introduced by Xue et al. (2014). The idea behind this procedure is straightforward. The identified jumps must be marked at the same location in the original data and denoised data. Our procedure is given as follows. 12

13 Step 1. Applying SOWDA to the original time series, then we separate the original data to the trend series and noise series. Step 2. As Xue et al. (2014) assume J w (i) following a normal distribution, we set the threshold level at 1% for testing the null hypothesis that there is no jump. Step 3. We run the XGF test following a non-overlapping moving window with length n to the original, the trend, and the noise. We mark 1 when the XGF test rejects the null hypothesis; otherwise 0 is marked. We then have three test series containing only 0 and 1 elements with a length of N/n, where N is the total data length and n << N. Step 4. We compute the difference series (e.g., the test series of the original minus the test series of the trend) of three series we obtained by Step 3. We next take the absolute value of them. Step 5. We run a T test to check if the mean of the absolute difference series obtained by Step 4 is different from 1. We reject the null hypothesis (i.e., equal to 1 ) with statistical significance when the jumps occur simultaneously in the two series investigated. Following Xue et al. (2014), we conduct this procedure respectively for positive jumps and negative jumps. If SOWDA efficiently decomposes the original time series, then we reject the null hypothesis for testing the difference series of original data and trend data, but cannot reject the null hypothesis for the difference series of the original data and noise data. 4 Simulation Study We conduct a simulation study to investigate the performance of the proposed algorithm. The purpose of this simulation study is two-fold. First, we show that for any arbitrary signal the proposed method will result in a better performance than the non-optimized method (i.e., arbitrarily determining the union of wavelet, level of decomposition, and the thresholding rule). Second, we shall illustrate the properties of our algorithm by, particularly, showing the consistency of our algorithm - that is, the error generated by our method is bounded and less than those of the non-optimized method. 4.1 The data In this study we perform the Monte Carlo simulations where errors (jumps) are generated from three different patterns to describe (1) moderate volatility, (2) excessive volatility, and (3) excessive volatility with mean level shifts. We create time series data of length This trend is based on a sine function, whose amplitude and frequency are drawn from a uniform distribution. For generating the pattern (1) signals, following the simulation introduced by Sun and Meinl (2012), we add jumps to this trend. Jump occurrences are uniformly distributed (coinciding 13

14 with a Poisson arrival rate observed in many systems), with the jump heights being a random number drawn from a normal distribution with mean 0 and variance 1. The signal is constant between the jumps. White Gaussian noise is added to the signal afterwards. For pattern (2) signals, we repeat the method used for simulating pattern (1), but replace the Gaussian noise with skewed contaminated normal noise, which has heavy tails to capture the excessive volatility (see Chun et al. (2012)). For pattern (3) signals, we repeat the method used for pattern (2) signals, but shift the trend up and down once in order to generate a signal characterized as excessive volatility with mean level shifts. The amplitude of the shift is four times the previous trend. Figure 1 illustrates the Q-Q plots of these three different signals. 4.2 The methodology In our simulation study we choose Haar, Daubechies (DB), Symlet (LA) and Coiflets (Coif) as wavelet functions (see Percival and Walden (2006)). We apply the pyramid algorithm in our empirical study. With each iteration of the pyramid algorithm, we increase the scaling level, that is, given a signal of length N = 2 J, the j-th iteration computes detail coefficients associated with changes on a scale of length λ j = 2 j 1 (see Gençay et al. (2002)). In this simulation we consider several thresholding rules. Donoho and Johnstone (1994) suggest the universal thresholding rule based on the following equation: ϑ U = ˆσ ɛ 2 log n. The idea behind this selection rule is that, for a sequence of n independently and identical distributed (i.i.d.) N(0, σ 2 ɛ ) random variables, it holds that the probability that the largest value in absolute terms is smaller than ˆσ ɛ 2 log n converges to one for large n. Donoho and Johnstone (1998) suggest the minimax. Donoho (1994) proposes a method based on minimizing Steins unbiased estimate of risk (i.e., SURE). Suppose there is a sequence of k i.i.d. random variables z i N(µ i, 1). Let ˆµ be an estimator of µ = (µ 1,..., µ k ). Given a weakly differentiable function g, an unbiased estimator of µ, ˆµ = z + g(z) is given by: where g(z) = i E ˆµ µ 2 = k + [E g(z) g(z)], i g i (z). With soft thresholding we obtain: SURE(z, ϑ) = k 2 #(i : z i ϑ + k min 2 ( z i, ϑ). Here, #S equals the cardinality of a given set S. The SURE thresholding is the one minimizing the estimated risk: i=1 ϑ S = arg min ϑ 0 SURE(z, ϑ). Donoho and Johnstone (1995) suggest the heuristic thresholding that applies the SURE thresholding rule to some levels of decomposition and universal thresholding to others. The 14

15 decision for which rule will be used on which level is made heuristically. Birgé and Massart (1998) suggest a thresholding rule based on the Birgé-Massart strategy using a penalized projection estimator (PPE). For each level i, q i is calculated as: q i = m (j + 2 i) α, where j is the maximal level of decomposition, m is a constant proposed to equal the length of the data (i.e., the number of observations), and α is a controlling constant. On each level i the q i largest coefficients are kept. The larger the α value is, the more coefficients that remain. A typical choice for the α value is 1.5 for compression and 3 for denoising. Based on the above-mentioned thresholding rules, wavelet coefficients are thresholded term by term on the basis of their individual magnitudes. Information on other coefficients has no influence on the treatment of particular coefficients. Cai and Silverman (2001) propose a block thresholding method, which is a shrinkage method capturing information on neighboring coefficients. When applying the block thresholding rule, the coefficients are considered in overlapping blocks and the treatment of coefficients in the middle of each block depends on the data in the whole block. The candidates for denoising factors of SOWDA in our simulation are: F { Haar, DB(2), DB(4), DB(8), LA(2), LA(4), LA(8), Coif(4), Coif(6), Coif(8)}; L {i : i = 1, 2, 3}; S { Block, Universal, SURE, heuristic Sure, Minimax, Birgé-Massart}. The linear score function T ( ) of τ 1 and τ 2 has the following form: T (τ 1, τ 2 ) = 0.5 τ τ 2. When we compute τ 1, we set α = 0.05 for the Grubbs t-statistic. When we compute τ 2, the one-sigma rule is applied to detect the local extrema - that is, an observation is considered to be the local extrema (of a given sequence) if it lies in the region at a distance from its mathematical expectation of more than the standard deviation. In this simulation the alternative methods we compare with SOWDA are five single wavelet functions, i.e., Haar, DB(4), DB(8), LA(8), and Coif(6), that work with both DWT and MODWT. We run the simulation for the three different data patterns described in Section 3.1. We use our algorithm to identify the best denoising method that optimally combines the wavelet function, level of decomposition, and the thresholding rule. For each pattern, we conduct the simulation based on a moving window design illustrated with Figure 2. We investigate our 15

16 algorithm for both an in-sample approximation and out-of-sample forecasting. For the out-ofsample forecasting, we work with one-step and two-step forecasting. Since the true trend (for both in-sample and out-of-sample) of the simulated stylized data is known, we then use SOWDA and alternative methods to denoise the simulated data and compare the approximated trend and forecasted trend with their true counterparts. Obviously, the smaller the difference is compared with the true trend, the better the performance will be by the underlying algorithm. 4.3 Simulation results As we mentioned in Section 3.1, the data length is 8,192 (2 13 ) for each pattern. For the moving window design, we set the in-sample size as 1,000 and the out-of-sample size as 10 for both one-step ahead and two-step ahead forecasting. The number of window moves is then 720, and we generate 100 data series for each pattern. Therefore, for each pattern we test our algorithm 72,000 times for in-sample approximation, one-step forecasting (validation), and two-step forecasting. In our simulation, we have 216,000 runs in total. For each run we compute the root mean squared error (RMSE). We report the mean value of RMSE and its corresponding variance (in parenthesis) for each method in Table 1. The smaller the MSE mean value is, the better the denoising performance. For DWT, we find that the RMSE mean values of SOWDA are smaller than that of other alternatives for both in-sample and out-of-sample performances. We note that the proposed SOWDA performs better than the non-optimized alternative methods. In order to illustrate the result reported in Table 1, we show some results (the best three) by Figures 3 and 4. We see that the mean values of RMSE of SOWDA are smaller than that of the alternative methods and the variance of RMSE for SOWDA turns out to be smaller than that obtained by using alternative methods. Additionally, we identify that when increasing the number of simulation runs, the variance of RMSE decreases. The speed of the decrease in variance (i.e., the speed of error convergence to its limit) of SOWDA is relatively faster than that of alternative methods. This result confirms the consistency property of SOWDA that we have analytically shown in Section The results we obtain in this simulation conclude that SOWDA shows better performance than alternative methods. 5 Empirical Study In this empirical study we investigate the performance of our proposed algorithm (SOWDA) by analyzing high-frequency data of German DAX 30 component stocks. We want to see if SOWDA can be used as a denoising method to improve the performance of some classic econometric models (i.e., AR, ARMA, and ARMA-GARCH) for in-sample estimation and out-of-sample forecasting. 16

17 5.1 The data The analysis is performed on high-frequency German DAX 30 component stock prices from January to December We aggregate the tick-by-tick data to the homogeneous (i. e., equally spaced) time series data at the 5-minute level by using the linear interpolation method - that is, the inhomogeneous series where times t i are given by x(t i ), while the target homogeneous time series x shall be defined at times τ j := t 0 + j t, j N, with t > 0 fixed. Every regular τ j is bounded by two times of irregularly spaced series, i.e.: t Ij τ j < t Ij +1, with I j := max{i t i τ j }, and data point τ j is interpolated between t Ij and t Ij +1 by: x(τ j ) = x Ij + τ j t Ij t Ij +1 t Ij (x Ij +1 x Ij ). In our sample, the 5-minute data contains 26,686 data points for each DAX stock in Methodology and results In this empirical study we have two experiments: the in-sample modeling and out-of-sample forecasting. For the in-sample training experiment, we first apply the smoothness-oriented wavelet denoising algorithm (SOWDA) proposed in this paper (see Algorithm chart 1 in Section 2.2.1) to denoise the data. We then use both the original data and denoised data to estimate the AR(2), ARMA(2,1), and ARMA(2,1)-GARCH(1,1) models with the maximum likelihood approach. We compare the performance of model fitting by using the root mean squared error (RMSE) for each model as the goodness-of-fit measure. For the out-of-sample forecasting, we use the established model to conduct one-step-ahead and two-step ahead forecasting following the method suggested by Sun and Meinl (2012). We then evaluate the forecasting performance by computing the root mean square error (RMSE) of the forecasted values compared with the observed values in the original data. We use the same SOWDA setting and moving window design as in the simulation. For the moving window design, we set the in-sample size as one month and the out-of-sample size as one day for both one-step ahead and two-step ahead forecasting. Table 2 reports the in-sample results for AR(2), ARMA(2,1), and ARMA(2,1)-GARCH(1,1). We see that the mean, median, and variance of RMSE are all reduced after applying SOWDA denoising, showing a significant improvement of denoised data (for both DWT and MODWT SOWDA) over the original data for model fitting. Table 3 reports the results of out-of-sample forecasting for AR(2), ARMA(2,1), and ARMA(2,1)- GARCH(1,1) based on SOWDA DWT denoising. Table 4 shows the forecasting results based on 17

18 SOWDA MODWT denoising. For the out-of-sample forecasting results, we also identify a generally significant improvement provided by SOWDA denoised data versus the original data. The empirical results coincide with our previous simulation results when considering the robustness of SOWDA s denoising performance. We next conduct the performance evaluation proposed in Section 3.3. If SOWDA efficiently decomposes the trend and noise from the original data, then all the jumps possessed by the original data, the trend, and the noise should coincide - that is, the null hypothesis that the mean of the absolute difference series equals one should be rejected with the T test. In our assessment, we consider not only the overall jumps, but also the negative and positive jumps individually. We report our results in Table 5. From Table 5, we note that the null hypothesis that the mean of the absolute difference series equals one is rejected with significant T test statistics, illustrating that the jumps occur simultaneously in the original data, the trend, and the noise obtained with SOWDA denoising. Our result implies that SOWDA performs efficiently in the decomposition of the original data based on the jump detection test. 6 Conclusion When conducting econometric analysis of data with classic parametric models built on i.i.d. white noise assumption, once one encounters unexpectedly extreme observations, e.g., outliers, the performance of such a parametric model will be distorted. To overcome the challenging disturbance of outliers, one can usually either find a more robust way to build the model or remove the outliers. The former might complicate the situation when the outlier s pattern is unknown, while the latter could remove some information if such outliers are closely related to the underlying true dynamics. In order to optimally detect and remove the real noise from the underlying dynamics, wavelet methods have been suggested to denoise the data not only in economics, but also in other areas, such as electronic signal processing. Applying the wavelet denoising method requires that the wavelet function, level of decomposition, and thresholding rule be determined. We call it trinity of wavelet denoising operation. Identifying the optimal combination of wavelet function, level of decomposition, and thresholding rule challenges the efficiency of the classical methods of denoising data based on the wavelet transform (e.g., DWT and MODWT). An inefficient decomposition of the systematic pattern (the trend) and noise of the target data will tremendously reduce the efficiency and effectiveness of any decision support system. When working with highfrequency financial data, their irregularities and roughness reinforce the necessity of more efficient tools for data mining. In this paper we propose a new denoising method for high-frequency financial data, named the smoothness-oriented wavelet denoising algorithm (SOWDA), which optimally determines the combination of denoising factors (i.e., wavelet function, level of decomposition, and thresholding 18

19 rule) based on a smoothness-oriented score function that is designed for detecting global and local extremum. The method can be applied with the classic DWT or MODWT approach. When applying it to high-frequency financial data, this algorithm is able to preserve a smooth trend and effectively displays the noise since all information can be contained in the wavelet multiresolution decomposition optimally after specifying the level of smoothness in reconstructing wavelet coefficients. In this paper we analytically show SOWDA s properties and propose a new performance evaluation procedure based on the jump detection test. In order to show the efficient performance of SOWDA, we first conduct an experiment based on simulations. We consider three different stylized data patterns that are often observed in high-frequency financial data, such as heavy tails and regime switching. In all simulation settings that we have investigated, SOWDA illustrates its robustness independent of the input data. We then empirically show the potential application of SOWDA by fitting and forecasting real high-frequency financial data (aggregated at the 5-minute-frequency level) from German DAX 30 component stock prices with three classic econometric models. The results confirm our conclusion that SOWDA significantly (based on the RMSE comparison) improves the efficiency of data denoising based on both DWT and MODWT transforms for these classic econometric models. From the results herein, we conclude that SOWDA is a robust algorithm that enriches the class of intelligent denoising methods in data mining. The proposed algorithm is expected to provide a significant improvement to the accuracy of econometric models in modeling and forecasting highfrequency financial data. We also believe SOWDA can be applied on other data with unknown noise patterns. Since data quality is fundamental for intelligent decision making, performing econometric models efficiently and effectively on quality data will help us build more reliable decision making system. References Aguiar-Conraria, L., Martins, M., Soares, M., The yield curve and the macro-economy across time and frequencies. Journal of Economic Dynamics and Control 36, Birgé, L., Massart, P., Minimum contrast estimators on sieves: exponential bounds and rates of convergence. Bernoulli 4(3), Cai, T., Silverman, B., Incorporating information on neighboring coefficients into wavelet estimation. Sankhya: The Indian Journal of Statistics 63, Chun, S., Shapiro, A., Uryasev, S., Conditional value-at-risk and average value-at-risk: estimation and asymptotics. Operations Research 60(4), Connor, J., Rossiter, R., Wavelet transforms and commodity prices. Studies in Nonlinear Dynamics & Econometrics 9,

20 Crowley, P., A guide to wavelets for economists. Journal of Economic Surveys 21(2), Daubechies, I., Ten lectures on wavelets, volume 61 of CBMS-NSF Regional Conference Series in Applied Mathematics. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA 118. Donoho, D., Asymptotic minimax risk for sup-norm loss: solution via optimal recovery. Probability Theory and Related Fields 99, Donoho, D., Johnstone, I., Ideal spatial adaptation by wavelet shrinkage. Biometrika 81, Donoho, D., Johnstone, I., Adapting to unknown smoothness via wavelet shrinkage. Journal of the American Statistical Association 90(432), Donoho, D., Johnstone, I., Minimax estimation via wavelet shrinkage. Annals of Statistics 26(3), Esteban-Bravo, M., Vidal-Sanz, J., Computing continuous-time growth models with boundary conditions via wavelets. Journal of Economic Dynamics and Control 31, Fan, J., Wang, Y., Multi-scale jump and volatility analysis for high-frequency financial data. Journal of the American Statistical Association 102, Fan, Y., Gençay, R., Unit root tests with wavelets. Econometric Theory 26, Gençay, R., Gradojevic, N., Errors-in-variables estimation with wavelets. Journal of Statistical Computation and Simulation 81(11), Gençay, R., Gradojevic, N., Selcuk, F., Whitcher, B., Asymmetry of information flow between volatilities across time scales. Quantitative Finance 10, Gençay, R., Selçuc, F., Whitcher, B., Multiscale systematic risk. Journal of International Money and Finance 24, Gençay, R., Selçuk, F., Whitcher, B., An introduction to wavelets and other filtering methods in finance and economics. Academic Press. Gençay, R., Selçuk, F., Whitcher, B., Systematic risk and timescales. Quantitative Finance 3(2), Grubbs, F., Procedures for detecting outlying observations in samples. Technometrics 11, Haven, E., Liu, X., Shen, L., De-noising option prices with the wavelet method. European Journal of Operational Research 222(1), Hong, Y., Kao, C., Wavelet-based testing for serial correlation of unknown form in panel models. Econometrica 72, In, F., Kim, S., Gençay, R., Investment horizon effect on asset allocation between value and growth strategies. Economic Modelling 28,

21 Jensen, M., An alternative maximum likelihood estimator of long-memory processes using compactly supported wavelets. Journal of Economic Dynamics and Control 24, Keinert, F., Wavelets and Multiwavelets. Chapman & Hall/CRC. Kelly, D., Steigerwald, D., Private information and high-frequency stochastic volatility. Studies in Nonlinear Dynamics & Econometrics 8, Kim, S., In, F., The relationship between financial variables and real economic activity: Evidence from spectral and wavelet analysis. Studies in Nonlinear Dynamics and Econometrics 7, Lada, E., Wilson, J., A wavelet-based spectral procedure for steady-state simulation analysis. European Journal of Operational Research 174(3), Laukaitis, A., Functional data analysis for cash flow and transactions intensity continuous-time prediction using Hilbert-valued autoregressive processes. European Journal of Operational Research 185(3), Lee, S., Mykland, P., Jumps in financial markets: A new nonparametric test and jump dynamics. Review of Financial Studies 21, Mallat, S., A theory for multiresolution signal decomposition: The wavelet representation. IEEE Transactions on Pattern Analysis and Machine Intelligence 11, Mallat, S., Hwang, W., Singularity detection and processing with wavelets. IEEE Transactions on Information Theory 38, Mandelbrot, B., The Fractal Geometry of Nature. W. H. Freeman & Co. Ltd. McCulloch, R., Tsay, R., Nonlinearity in high-frequency financial data and hierarchical models. Studies in Nonlinear Dynamics & Econometrics 5, Morlet, J., Sampling theory and wave propagation. Issues in Acoustic Signal/Image Processing and Recognition 1, Percival, D., Walden, A., Wavelet methods for time series analysis. Cambridge University Press. Ramsey, J., Wavelets in economics and finance: past and future. Studies in Nonlinear Dynamics and Econometrics 6, Ramsey, J., Lampart, C., The decomposition of economic relationships by time scale using wavelets: expenditure and income. Studies in Nonlinear Dynamics and Econometrics 3, Sun, E., Meinl, T., A new wavelet-based denoising algorithm for high-frequency financial data mining. European Journal of Operational Research 217, Sun, E., Rezania, O., Rachev, S., Fabozzi, F., Analysis of the intraday effects of economic releases on the currency market. Journal of International Money and Finance 30(4),

22 Sun, W., Rachev, S., Fabozzi, F., Fractals or I.I.D.: evidence of long-range dependence and heavy tailedness from modeling German equity market returns. Journal of Economics and Business 59, Sun, W., Rachev, S., Fabozzi, F., A new approach for using lèvy processes for determining high-frequency value-at-risk predictions. European Financial Management 15, Xue, Y., Gençay, R., Fagan, S., Jump detection with wavelets for high-frequency financial time series. Quantitative Finance 14,

23 Data pattern 1 Data pattern 2 Data pattern 3 Figure 1: Normal Q-Q plot for three different simulated data patterns. 23

24 Figure 2: Moving window design for the numerical studies. E is the length of the data used for training (approximation), V is the length for one-step ahead forecasting (validation), and F is the length for the two-step ahead forecasting. Given T as the total length of the data, the number of window moves is then floor((t E)/V )

Master of Mathematical Finance: Course Descriptions

Master of Mathematical Finance: Course Descriptions Master of Mathematical Finance: Course Descriptions CS 522 Data Mining Computer Science This course provides continued exploration of data mining algorithms. More sophisticated algorithms such as support

More information

Introduction to time series analysis

Introduction to time series analysis Introduction to time series analysis Margherita Gerolimetto November 3, 2010 1 What is a time series? A time series is a collection of observations ordered following a parameter that for us is time. Examples

More information

A Wavelet Based Prediction Method for Time Series

A Wavelet Based Prediction Method for Time Series A Wavelet Based Prediction Method for Time Series Cristina Stolojescu 1,2 Ion Railean 1,3 Sorin Moga 1 Philippe Lenca 1 and Alexandru Isar 2 1 Institut TELECOM; TELECOM Bretagne, UMR CNRS 3192 Lab-STICC;

More information

Booth School of Business, University of Chicago Business 41202, Spring Quarter 2015, Mr. Ruey S. Tsay. Solutions to Midterm

Booth School of Business, University of Chicago Business 41202, Spring Quarter 2015, Mr. Ruey S. Tsay. Solutions to Midterm Booth School of Business, University of Chicago Business 41202, Spring Quarter 2015, Mr. Ruey S. Tsay Solutions to Midterm Problem A: (30 pts) Answer briefly the following questions. Each question has

More information

Statistical Modeling by Wavelets

Statistical Modeling by Wavelets Statistical Modeling by Wavelets BRANI VIDAKOVIC Duke University A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York / Chichester / Weinheim / Brisbane / Singapore / Toronto Contents Preface

More information

OTHER PROFESSIONAL EXPERIENCE IN TEACHING AND RESEARCH Associate Editor of Journal of Business and Policy Research

OTHER PROFESSIONAL EXPERIENCE IN TEACHING AND RESEARCH Associate Editor of Journal of Business and Policy Research Prof. Dr. Edward W. Sun Professeur Senior en Finance KEDGE Business School France 680 cours de la Liberation, 33405 Talence Cedex, France PROFESSIONAL +33 (0)556 842 277 edward.sun@kedgebs.com EDUCATION

More information

WAVEFORM DICTIONARIES AS APPLIED TO THE AUSTRALIAN EXCHANGE RATE

WAVEFORM DICTIONARIES AS APPLIED TO THE AUSTRALIAN EXCHANGE RATE Sunway Academic Journal 3, 87 98 (26) WAVEFORM DICTIONARIES AS APPLIED TO THE AUSTRALIAN EXCHANGE RATE SHIRLEY WONG a RAY ANDERSON Victoria University, Footscray Park Campus, Australia ABSTRACT This paper

More information

Wavelet Analysis Based Estimation of Probability Density function of Wind Data

Wavelet Analysis Based Estimation of Probability Density function of Wind Data , pp.23-34 http://dx.doi.org/10.14257/ijeic.2014.5.3.03 Wavelet Analysis Based Estimation of Probability Density function of Wind Data Debanshee Datta Department of Mechanical Engineering Indian Institute

More information

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its

More information

Threshold Autoregressive Models in Finance: A Comparative Approach

Threshold Autoregressive Models in Finance: A Comparative Approach University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Informatics 2011 Threshold Autoregressive Models in Finance: A Comparative

More information

Java Modules for Time Series Analysis

Java Modules for Time Series Analysis Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series

More information

Maximum likelihood estimation of mean reverting processes

Maximum likelihood estimation of mean reverting processes Maximum likelihood estimation of mean reverting processes José Carlos García Franco Onward, Inc. jcpollo@onwardinc.com Abstract Mean reverting processes are frequently used models in real options. For

More information

Time Series Analysis

Time Series Analysis Time Series Analysis Identifying possible ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos

More information

An Evaluation of Irregularities of Milled Surfaces by the Wavelet Analysis

An Evaluation of Irregularities of Milled Surfaces by the Wavelet Analysis An Evaluation of Irregularities of Milled Surfaces by the Wavelet Analysis Włodzimierz Makieła Abstract This paper presents an introductory to wavelet analysis and its application in assessing the surface

More information

A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University

A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses Michael R. Powers[ ] Temple University and Tsinghua University Thomas Y. Powers Yale University [June 2009] Abstract We propose a

More information

A Simple Model for Intra-day Trading

A Simple Model for Intra-day Trading A Simple Model for Intra-day Trading Anton Golub 1 1 Marie Curie Fellow, Manchester Business School April 15, 2011 Abstract Since currency market is an OTC market, there is no information about orders,

More information

How To Understand The Theory Of Probability

How To Understand The Theory Of Probability Graduate Programs in Statistics Course Titles STAT 100 CALCULUS AND MATR IX ALGEBRA FOR STATISTICS. Differential and integral calculus; infinite series; matrix algebra STAT 195 INTRODUCTION TO MATHEMATICAL

More information

Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes

Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes Yong Bao a, Aman Ullah b, Yun Wang c, and Jun Yu d a Purdue University, IN, USA b University of California, Riverside, CA, USA

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization

Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization Jean- Damien Villiers ESSEC Business School Master of Sciences in Management Grande Ecole September 2013 1 Non Linear

More information

Co-integration of Stock Markets using Wavelet Theory and Data Mining

Co-integration of Stock Markets using Wavelet Theory and Data Mining Co-integration of Stock Markets using Wavelet Theory and Data Mining R.Sahu P.B.Sanjeev rsahu@iiitm.ac.in sanjeev@iiitm.ac.in ABV-Indian Institute of Information Technology and Management, India Abstract

More information

Practical Calculation of Expected and Unexpected Losses in Operational Risk by Simulation Methods

Practical Calculation of Expected and Unexpected Losses in Operational Risk by Simulation Methods Practical Calculation of Expected and Unexpected Losses in Operational Risk by Simulation Methods Enrique Navarrete 1 Abstract: This paper surveys the main difficulties involved with the quantitative measurement

More information

Affine-structure models and the pricing of energy commodity derivatives

Affine-structure models and the pricing of energy commodity derivatives Affine-structure models and the pricing of energy commodity derivatives Nikos K Nomikos n.nomikos@city.ac.uk Cass Business School, City University London Joint work with: Ioannis Kyriakou, Panos Pouliasis

More information

Supplement to Call Centers with Delay Information: Models and Insights

Supplement to Call Centers with Delay Information: Models and Insights Supplement to Call Centers with Delay Information: Models and Insights Oualid Jouini 1 Zeynep Akşin 2 Yves Dallery 1 1 Laboratoire Genie Industriel, Ecole Centrale Paris, Grande Voie des Vignes, 92290

More information

Particle Filter-Based On-Line Estimation of Spot (Cross-)Volatility with Nonlinear Market Microstructure Noise Models. Jan C.

Particle Filter-Based On-Line Estimation of Spot (Cross-)Volatility with Nonlinear Market Microstructure Noise Models. Jan C. Particle Filter-Based On-Line Estimation of Spot (Cross-)Volatility with Nonlinear Market Microstructure Noise Models Jan C. Neddermeyer (joint work with Rainer Dahlhaus, University of Heidelberg) DZ BANK

More information

A Regime-Switching Model for Electricity Spot Prices. Gero Schindlmayr EnBW Trading GmbH g.schindlmayr@enbw.com

A Regime-Switching Model for Electricity Spot Prices. Gero Schindlmayr EnBW Trading GmbH g.schindlmayr@enbw.com A Regime-Switching Model for Electricity Spot Prices Gero Schindlmayr EnBW Trading GmbH g.schindlmayr@enbw.com May 31, 25 A Regime-Switching Model for Electricity Spot Prices Abstract Electricity markets

More information

Operations Research and Financial Engineering. Courses

Operations Research and Financial Engineering. Courses Operations Research and Financial Engineering Courses ORF 504/FIN 504 Financial Econometrics Professor Jianqing Fan This course covers econometric and statistical methods as applied to finance. Topics

More information

LOGNORMAL MODEL FOR STOCK PRICES

LOGNORMAL MODEL FOR STOCK PRICES LOGNORMAL MODEL FOR STOCK PRICES MICHAEL J. SHARPE MATHEMATICS DEPARTMENT, UCSD 1. INTRODUCTION What follows is a simple but important model that will be the basis for a later study of stock prices as

More information

Wavelet analysis. Wavelet requirements. Example signals. Stationary signal 2 Hz + 10 Hz + 20Hz. Zero mean, oscillatory (wave) Fast decay (let)

Wavelet analysis. Wavelet requirements. Example signals. Stationary signal 2 Hz + 10 Hz + 20Hz. Zero mean, oscillatory (wave) Fast decay (let) Wavelet analysis In the case of Fourier series, the orthonormal basis is generated by integral dilation of a single function e jx Every 2π-periodic square-integrable function is generated by a superposition

More information

SYSTEMS OF REGRESSION EQUATIONS

SYSTEMS OF REGRESSION EQUATIONS SYSTEMS OF REGRESSION EQUATIONS 1. MULTIPLE EQUATIONS y nt = x nt n + u nt, n = 1,...,N, t = 1,...,T, x nt is 1 k, and n is k 1. This is a version of the standard regression model where the observations

More information

The term structure of Russian interest rates

The term structure of Russian interest rates The term structure of Russian interest rates Stanislav Anatolyev New Economic School, Moscow Sergey Korepanov EvrazHolding, Moscow Corresponding author. Address: Stanislav Anatolyev, New Economic School,

More information

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data Athanasius Zakhary, Neamat El Gayar Faculty of Computers and Information Cairo University, Giza, Egypt

More information

Four Essays on the Empirical Properties of Stock Market Volatility

Four Essays on the Empirical Properties of Stock Market Volatility Four Essays on the Empirical Properties of Stock Market Volatility Thesis Presented to the Faculty of Economics and Social Sciences of the University of Fribourg (Switzerland) in fulfillment of the requirements

More information

STATISTICA Formula Guide: Logistic Regression. Table of Contents

STATISTICA Formula Guide: Logistic Regression. Table of Contents : Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

More information

Contents. List of Figures. List of Tables. List of Examples. Preface to Volume IV

Contents. List of Figures. List of Tables. List of Examples. Preface to Volume IV Contents List of Figures List of Tables List of Examples Foreword Preface to Volume IV xiii xvi xxi xxv xxix IV.1 Value at Risk and Other Risk Metrics 1 IV.1.1 Introduction 1 IV.1.2 An Overview of Market

More information

PHASE ESTIMATION ALGORITHM FOR FREQUENCY HOPPED BINARY PSK AND DPSK WAVEFORMS WITH SMALL NUMBER OF REFERENCE SYMBOLS

PHASE ESTIMATION ALGORITHM FOR FREQUENCY HOPPED BINARY PSK AND DPSK WAVEFORMS WITH SMALL NUMBER OF REFERENCE SYMBOLS PHASE ESTIMATION ALGORITHM FOR FREQUENCY HOPPED BINARY PSK AND DPSK WAVEFORMS WITH SMALL NUM OF REFERENCE SYMBOLS Benjamin R. Wiederholt The MITRE Corporation Bedford, MA and Mario A. Blanco The MITRE

More information

TEMPORAL CAUSAL RELATIONSHIP BETWEEN STOCK MARKET CAPITALIZATION, TRADE OPENNESS AND REAL GDP: EVIDENCE FROM THAILAND

TEMPORAL CAUSAL RELATIONSHIP BETWEEN STOCK MARKET CAPITALIZATION, TRADE OPENNESS AND REAL GDP: EVIDENCE FROM THAILAND I J A B E R, Vol. 13, No. 4, (2015): 1525-1534 TEMPORAL CAUSAL RELATIONSHIP BETWEEN STOCK MARKET CAPITALIZATION, TRADE OPENNESS AND REAL GDP: EVIDENCE FROM THAILAND Komain Jiranyakul * Abstract: This study

More information

Financial TIme Series Analysis: Part II

Financial TIme Series Analysis: Part II Department of Mathematics and Statistics, University of Vaasa, Finland January 29 February 13, 2015 Feb 14, 2015 1 Univariate linear stochastic models: further topics Unobserved component model Signal

More information

CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen

CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 3: DATA TRANSFORMATION AND DIMENSIONALITY REDUCTION Chapter 3: Data Preprocessing Data Preprocessing: An Overview Data Quality Major

More information

Geostatistics Exploratory Analysis

Geostatistics Exploratory Analysis Instituto Superior de Estatística e Gestão de Informação Universidade Nova de Lisboa Master of Science in Geospatial Technologies Geostatistics Exploratory Analysis Carlos Alberto Felgueiras cfelgueiras@isegi.unl.pt

More information

The CUSUM algorithm a small review. Pierre Granjon

The CUSUM algorithm a small review. Pierre Granjon The CUSUM algorithm a small review Pierre Granjon June, 1 Contents 1 The CUSUM algorithm 1.1 Algorithm............................... 1.1.1 The problem......................... 1.1. The different steps......................

More information

Univariate and Multivariate Methods PEARSON. Addison Wesley

Univariate and Multivariate Methods PEARSON. Addison Wesley Time Series Analysis Univariate and Multivariate Methods SECOND EDITION William W. S. Wei Department of Statistics The Fox School of Business and Management Temple University PEARSON Addison Wesley Boston

More information

When to Refinance Mortgage Loans in a Stochastic Interest Rate Environment

When to Refinance Mortgage Loans in a Stochastic Interest Rate Environment When to Refinance Mortgage Loans in a Stochastic Interest Rate Environment Siwei Gan, Jin Zheng, Xiaoxia Feng, and Dejun Xie Abstract Refinancing refers to the replacement of an existing debt obligation

More information

Statistics in Retail Finance. Chapter 6: Behavioural models

Statistics in Retail Finance. Chapter 6: Behavioural models Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural

More information

On Correlating Performance Metrics

On Correlating Performance Metrics On Correlating Performance Metrics Yiping Ding and Chris Thornley BMC Software, Inc. Kenneth Newman BMC Software, Inc. University of Massachusetts, Boston Performance metrics and their measurements are

More information

Part II Redundant Dictionaries and Pursuit Algorithms

Part II Redundant Dictionaries and Pursuit Algorithms Aisenstadt Chair Course CRM September 2009 Part II Redundant Dictionaries and Pursuit Algorithms Stéphane Mallat Centre de Mathématiques Appliquées Ecole Polytechnique Sparsity in Redundant Dictionaries

More information

Using simulation to calculate the NPV of a project

Using simulation to calculate the NPV of a project Using simulation to calculate the NPV of a project Marius Holtan Onward Inc. 5/31/2002 Monte Carlo simulation is fast becoming the technology of choice for evaluating and analyzing assets, be it pure financial

More information

Volatility modeling in financial markets

Volatility modeling in financial markets Volatility modeling in financial markets Master Thesis Sergiy Ladokhin Supervisors: Dr. Sandjai Bhulai, VU University Amsterdam Brian Doelkahar, Fortis Bank Nederland VU University Amsterdam Faculty of

More information

Bootstrapping Big Data

Bootstrapping Big Data Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu

More information

Time Series Analysis

Time Series Analysis Time Series Analysis Forecasting with ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos (UC3M-UPM)

More information

A Confidence Interval Triggering Method for Stock Trading Via Feedback Control

A Confidence Interval Triggering Method for Stock Trading Via Feedback Control A Confidence Interval Triggering Method for Stock Trading Via Feedback Control S. Iwarere and B. Ross Barmish Abstract This paper builds upon the robust control paradigm for stock trading established in

More information

Charles University, Faculty of Mathematics and Physics, Prague, Czech Republic.

Charles University, Faculty of Mathematics and Physics, Prague, Czech Republic. WDS'09 Proceedings of Contributed Papers, Part I, 148 153, 2009. ISBN 978-80-7378-101-9 MATFYZPRESS Volatility Modelling L. Jarešová Charles University, Faculty of Mathematics and Physics, Prague, Czech

More information

Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections

Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Maximilian Hung, Bohyun B. Kim, Xiling Zhang August 17, 2013 Abstract While current systems already provide

More information

SPARE PARTS INVENTORY SYSTEMS UNDER AN INCREASING FAILURE RATE DEMAND INTERVAL DISTRIBUTION

SPARE PARTS INVENTORY SYSTEMS UNDER AN INCREASING FAILURE RATE DEMAND INTERVAL DISTRIBUTION SPARE PARS INVENORY SYSEMS UNDER AN INCREASING FAILURE RAE DEMAND INERVAL DISRIBUION Safa Saidane 1, M. Zied Babai 2, M. Salah Aguir 3, Ouajdi Korbaa 4 1 National School of Computer Sciences (unisia),

More information

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium

More information

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010 Simulation Methods Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Simulation Methods 15.450, Fall 2010 1 / 35 Outline 1 Generating Random Numbers 2 Variance Reduction 3 Quasi-Monte

More information

Evolutionary denoising based on an estimation of Hölder exponents with oscillations.

Evolutionary denoising based on an estimation of Hölder exponents with oscillations. Evolutionary denoising based on an estimation of Hölder exponents with oscillations. Pierrick Legrand,, Evelyne Lutton and Gustavo Olague CICESE, Research Center, Applied Physics Division Centro de Investigación

More information

7 Time series analysis

7 Time series analysis 7 Time series analysis In Chapters 16, 17, 33 36 in Zuur, Ieno and Smith (2007), various time series techniques are discussed. Applying these methods in Brodgar is straightforward, and most choices are

More information

Detection of changes in variance using binary segmentation and optimal partitioning

Detection of changes in variance using binary segmentation and optimal partitioning Detection of changes in variance using binary segmentation and optimal partitioning Christian Rohrbeck Abstract This work explores the performance of binary segmentation and optimal partitioning in the

More information

The information content of lagged equity and bond yields

The information content of lagged equity and bond yields Economics Letters 68 (2000) 179 184 www.elsevier.com/ locate/ econbase The information content of lagged equity and bond yields Richard D.F. Harris *, Rene Sanchez-Valle School of Business and Economics,

More information

Summary Nonstationary Time Series Multitude of Representations Possibilities from Applied Computational Harmonic Analysis Tests of Stationarity

Summary Nonstationary Time Series Multitude of Representations Possibilities from Applied Computational Harmonic Analysis Tests of Stationarity Nonstationary Time Series, Priestley s Evolutionary Spectra and Wavelets Guy Nason, School of Mathematics, University of Bristol Summary Nonstationary Time Series Multitude of Representations Possibilities

More information

Aachen Summer Simulation Seminar 2014

Aachen Summer Simulation Seminar 2014 Aachen Summer Simulation Seminar 2014 Lecture 07 Input Modelling + Experimentation + Output Analysis Peer-Olaf Siebers pos@cs.nott.ac.uk Motivation 1. Input modelling Improve the understanding about how

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

MARKETS, INFORMATION AND THEIR FRACTAL ANALYSIS. Mária Bohdalová and Michal Greguš Comenius University, Faculty of Management Slovak republic

MARKETS, INFORMATION AND THEIR FRACTAL ANALYSIS. Mária Bohdalová and Michal Greguš Comenius University, Faculty of Management Slovak republic MARKETS, INFORMATION AND THEIR FRACTAL ANALYSIS Mária Bohdalová and Michal Greguš Comenius University, Faculty of Management Slovak republic Abstract: We will summarize the impact of the conflict between

More information

Stephane Crepey. Financial Modeling. A Backward Stochastic Differential Equations Perspective. 4y Springer

Stephane Crepey. Financial Modeling. A Backward Stochastic Differential Equations Perspective. 4y Springer Stephane Crepey Financial Modeling A Backward Stochastic Differential Equations Perspective 4y Springer Part I An Introductory Course in Stochastic Processes 1 Some Classes of Discrete-Time Stochastic

More information

Statistics Graduate Courses

Statistics Graduate Courses Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.

More information

Monte Carlo testing with Big Data

Monte Carlo testing with Big Data Monte Carlo testing with Big Data Patrick Rubin-Delanchy University of Bristol & Heilbronn Institute for Mathematical Research Joint work with: Axel Gandy (Imperial College London) with contributions from:

More information

Quantitative Methods for Finance

Quantitative Methods for Finance Quantitative Methods for Finance Module 1: The Time Value of Money 1 Learning how to interpret interest rates as required rates of return, discount rates, or opportunity costs. 2 Learning how to explain

More information

Measuring downside risk of stock returns with time-dependent volatility (Downside-Risikomessung für Aktien mit zeitabhängigen Volatilitäten)

Measuring downside risk of stock returns with time-dependent volatility (Downside-Risikomessung für Aktien mit zeitabhängigen Volatilitäten) Topic 1: Measuring downside risk of stock returns with time-dependent volatility (Downside-Risikomessung für Aktien mit zeitabhängigen Volatilitäten) One of the principal objectives of financial risk management

More information

A comparison between different volatility models. Daniel Amsköld

A comparison between different volatility models. Daniel Amsköld A comparison between different volatility models Daniel Amsköld 211 6 14 I II Abstract The main purpose of this master thesis is to evaluate and compare different volatility models. The evaluation is based

More information

Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts

Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts Page 1 of 20 ISF 2008 Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts Andrey Davydenko, Professor Robert Fildes a.davydenko@lancaster.ac.uk Lancaster

More information

Multiple Linear Regression in Data Mining

Multiple Linear Regression in Data Mining Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple

More information

Wavelet Analysis on Stochastic Time Series A visual introduction with an examination of long term financial time series

Wavelet Analysis on Stochastic Time Series A visual introduction with an examination of long term financial time series Wavelet Analysis on Stochastic Time Series A visual introduction with an examination of long term financial time series Tobias Setz supervised by PD Diethelm Würtz ETH Zürich Computational Science and

More information

A VARIANT OF THE EMD METHOD FOR MULTI-SCALE DATA

A VARIANT OF THE EMD METHOD FOR MULTI-SCALE DATA Advances in Adaptive Data Analysis Vol. 1, No. 4 (2009) 483 516 c World Scientific Publishing Company A VARIANT OF THE EMD METHOD FOR MULTI-SCALE DATA THOMAS Y. HOU and MIKE P. YAN Applied and Computational

More information

3. Regression & Exponential Smoothing

3. Regression & Exponential Smoothing 3. Regression & Exponential Smoothing 3.1 Forecasting a Single Time Series Two main approaches are traditionally used to model a single time series z 1, z 2,..., z n 1. Models the observation z t as a

More information

Measuring Line Edge Roughness: Fluctuations in Uncertainty

Measuring Line Edge Roughness: Fluctuations in Uncertainty Tutor6.doc: Version 5/6/08 T h e L i t h o g r a p h y E x p e r t (August 008) Measuring Line Edge Roughness: Fluctuations in Uncertainty Line edge roughness () is the deviation of a feature edge (as

More information

The Combination Forecasting Model of Auto Sales Based on Seasonal Index and RBF Neural Network

The Combination Forecasting Model of Auto Sales Based on Seasonal Index and RBF Neural Network , pp.67-76 http://dx.doi.org/10.14257/ijdta.2016.9.1.06 The Combination Forecasting Model of Auto Sales Based on Seasonal Index and RBF Neural Network Lihua Yang and Baolin Li* School of Economics and

More information

From the help desk: Bootstrapped standard errors

From the help desk: Bootstrapped standard errors The Stata Journal (2003) 3, Number 1, pp. 71 80 From the help desk: Bootstrapped standard errors Weihua Guan Stata Corporation Abstract. Bootstrapping is a nonparametric approach for evaluating the distribution

More information

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic

More information

Local outlier detection in data forensics: data mining approach to flag unusual schools

Local outlier detection in data forensics: data mining approach to flag unusual schools Local outlier detection in data forensics: data mining approach to flag unusual schools Mayuko Simon Data Recognition Corporation Paper presented at the 2012 Conference on Statistical Detection of Potential

More information

Time series analysis of data from stress ECG

Time series analysis of data from stress ECG Communications to SIMAI Congress, ISSN 827-905, Vol. 3 (2009) DOI: 0.685/CSC09XXX Time series analysis of data from stress ECG Camillo Cammarota Dipartimento di Matematica La Sapienza Università di Roma,

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

Monte Carlo Simulation

Monte Carlo Simulation 1 Monte Carlo Simulation Stefan Weber Leibniz Universität Hannover email: sweber@stochastik.uni-hannover.de web: www.stochastik.uni-hannover.de/ sweber Monte Carlo Simulation 2 Quantifying and Hedging

More information

Monte Carlo Methods in Finance

Monte Carlo Methods in Finance Author: Yiyang Yang Advisor: Pr. Xiaolin Li, Pr. Zari Rachev Department of Applied Mathematics and Statistics State University of New York at Stony Brook October 2, 2012 Outline Introduction 1 Introduction

More information

Predictive Indicators for Effective Trading Strategies By John Ehlers

Predictive Indicators for Effective Trading Strategies By John Ehlers Predictive Indicators for Effective Trading Strategies By John Ehlers INTRODUCTION Technical traders understand that indicators need to smooth market data to be useful, and that smoothing introduces lag

More information

A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA

A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA Krish Muralidhar University of Kentucky Rathindra Sarathy Oklahoma State University Agency Internal User Unmasked Result Subjects

More information

Analysis of a Production/Inventory System with Multiple Retailers

Analysis of a Production/Inventory System with Multiple Retailers Analysis of a Production/Inventory System with Multiple Retailers Ann M. Noblesse 1, Robert N. Boute 1,2, Marc R. Lambrecht 1, Benny Van Houdt 3 1 Research Center for Operations Management, University

More information

Analysis of Financial Time Series

Analysis of Financial Time Series Analysis of Financial Time Series Analysis of Financial Time Series Financial Econometrics RUEY S. TSAY University of Chicago A Wiley-Interscience Publication JOHN WILEY & SONS, INC. This book is printed

More information

Tutorial 5: Hypothesis Testing

Tutorial 5: Hypothesis Testing Tutorial 5: Hypothesis Testing Rob Nicholls nicholls@mrc-lmb.cam.ac.uk MRC LMB Statistics Course 2014 Contents 1 Introduction................................ 1 2 Testing distributional assumptions....................

More information

ANALYZER BASICS WHAT IS AN FFT SPECTRUM ANALYZER? 2-1

ANALYZER BASICS WHAT IS AN FFT SPECTRUM ANALYZER? 2-1 WHAT IS AN FFT SPECTRUM ANALYZER? ANALYZER BASICS The SR760 FFT Spectrum Analyzer takes a time varying input signal, like you would see on an oscilloscope trace, and computes its frequency spectrum. Fourier's

More information

Forecasting methods applied to engineering management

Forecasting methods applied to engineering management Forecasting methods applied to engineering management Áron Szász-Gábor Abstract. This paper presents arguments for the usefulness of a simple forecasting application package for sustaining operational

More information

Big Data, Socio- Psychological Theory, Algorithmic Text Analysis, and Predicting the Michigan Consumer Sentiment Index

Big Data, Socio- Psychological Theory, Algorithmic Text Analysis, and Predicting the Michigan Consumer Sentiment Index Big Data, Socio- Psychological Theory, Algorithmic Text Analysis, and Predicting the Michigan Consumer Sentiment Index Rickard Nyman *, Paul Ormerod Centre for the Study of Decision Making Under Uncertainty,

More information

The relation between news events and stock price jump: an analysis based on neural network

The relation between news events and stock price jump: an analysis based on neural network 20th International Congress on Modelling and Simulation, Adelaide, Australia, 1 6 December 2013 www.mssanz.org.au/modsim2013 The relation between news events and stock price jump: an analysis based on

More information

A Primer on Forecasting Business Performance

A Primer on Forecasting Business Performance A Primer on Forecasting Business Performance There are two common approaches to forecasting: qualitative and quantitative. Qualitative forecasting methods are important when historical data is not available.

More information

How To Analyze The Time Varying And Asymmetric Dependence Of International Crude Oil Spot And Futures Price, Price, And Price Of Futures And Spot Price

How To Analyze The Time Varying And Asymmetric Dependence Of International Crude Oil Spot And Futures Price, Price, And Price Of Futures And Spot Price Send Orders for Reprints to reprints@benthamscience.ae The Open Petroleum Engineering Journal, 2015, 8, 463-467 463 Open Access Asymmetric Dependence Analysis of International Crude Oil Spot and Futures

More information

APPENDIX N. Data Validation Using Data Descriptors

APPENDIX N. Data Validation Using Data Descriptors APPENDIX N Data Validation Using Data Descriptors Data validation is often defined by six data descriptors: 1) reports to decision maker 2) documentation 3) data sources 4) analytical method and detection

More information

Hedging Illiquid FX Options: An Empirical Analysis of Alternative Hedging Strategies

Hedging Illiquid FX Options: An Empirical Analysis of Alternative Hedging Strategies Hedging Illiquid FX Options: An Empirical Analysis of Alternative Hedging Strategies Drazen Pesjak Supervised by A.A. Tsvetkov 1, D. Posthuma 2 and S.A. Borovkova 3 MSc. Thesis Finance HONOURS TRACK Quantitative

More information

Predictability of Non-Linear Trading Rules in the US Stock Market Chong & Lam 2010

Predictability of Non-Linear Trading Rules in the US Stock Market Chong & Lam 2010 Department of Mathematics QF505 Topics in quantitative finance Group Project Report Predictability of on-linear Trading Rules in the US Stock Market Chong & Lam 010 ame: Liu Min Qi Yichen Zhang Fengtian

More information

TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS

TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS 1. Bandwidth: The bandwidth of a communication link, or in general any system, was loosely defined as the width of

More information

The Variability of P-Values. Summary

The Variability of P-Values. Summary The Variability of P-Values Dennis D. Boos Department of Statistics North Carolina State University Raleigh, NC 27695-8203 boos@stat.ncsu.edu August 15, 2009 NC State Statistics Departement Tech Report

More information