Modelling and Forecasting financial time series of «tick data» by functional analysis and neural networks
|
|
|
- Harriet Stanley
- 10 years ago
- Views:
Transcription
1 Modelling and Forecasting financial time series of «tick data» by functional analysis and neural networks S DABLEMONT S VAN BELLEGEM M VERLEYSEN Université catholique de Louvain Machine Learning Group DICE 3 Place du Levant B-1348 Louvain-la-Neuve - BELGIUM Tel : Fax : {dablemont verleysen}@diceuclacbe vanbellegem@statuclacbe Subjects Finances Neural Networks Forecasting Nonlinear Time Series Models Tick Data Abstract The analysis of financial time series is of primary importance in the economic world This paper deals with a data-driven empirical analysis of financial time series The goal is to obtain insights into the dynamics of series and out-of-sample forecasting In this paper we present a forecasting method based on an empirical functional analysis of the past of series An originality of this method is that it does not make the assumption that a single model is able to capture the dynamics of the whole series On the contrary it splits the past of the series into clusters and generates a specific local neural model for each of them The local models are then combined in a probabilistic way according to the distribution of the series in the past This forecasting method can be applied to any time series forecasting problem but is particularly suited for data showing nonlinear dependencies cluster effects and observed at irregularly and randomly spaced times like high-frequency financial time series do One way to overcome the irregular and random sampling of "tick-data" is to resample them at low-frequency as it is done with "Intraday" However even with optimal resampling using say five minute returns when transactions are recorded every second a vast amount of data is discarded in contradiction to basic statistical principles Thus modelling the noise and using all the data is a better solution even if one misspecifies the noise distribution The method is applied to the forecasting of financial time series of «tick data» of assets on a short horizon in order to be useful for speculators
2 1 Introduction The analysis of financial time series is of primary importance in the economic world This paper deals with a data-driven empirical analysis of financial time series the goal is to obtain insights into the dynamics of series and out-of-sample forecasting Forecasting future returns on assets is of obvious interest in empirical finance If one were able to forecast tomorrow s returns on an asset with some degree of precision one could use this information in an investment today Unfortunately we are seldom able to generate a very accurate prediction for asset returns Financial time series display typical nonlinear characteristics it exists clusters within which returns and volatility display specific dynamic behavior For this reason we consider here nonlinear forecasting models based on local analysis into clusters Although financial theory does not provide many motivations for nonlinear models analyzing data by nonlinear tools seems to be appropriate and is at least as much informative as an analysis by more restrictive linear methods Time series of asset returns can be characterized as serial dependent This is revealed by the presence of positive autocorrelation in squared returns and sometimes in the returns too The increased importance played by risk and uncertainty considerations in modern economic theory has necessitated the development of new econometric time series techniques that allow for modelling of time varying means variances and covariances Given the apparent lack of any structural dynamic economic theory explaining the variation in the second moment econometricians have thus extended traditional time series tools such as AutoRegressive Moving Average ARMA) models Box and Jenkins 1970) for the conditional means and equivalent models for the conditional variance Indeed the dynamics observed in the dispersion is clearly the dominating feature in the data The most widespread modelling approach to capture these properties is to specify a dynamic model for the conditional means and the conditional variance such as an ARMA-GARCH model or one of its various extensions Engle 1982) Hamilton 1994) The Gaussian random walk paradigm - under the form of the diffusion geometric Wiener process - is the core of modelling of financial time series Its robustness mostly suffices to keep it as the best foundation for any development in financial modelling in addition to the fact that in the long run and with enough spaced out data it is almost verified by the facts Failures in its application are however well admitted on the very) short term market microstructure) Fama 1991) Olsen and Dacorogna 1992) Franke et al 2002) We claim that to some extent such failures are actually caused by the uniqueness of the modelling process The first breach in such a unique process has appeared with two-regime or switching processes Diebold et al 1994) which recognize that a return process could be originated by two different stochastic differential equations But in such a case the switch is governed by an exogenous cause for example in the case of exchange rates the occurrence of a central bank decision to modify its leading interest rate or to organize a huge buying or selling of its currency through major banks)
3 Market practitioners Engle 1982) however have always observed that financial markets can follow different behaviors over time such as overreaction mean reversion etc which look like succeeding each other with the passing of the time Such observations would justify a rather fundamental divergence from the classic modelling foundations That is financial markets should not be modeled by a single process but rather by a succession of different processes even in absence of the exogenous causes retained by existing switching process Such a multiple switching process should imply first the determination of a limited number of competitive sub-processes and secondly the identification of the factors) causing the switch from one to another sub-processes The resulting model should not be Markovian and without doubt would be hard to determine The aim of this paper is as a first step to at least empirically verify with the help of functional clustering and neural networks that a multiple switching process leads to better short term forecasting In this paper we present a forecasting method based on an empirical functional analysis of the past of series An originality of this method is that it does not make the assumption that a single model is able to capture the dynamics of the whole series On the contrary it splits the past of the series into clusters and generates a specific local neural model for each of them The local models are then combined in a probabilistic way according to the distribution of the series in the past This forecasting method can be applied to any time series forecasting problem but is particularly suited for data showing nonlinear dependencies cluster effects and observed at irregularly and randomly spaced times like high-frequency financial time series do One way to overcome the irregular and random sampling of "tick-data" is to resample them at low frequency as it is done with "Intraday" However even with optimal resampling using say five minute returns when transactions are recorded every second a vast amount of data is discarded in contradiction to basic statistical principles Thus modelling the noise and using all the data is a better solution even if one misspecifies the noise distribution Ait-Sahalia and Myland 2003) And one way to get to this goal is by using Functional Analysis as done in this paper Further in this paper we first describe how Functional Analysis can be applied to time series data section 2) and briefly introduce the Radial-Basis Functions Networks we use as nonlinear models section 3) Then we describe the forecasting method itself section 4) and illustrate its results on the IBM series of "tick data" section 5) 2 Functional Modelling and Clustering Our purpose is to realize the clustering of the observations into classes having homogenous properties in order to build nonlinear local forecasting models in each class When the observations are sparse irregularly spaced or occur at different time points for each subject as with high-frequency financial series standards statistical tools can not be used because we do not have the same number of observations for each series and the observations are not at the same time-point In this case it will be necessary to
4 represent these data by a fixed number of features One way to get to this purpose is to smooth the rough data by projecting them onto a functional basis for example cubic splines Then the coefficients of this projection may be used in a more standard way for clustering purposes In section 21) we introduce the generic problem of clustering in section 22) we explain why we have to use this tool for high-frequency financial series in section 23) we define the functional model we use in section 24) we build the likelihood function and in section 25) we have the procedures to optimize the parameters 21 Clustering Cluster analysis consists in identifying groups in data; it is the dual form of discriminant analysis but in cluster analysis the group labels are not known a priori It is an unsupervised process We assume that the observations {y 1 y N } are generated according to a mixture distribution with G clusters Let f k y θ k ) be the density distribution function corresponding to cluster k with parameters θ k and let 1 {k} i) be the cluster membership indicator function of cluster k) for the observation i where 1 {k} i) = 1 if y i is a member of cluster k and 0 otherwise The indicators are unknown and 1 {k} i) is multinomial with parameters [π 1 π G ] and π k is the probability that an observation belongs to cluster k We can estimate the parameters by maximizing the likelihood Lθ 1 θ G ; π 1 π G y 1 y N ) = N G π k f k y i θ k ) 1) The maximum likelihood corresponds to the most probable model given the observations {y 1 y N } Such model can be used in finite dimensional problems but it is not appropriated to infinite dimensional data such as curves Hastie et al 2001) We could get around by discretizing the time interval but generally the resulting data vectors are highly correlated and high-dimensional and by resampling at low frequency we loose much information Ait-Sahalia and Myland 2003) Another approach is to project each curve onto a finite-dimensional basis φx) and find the best projection of each curve onto this basis The resulting basis coefficients can than be used as a finite-dimensional representation making it possible to use classical clustering methods on the basis coefficients Ramsay and Silverman 1997) These approaches can work well when every curve has been observed over the same fine grid of points but they break down if the individual curves are sparsely sampled The variance of the estimated basis coefficients is different for each individual because the curves are measured at different time points And for sparse data sets many of the basis coefficients would have infinite variance making it impossible to produce reasonable estimates
5 In this case we convert the original infinite dimensional problem into a finite dimensional one using basic functions and we use a random effects model for the coefficients Rice 2001) 22 Modelling functional data Modern financial data sets may contain tens of thousands of transactions per day stamped to the nearest second The analysis of these data are complicated due to stochastic temporal spacing diurnal patterns prices discreteness and complex temporal dependence Let data coming from an interval [t 0 t N ] If we use very liquid series like stocks with observations onto the whole interval of time we could realize the smoothing by splines and afterwards the clustering from the spline coefficients in two separate steps If we don t have enough data near the limits t 0 or t N of the interval the smoothing by splines will not be fine at these limits we will have many outliers and the clustering will be very poor If we use poorly liquid data with large fragment of curve without observations like options the smoothing could be very chaotic In those cases we have to realize the smoothing of the observations and the clustering in the same iterative steps Thus when the observations are sparse irregularly spaced or occur at different time points for each subject and moreover when only fragments of the function are available with the Linear Discriminant Analysis LDA) many of the basis coefficients would have infinite variance making it impossible to produce reasonable estimates Gareth et al 2000) and similar problem arise with the clustering methods In this case we will use a random effects model for the coefficients and we will realize in the same step the estimation of the splines coefficients and the clustering Gareth and Sugar 2003) 23 Functional Clustering We will use basis functions in order to convert the original infinite dimensional problem into a finite dimensional one but instead of treating the basis coefficients as parameters and fitting a separate spline for each individual we will use a random effects model for the coefficients This procedure borrows information across curves and produces far better results no matter how sparsely or irregularly the individual curves are sampled provided that the total number of observations is large enough Moreover it automatically weights the estimated spline coefficients according to their variances which is highly efficient because it requires fitting few parameters and it can be used to produce estimates of individual curves that are optimal in terms of mean square errors Let g i t) the true value for the curve i at time t and g i y i and ɛ i the vectors of true values measurements and errors at times t i1 t i2 t ini We have got : y i = g i + ɛ i i = 1 N 2) where N is the number of curves The errors are assumed to have mean zero and uncorrelated with each other and with g i Let : T g i = g i t 1 ) g i t j ) g i t ni )) T y i = y i t 1 ) y i t j ) y i t ni ))
6 T ɛ i = ɛ i t 1 ) ɛ i t j ) ɛ i t ni )) where t j is the time-point for the observation j and n i is the number of observations for the curve i For the true values g i we use a functional basis for their representation and we have : g i t) = s T t) η i 3) where st) is the spline basis vector with q dimensions and η i is a vector of spline coefficients η i is a Gaussian random variable) Let : st) = η i = s 1 t) s q t)) T η i1 η iq ) T If we used a "power cubic spline" with 1 knot then q = 5 and we would have : T st) = s 1 t) s 5 t)) with : s 1 t) = 1 s 2 t) = t s 3 t) = t 2 s 4 t) = t 3 s 5 t) = t τ) 3 + where τ is the knot For the Gaussian coefficients η i we have : η i = µ zi + γ i γ N0 Γ) 4) where z i denotes the unknown cluster membership for the curve i and it will be treated as missing data { 1 if curve i belongs to cluster k z ki = 0 otherwise then we have : P z ki = 1 ) = π k i ) T µ zi = µ i1zi µ iqzi µ zi = { µ 1) ) z i1 µ k) ) z ik µ G) ) ig} z ) T γ i = γ i1 γ iq We have split η i into two terms µ zi represents the centroid of the cluster and γ i represents the curve in its cluster Also in the same way we can represent the centroid of the cluster from the global mean of the population by : µ k = λ 0 + Λα k 5) where λ 0 is a q 1) vector and α k h 1) Λ is q h) matrix with h minq G 1) λ 0 = α k = λ 01 λ 0q ) T α k1 α kh ) T
7 Γ 11 Γ 1h Λ = Γ p1 Γ qh With this formulation the functional clustering model can be written as : y i = S i λ 0 + Λα zi + γ i ) + ɛ i i = 1 N 6) ɛ i N0 R) γ i N0 Γ) [ T where S i = st i1 ) st ini )] is the splines basis matrix for the individual i s 1 t 1 ) s q t 1 ) S i = s 1 t ni ) s q t ni ) α zi = { α 1) ) z i1 α k) ) z ik α G) ) z ig} but λ 0 α k and Λ could be confounded if no constraints were imposed : Hastie et al 2001) Therefore we require that : α k = 0 7) k that means that st) T λ 0 may be interpreted as the overall mean curve and with : Λ T S T Σ 1 SΛ = I 8) Σ = σ 2 I + SΓS T where S is the splines basis matrix on a fine grid of time points over the full range of the data and we will put R = σ 2 I and with Γ the same for every cluster Then we have : st) T λ 0 the representation of the global mean curve st) T λ 0 + Λα k ) the global representation of the centroid of cluster k st) T Λα k the local representation of the centroid of cluster k in connection with the global mean curve st) T γ i the local representation of the curve i in connection with the centroid of its cluster k
8 231 Example : Equations for curve i Let the curve i with n i observations at times {t i1 t i2 t ini } with g i t) the unknown true value y i t) the measurement and ɛ i t) the measurement error at time t and G the number of clusters We have : y i t 1 ) g i t 1 ) ɛ i t 1 ) = + 9) y i t ni ) g i t ni ) ɛ i t ni ) We represent the value g i t) at time t on a spline basis vector st) with q dimensions and η i the random spline coefficients vector Then we have : η i1 g i t) = s 1 t) s q t) ) η iq 10) and an exhaustive description : g i t 1 ) s 1 t 1 ) s q t 1 ) η i1 = 11) g i t ni ) s 1 t ni ) s q t ni ) We split the random spline coefficients vector η i into a deterministic coefficients vector µ zi if the curve i is part of the cluster k defined by the cluster membership z ik = 1 and z il = 0 for l k l = 1 G and a random coefficients vector γ i : η i1 η iq = µ 1 µ q zi + γ i1 γ iq η iq 12) We also split the deterministic spline coefficients vector µ i into a deterministic coefficients vector λ 0 which represents the coefficients of the global mean curve of all curves and Λα k which represents the centroid coefficients of the cluster k from the coefficients of the global mean curve of the population : µ k1 µ kq λ 01 = λ 0q + λ 11 λ 1h λ q1 λ qh With all these representations we have got : y i t 1 ) s 1 t 1 ) s q t 1 ) = y i t ni ) s 1 t ni ) s q t ni ) λ 01 λ 11 λ 1h + λ q1 λ qh λ 0q α k1 α kh α k1 α kh 13) γ i1 + γ iq ɛ i t 1 ) + 14) ɛ i t ni )
9 24 Parametric Identification Now we have to estimate the parameters λ 0 Λ α k Γ σ 2 et π k by maximization of a likelihood function For y i we have a conditional distribution : ) y i N S i λ 0 + Λα zi ) Σ i where Σ i = σ 2 I + S i ΓS T i Since the observations of the different curves are independent the joint distribution of y and z is given by : fy z) = G π k 1 2π) n 2 Σ 1 2 [ exp 1 y Sλ0 + Λα z ) 2 ) T Σ 1 y Sλ 0 + Λα z ) )] and the likelihood for the parameters π k λ 0 Λ α k Γ σ 2 given the observations y i z i is : Lπ k λ 0 Λ α k Γ σ 2 y i z i ) = N G 1 [ π k exp 1 2π) n i 2 Σ i ) yi S i λ 0 + Λα zi ) ) T Σ 1 i yi S i λ 0 + Λα zi ) )] Maximizing this likelihood would give us the parameters λ 0 Λ α k π k Γ and σ 2 but unfortunately a direct maximization of this likelihood is a difficult non-convex optimization problem If the γ i had been observed then the joint likelihood of y i z i and γ i would simplify and like z i and γ i are independent the joint distribution can be written as : fy z γ) = fy z γ)fz)fγ) where z i are multinomial π k ) γ i are N0 Γ) and y i are conditional N S i λ 0 +Λα k + ) γ i ); σ 2 I The joint distribution is now written as : fy z γ) = 1 2π) n+q 2 Γ 1 2 exp 1 2 γt Γ 1 γ ) G { π k exp{ 1 2 n logσ2 )} 16) [ exp 1 y Sλ0 + Λα 2σ 2 k + γ) ) T y Sλ0 + Λα k + γ) )]} zk 17) and the likelihood of the parameters is given by : Lπ k λ 0 Λ α k Γ σ 2 y i z i γ i ) = N 1 exp 1 ) G 2π) n i +q 2 Γ γt i Γ 1 γ i [ exp { π k exp{ 1 2 n i logσ 2 )} 1 2σ 2 yi S i λ 0 + Λα k + γ i ) ) T yi S i λ 0 + Λα k + γ i ) )]} zik 18)
10 In the paper we will use the log likelihood : lπ k λ 0 Λ α k Γ σ 2 y i z i γ i ) = 1 N n i + q) log2π) N G z ik logπ k ) 19) N [ ] log Γ ) + γ T i Γ 1 γ i N 25 EM algorithm 20) G z ik [n i logσ 2 ) + 1 y σ 2 i S i λ 0 + Λα k + γ i ) 2] 21) The EM algorithm consists of iteratively maximizing the expected values of 19) 20) and 21) given y i and the current parameters estimates As these three parts involve separate parameters we can optimize them separately : 251 E step The E step is realized from : } γ i = E {γ i y i λ 0 Λ α Γ σ 2 z ik For the curve i we have the model : Let : y i = S i λ 0 + Λα k + γ i ) + ɛ i u i = y i S i λ 0 + Λα k ) then the joint distribution of u i and of γ i is written as : ) [ ] [ ui 0 SΓS N T + σ 2 I SΓ 0 ΓS T Γ γ i and the conditional distribution of γ i given u i is : ) γ i u i = N γ i ; Σ γi where : and : γ i = S T i S i + σ 2 Γ 1) 1 S T i u i Σ γi = σ 2 S T i S i + σ 2 Γ 1) 1 ]) Then we have got the conditional distribution for γ i y i z ik = 1) : with : γ i y i z ik = 1) N γ i ; Σ γi ) 22) γ i = S T i S i + σ 2 Γ 1) 1 S T i yi S i λ 0 S i Λα k ) 23) Σ γi = σ 2 S T i S i + σ 2 Γ 1) 1 24)
11 252 M step The M step involve maximizing : { } Q = E lπ k λ 0 Λ α k Γ σ 2 y i z i γ i ) holding γ i fixed 253 Estimation of π k The expected value of 19) is maximized by setting : π k = 1 N N π k i 25) with : π k i = P z ik = 1 y i ) = fy z ik = 1)π k G j=1 fy z ij = 1)π j with : fy z ik = 1) given by : y i NS i λ 0 + Λα zi ) Σ i ) where Σ i = σ 2 I + S i ΓS T i 254 Estimation of Γ The expected value of 20) is maximized by setting : Γ = 1 N = 1 N N N ] E [ γ i γ Ti Y i G [ ] E γ i γ T i y i z ik = 1 26) with γ i y i z ik = 1) given by the E step 255 Estimation of λ 0 α k Λ To maximize 21) we need an iterative procedure where λ 0 α k and the columns of Λ are repeatedly optimized while holding all other parameters fixed
12 256 Estimation of λ 0 From the functional model : y i = S i λ 0 + Λα zi + γ i ) + ɛ i i = 1 N we have got by Generalized Least Squares GLS) : λ 0 = N ) 1 S T i S i N S T i [ y i } with γ ik = E {γ ik z ik = 1 y i given by the E step G ] π k i S i Λα k + γ ik ) 27) 257 Estimation of α k The α k are estimated from : N ) 1 N ] α k = π k i Λ T S T i S i Λ π k i Λ T S T i [y i S i λ0 S i γ ik 28) 258 Estimation of Λ By GLS we only have the possibility of estimating vectors and no matrix thus we will have to optimize each column of Λ separately holding all other fixed using : N G ) 1 Λ m = π k i α kms 2 T i S i where : N G π k i α km S T i Λ m is the column m of Λ α km is the component m of α k y i = y i S i λ0 y i G ) α km S i Λl S i γ ik 29) We iterate through 27) 28) 29) until all parameters have converged then we can optimize σ Estimation of σ 2 We have got : σ 2 = 1 N N Let : G l m yi ) T ) ] π k E[ S i Λα k S i γ i yi S i Λα k S i γ i yi z ik = 1 30) σ 2 = 1 N G { yi ) T ) π k S i Λα k S i γ N i yi S i Λα k S i γ i } +S i Cov[γ i y i z ik = 1]S T i 31) The algorithm iterates until all the parameters have converged
13 3 Radial Basis Function Networks Radial Basis Function Networks RBFN) are neural networks used in approximation and classification tasks They share with Multi-Layer Perceptrons the universal approximation property Haykin 1999) Classical RBF networks have their inputs fully connected to non-linear units in a single hidden layer The output of a RBFN is a linear combination of the hidden units outputs More precisely the output is a weighted sum of Gaussian functions or kernels ie the nonlinearities) applied to the inputs : y = I { λ i exp x c i 2 } 32) σ i where x is the input vector y is the scalar output of the RBFN c i 1 i I are the centers of the I Gaussian kernels σ i 1 i I are their widths and λ i 1 i I their weights Intuitively those last λ i parameters represent the relative importance of each kernel in the output y As shown in equation 32) the RBF network has three sets of parameters c i σ i λ i 1 i I One advantage of RBFN networks compared to other approximation models is that these three sets can be learned separately with suitable performances Moreover the learning of the λ i weights results from a linear system A description of learning algorithms for RBF networks can be found in Benoudjit and Verleysen 2003) 4 The Forecasting Method In this section we present a detailed model-based approach for clustering functional data and a time series forecasting method This method will first be sketched to give an intuition of how the forecasting is performed Then each step of the method will be detailed 41 Method Description The forecasting method is based on the "looking in the past" principle Let s the observations on the time interval [t 0 T ] To perform a functional prediction of the curve for the time interval [t t + t out ] we create two functional spaces A first functional space IN is built with past observations for the time interval [t t in t] the "regressors" and a similar second functional space OUT is built with observations for the time interval [t t in t+ t out ] These two spaces are built with all data corresponding to times t [t 0 + t in T t in t out ] These functional spaces are combined into a probabilistic way to build the functional prediction for the time interval [t t+ t out ] and are quantized using the functional clustering algorithm The relationship between the first and the second functional spaces issued from the clustering algorithms is encoded into a probability transition table constructed empirically on the datasets In each of the clusters determined by the second clustering OUT a local RBFN model is built to approximate the relationship between the functional output the local predic-
14 tion) and the functional input the regressor) Finally the global functional prediction at time t for the interval [t t + t out ] is performed by combining the local models results associated to clusters OUT according to their frequencies with respect to the class considered in the cluster IN 42 Quantizing the «inputs» Consider a scalar time series X where xt) is the value at time t t [t 0 T ] This original series is transformed into an array of observations X in for the time intervals [t t + t in ] for all t [t 0 t 0 + t in t t in T t in t out ] Then the clustering algorithm is applied to the input array X in ; after convergence it gives an IN map of K in codewords and the spline coefficients for the curves of each cluster in this IN map 43 Quantizing the «outputs» At each input vector of the matrix X in we aggregate the next observations to get a new array Y out for the time interval [t t + t in + t out ] for all t [t 0 t 0 + t in t t in T t in t out ] The clustering algorithm is applied to the new array Y out ; after convergence it gives an OUT map of K out codewords and the spline coefficients for the curves of each cluster in this OUT map Note that by construction there is a one-to-one relationship between each input and each output vector of spline coefficients 44 Probability transition table Both sets of codewords from maps IN and OUT only contain a static information This information does not reflect completely the evolution of the time series The idea is thus to create a data structure that represents the dynamics of the time series ie how each class of output vectors of spline coefficients including the values for the time interval [t t + t out ]) is associated to each class of input vectors of spline coefficients for the time interval [t t in t] This structure is the probability transition table T i j) with 1 i N in 1 j N out Each element T i j) of this table represents the proportion of output vectors that belongs to the j th class of the OUT map while their corresponding input vectors belong to class i of the IN map Those proportions are computed empirically for the given dataset and sum to one on each line of the table Intuitively the probability transition table represents all the possible evolutions at a given time t together with the probability that they effectively happen 45 Local RBFN models When applied to the «outputs» the functional clustering algorithm provides N out classes and the spline coefficients of the curves for the intervals [t t + t out ] In each of these classes a RBFN model is learned Each RBFN model has p inputs the spline s coefficients of the regressors) and q outputs
15 the spline coefficients of the prediction curve) These models represent the local evolution of the time series restricted to a specific class of regressors The local information provided by these models will be used when predicting the future evolution of the time series 46 Forecasting The relevant information has been extracted from the time series through both maps the probability transition table and the local RBFN models detailed in the previous sections Having this information it is now possible to perform the forecasting itself At each time t the goal is to estimate the functional curve for the time interval [t t + t out ] denoted ˆx[t t + t out ]) First the input at time t is built leading to Xt) This vector is presented to the IN map and the nearest codeword X kt) is identified 1 kt) N in ) In the frequency table in the kt) th line there are some columns corresponding to classes of the OUT map for which the proportions are non zero This means that those columns represent possible evolutions for the considered data Xt) since Xt) has the same shape than data in the kt) th class For each of those potential evolutions the respective RBFN models are considered one RBFN model has been built for each class in the OUT map) For each of them a local prediction ˆx j [t t + t out ]) is obtained 1 j N out ) The final prediction is a weighted sum of the different local predictions the weights being the proportions recorded in the probability transition table The final prediction is thus : ˆx[t t + t out ]) = N out T k j)ˆx j [t t + t out ]) 33) j=1 5 Experimental Results The examples presented here deal with the IBM stock time series of "tick data" for the period starting on January and ending on may with more than 3000 transactions per day on the New York Stock Exchange NYSE) The pricing model will use as inputs inhomogeneous and high-frequency time series of bid and ask prices and also implied volatility Such volatility time series can be obtained from market data derived from option market pricesfrom eight call and put near-the-money nearby and secondary nearby option contracts on the underlying asset)or computed from diverse model assumptions On Fig 1 we can see the evolution of the Prices top) and Volumes bottom) on one day
16 On Fig 2 we see the distribution of transactions for the same day Each point is a transaction with more transactions at the opening and closing of the NYSE The transactions are sampled discretely in time and like it is often the case with financial data the time separating successive observations is itself random On Fig 3 we can see two successive days of the stock IBM with a fine smoothing of the "tick data" by splines 166 TRANSACTIONS on Day TRANSACTIONS between 0930 hr and 1600 hr on Day Prices US $ :00 11:00 12:00 13:00 14:00 15:00 16:00 Time of the Day x Volumes :00 11:00 12:00 13:00 14:00 15:00 16:00 Time of the Day :00 Time of the Day Figure 1: Prices Top) et Volumes Bottom) Figure 2: Distributions of transactions Transactions on Day 10 Transactions on Day Prices US $ Prices US $ Time of the day in Secondes x Time of the day in Secondes x 10 4 Figure 3: Two days of transactions for IBM dashed curve)with smoothing splines solid curve) 51 Prediction We forecast the future transactions splines for three hours of day J between 1030Hr and 1330 Hr t out = 30Hr) from the past transactions days J 2 and J 1 and half an hour of day J between 1000 Hr and 1030 Hr) in this case t in = 1130Hr) We have eliminated the transactions at the opening and closing of the NYSE which are "outliers" without any correlation with the next hours On Fig 4 we can see four out-of-sample forecasting days superposed with the observations and smoothing splines not known by the model) There is a good correlation
17 between the out-of-sample forecasting and the observations Observations Spline and Prediction Observations Spline and Prediction Prices 1525 Prices :00 11:30 12:00 12:30 13:00 11:00 11:30 12:00 12:30 13:00 Observations Spline and Prediction Observations Spline and Prediction Prices 149 Prices :00 11:30 12:00 12:30 13:00 11:00 11:30 12:00 12:30 13:00 Figure 4: Four forecasting days for IBM stock price Observations Points); Smoothing splines solid curve); Out-of-sample forecasting by the model dashed curve) 6 Conclusion We have presented a functional method for the clustering modelling and forecasting of time series by functional analysis and neural networks This method can be applied to any types of time series but is particularly effective when the observations are sparse irregularly spaced occur at different time points for each curve or when only fragments of the curves are observed; standard methods completely fail in these circumstances By the functional clustering we can also realize the forecasting of multiple dynamic processes References Ait-Sahalia Y and PA MylandMarch 2003) The effects of random and discrete sampling when estimating continuous-time diffusions Econometrica Vol 71 pp Benoudjit N and M Verleysen On the kernel widths in Radial-Basis Function Networks Neural Processing Letters Kluwer academic pub vol 18 no 2 pp October 2003
18 Box G and GM Jenkins1970) Time series Forecasting and Control Holden-Day de Boor C 1978) A Practical Guide to Splines New York : Springer DempsterAP NM Laird and DB Rubin1977) Maximum likelihood from incomplete data via the EM algorithm Jounal of the Royal Statistical Society Ser B Vol 39 pp 1-22 Diebold F JH Lee and GWeinbach Regime switching with time-varying transition probabilities in non stationary time series analysis and cointegration ed by C Hargreaves 1994 pp Oxford University Press Engel RF Autoregressive Conditional Heteroscedasticity with Estimates of the Variance of United Kingdom Inflation Econometrica Vol 50 pp Fama EF Efficient Capital Markets : II The Journal of Finance Vol XLVI N 5 pp Hamilton JD1994) Time Series Analysis Princeton University Press Hastie T R Tibshirani and J Friedman 2001) The Elements of Statistical Learning dta Mining Inference and prediction Springer Haykin S1999) Neural Networks - A comprehensive foundation Prentice Hall Franke G R Olsen and W PohlmeierJune 2002) Overview of Forecasting Models University of Konstanz Seminar on High Frequency Finance paper SS 2002 James GM TH Hastie and CA Sugar2000) Principal component models for sparse functional data Biometrika Vol 87 pp James GM CA Sugar2003) Clustering for sparcely sampled functional data Journal of the Americal Association Vol 98 pp Ramsay JO and BW Silverman1997) Functional Data Analysis Springer Series in Statistics Rice JA and CO WuMarch 2001) Nonpoarametric Mixed Effects models for Unequally Sampled Noisy Curves Biometrics Vol 57 pp Olsen RB and MM DacorognaDecember 1992) going Back to the Basics - Rethinking Market Efficiency Olsen and Associates research Group Paper RBO
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht [email protected] 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht [email protected] 539 Sennott
CHARACTERISTICS IN FLIGHT DATA ESTIMATION WITH LOGISTIC REGRESSION AND SUPPORT VECTOR MACHINES
CHARACTERISTICS IN FLIGHT DATA ESTIMATION WITH LOGISTIC REGRESSION AND SUPPORT VECTOR MACHINES Claus Gwiggner, Ecole Polytechnique, LIX, Palaiseau, France Gert Lanckriet, University of Berkeley, EECS,
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
How To Understand The Theory Of Probability
Graduate Programs in Statistics Course Titles STAT 100 CALCULUS AND MATR IX ALGEBRA FOR STATISTICS. Differential and integral calculus; infinite series; matrix algebra STAT 195 INTRODUCTION TO MATHEMATICAL
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
Predict the Popularity of YouTube Videos Using Early View Data
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
STOCK MARKET VOLATILITY AND REGIME SHIFTS IN RETURNS
STOCK MARKET VOLATILITY AND REGIME SHIFTS IN RETURNS Chia-Shang James Chu Department of Economics, MC 0253 University of Southern California Los Angles, CA 90089 Gary J. Santoni and Tung Liu Department
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
Statistical Models in Data Mining
Statistical Models in Data Mining Sargur N. Srihari University at Buffalo The State University of New York Department of Computer Science and Engineering Department of Biostatistics 1 Srihari Flood of
NEURAL NETWORKS A Comprehensive Foundation
NEURAL NETWORKS A Comprehensive Foundation Second Edition Simon Haykin McMaster University Hamilton, Ontario, Canada Prentice Hall Prentice Hall Upper Saddle River; New Jersey 07458 Preface xii Acknowledgments
SYSTEMS OF REGRESSION EQUATIONS
SYSTEMS OF REGRESSION EQUATIONS 1. MULTIPLE EQUATIONS y nt = x nt n + u nt, n = 1,...,N, t = 1,...,T, x nt is 1 k, and n is k 1. This is a version of the standard regression model where the observations
Analysis of Bayesian Dynamic Linear Models
Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main
Forecasting Trade Direction and Size of Future Contracts Using Deep Belief Network
Forecasting Trade Direction and Size of Future Contracts Using Deep Belief Network Anthony Lai (aslai), MK Li (lilemon), Foon Wang Pong (ppong) Abstract Algorithmic trading, high frequency trading (HFT)
Forecasting of Economic Quantities using Fuzzy Autoregressive Model and Fuzzy Neural Network
Forecasting of Economic Quantities using Fuzzy Autoregressive Model and Fuzzy Neural Network Dušan Marček 1 Abstract Most models for the time series of stock prices have centered on autoregressive (AR)
Time Series Analysis
JUNE 2012 Time Series Analysis CONTENT A time series is a chronological sequence of observations on a particular variable. Usually the observations are taken at regular intervals (days, months, years),
Mixtures of Robust Probabilistic Principal Component Analyzers
Mixtures of Robust Probabilistic Principal Component Analyzers Cédric Archambeau, Nicolas Delannay 2 and Michel Verleysen 2 - University College London, Dept. of Computer Science Gower Street, London WCE
Statistical issues in the analysis of microarray data
Statistical issues in the analysis of microarray data Daniel Gerhard Institute of Biostatistics Leibniz University of Hannover ESNATS Summerschool, Zermatt D. Gerhard (LUH) Analysis of microarray data
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Probabilistic Forecasting of Medium-Term Electricity Demand: A Comparison of Time Series Models
Fakultät IV Department Mathematik Probabilistic of Medium-Term Electricity Demand: A Comparison of Time Series Kevin Berk and Alfred Müller SPA 2015, Oxford July 2015 Load forecasting Probabilistic forecasting
A FUZZY LOGIC APPROACH FOR SALES FORECASTING
A FUZZY LOGIC APPROACH FOR SALES FORECASTING ABSTRACT Sales forecasting proved to be very important in marketing where managers need to learn from historical data. Many methods have become available for
THE SVM APPROACH FOR BOX JENKINS MODELS
REVSTAT Statistical Journal Volume 7, Number 1, April 2009, 23 36 THE SVM APPROACH FOR BOX JENKINS MODELS Authors: Saeid Amiri Dep. of Energy and Technology, Swedish Univ. of Agriculture Sciences, P.O.Box
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
Using Mixtures-of-Distributions models to inform farm size selection decisions in representative farm modelling. Philip Kostov and Seamus McErlean
Using Mixtures-of-Distributions models to inform farm size selection decisions in representative farm modelling. by Philip Kostov and Seamus McErlean Working Paper, Agricultural and Food Economics, Queen
Factor analysis. Angela Montanari
Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number
New Ensemble Combination Scheme
New Ensemble Combination Scheme Namhyoung Kim, Youngdoo Son, and Jaewook Lee, Member, IEEE Abstract Recently many statistical learning techniques are successfully developed and used in several areas However,
Chapter 4: Vector Autoregressive Models
Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
Threshold Autoregressive Models in Finance: A Comparative Approach
University of Wollongong Research Online Applied Statistics Education and Research Collaboration (ASEARC) - Conference Papers Faculty of Informatics 2011 Threshold Autoregressive Models in Finance: A Comparative
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its
Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement
Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement Toshio Sugihara Abstract In this study, an adaptive
Univariate and Multivariate Methods PEARSON. Addison Wesley
Time Series Analysis Univariate and Multivariate Methods SECOND EDITION William W. S. Wei Department of Statistics The Fox School of Business and Management Temple University PEARSON Addison Wesley Boston
Exploratory Data Analysis Using Radial Basis Function Latent Variable Models
Exploratory Data Analysis Using Radial Basis Function Latent Variable Models Alan D. Marrs and Andrew R. Webb DERA St Andrews Road, Malvern Worcestershire U.K. WR14 3PS {marrs,webb}@signal.dera.gov.uk
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
1 Teaching notes on GMM 1.
Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in
State Space Time Series Analysis
State Space Time Series Analysis p. 1 State Space Time Series Analysis Siem Jan Koopman http://staff.feweb.vu.nl/koopman Department of Econometrics VU University Amsterdam Tinbergen Institute 2011 State
The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.
Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium
Maximum likelihood estimation of mean reverting processes
Maximum likelihood estimation of mean reverting processes José Carlos García Franco Onward, Inc. [email protected] Abstract Mean reverting processes are frequently used models in real options. For
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression Saikat Maitra and Jun Yan Abstract: Dimension reduction is one of the major tasks for multivariate
A Simple Model for Intra-day Trading
A Simple Model for Intra-day Trading Anton Golub 1 1 Marie Curie Fellow, Manchester Business School April 15, 2011 Abstract Since currency market is an OTC market, there is no information about orders,
Optimization of technical trading strategies and the profitability in security markets
Economics Letters 59 (1998) 249 254 Optimization of technical trading strategies and the profitability in security markets Ramazan Gençay 1, * University of Windsor, Department of Economics, 401 Sunset,
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
1 Short Introduction to Time Series
ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING. ----Changsheng Liu 10-30-2014
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING ----Changsheng Liu 10-30-2014 Agenda Semi Supervised Learning Topics in Semi Supervised Learning Label Propagation Local and global consistency Graph
Neural Network Add-in
Neural Network Add-in Version 1.5 Software User s Guide Contents Overview... 2 Getting Started... 2 Working with Datasets... 2 Open a Dataset... 3 Save a Dataset... 3 Data Pre-processing... 3 Lagging...
Knowledge Discovery in Stock Market Data
Knowledge Discovery in Stock Market Data Alfred Ultsch and Hermann Locarek-Junge Abstract This work presents the results of a Data Mining and Knowledge Discovery approach on data from the stock markets
Volatility modeling in financial markets
Volatility modeling in financial markets Master Thesis Sergiy Ladokhin Supervisors: Dr. Sandjai Bhulai, VU University Amsterdam Brian Doelkahar, Fortis Bank Nederland VU University Amsterdam Faculty of
Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts
Page 1 of 20 ISF 2008 Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts Andrey Davydenko, Professor Robert Fildes [email protected] Lancaster
NTC Project: S01-PH10 (formerly I01-P10) 1 Forecasting Women s Apparel Sales Using Mathematical Modeling
1 Forecasting Women s Apparel Sales Using Mathematical Modeling Celia Frank* 1, Balaji Vemulapalli 1, Les M. Sztandera 2, Amar Raheja 3 1 School of Textiles and Materials Technology 2 Computer Information
Automated Stellar Classification for Large Surveys with EKF and RBF Neural Networks
Chin. J. Astron. Astrophys. Vol. 5 (2005), No. 2, 203 210 (http:/www.chjaa.org) Chinese Journal of Astronomy and Astrophysics Automated Stellar Classification for Large Surveys with EKF and RBF Neural
Learning outcomes. Knowledge and understanding. Competence and skills
Syllabus Master s Programme in Statistics and Data Mining 120 ECTS Credits Aim The rapid growth of databases provides scientists and business people with vast new resources. This programme meets the challenges
Prediction of Stock Performance Using Analytical Techniques
136 JOURNAL OF EMERGING TECHNOLOGIES IN WEB INTELLIGENCE, VOL. 5, NO. 2, MAY 2013 Prediction of Stock Performance Using Analytical Techniques Carol Hargreaves Institute of Systems Science National University
Machine Learning in Statistical Arbitrage
Machine Learning in Statistical Arbitrage Xing Fu, Avinash Patra December 11, 2009 Abstract We apply machine learning methods to obtain an index arbitrage strategy. In particular, we employ linear regression
Introduction to Engineering System Dynamics
CHAPTER 0 Introduction to Engineering System Dynamics 0.1 INTRODUCTION The objective of an engineering analysis of a dynamic system is prediction of its behaviour or performance. Real dynamic systems are
Statistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural
Introduction to time series analysis
Introduction to time series analysis Margherita Gerolimetto November 3, 2010 1 What is a time series? A time series is a collection of observations ordered following a parameter that for us is time. Examples
11.2 Monetary Policy and the Term Structure of Interest Rates
518 Chapter 11 INFLATION AND MONETARY POLICY Thus, the monetary policy that is consistent with a permanent drop in inflation is a sudden upward jump in the money supply, followed by low growth. And, in
Statistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
The term structure of Russian interest rates
The term structure of Russian interest rates Stanislav Anatolyev New Economic School, Moscow Sergey Korepanov EvrazHolding, Moscow Corresponding author. Address: Stanislav Anatolyev, New Economic School,
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
PS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng [email protected] The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
A Regime-Switching Model for Electricity Spot Prices. Gero Schindlmayr EnBW Trading GmbH [email protected]
A Regime-Switching Model for Electricity Spot Prices Gero Schindlmayr EnBW Trading GmbH [email protected] May 31, 25 A Regime-Switching Model for Electricity Spot Prices Abstract Electricity markets
Data Mining mit der JMSL Numerical Library for Java Applications
Data Mining mit der JMSL Numerical Library for Java Applications Stefan Sineux 8. Java Forum Stuttgart 07.07.2005 Agenda Visual Numerics JMSL TM Numerical Library Neuronale Netze (Hintergrund) Demos Neuronale
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Time Series Analysis
Time Series Analysis Forecasting with ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos (UC3M-UPM)
Trading activity as driven Poisson process: comparison with empirical data
Trading activity as driven Poisson process: comparison with empirical data V. Gontis, B. Kaulakys, J. Ruseckas Institute of Theoretical Physics and Astronomy of Vilnius University, A. Goštauto 2, LT-008
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Adaptive Equalization of binary encoded signals Using LMS Algorithm
SSRG International Journal of Electronics and Communication Engineering (SSRG-IJECE) volume issue7 Sep Adaptive Equalization of binary encoded signals Using LMS Algorithm Dr.K.Nagi Reddy Professor of ECE,NBKR
A Learning Algorithm For Neural Network Ensembles
A Learning Algorithm For Neural Network Ensembles H. D. Navone, P. M. Granitto, P. F. Verdes and H. A. Ceccatto Instituto de Física Rosario (CONICET-UNR) Blvd. 27 de Febrero 210 Bis, 2000 Rosario. República
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
Note 2 to Computer class: Standard mis-specification tests
Note 2 to Computer class: Standard mis-specification tests Ragnar Nymoen September 2, 2013 1 Why mis-specification testing of econometric models? As econometricians we must relate to the fact that the
Machine Learning in FX Carry Basket Prediction
Machine Learning in FX Carry Basket Prediction Tristan Fletcher, Fabian Redpath and Joe D Alessandro Abstract Artificial Neural Networks ANN), Support Vector Machines SVM) and Relevance Vector Machines
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
ADVANCED FORECASTING MODELS USING SAS SOFTWARE
ADVANCED FORECASTING MODELS USING SAS SOFTWARE Girish Kumar Jha IARI, Pusa, New Delhi 110 012 [email protected] 1. Transfer Function Model Univariate ARIMA models are useful for analysis and forecasting
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
Genetic algorithms for solving portfolio allocation models based on relative-entropy, mean and variance
Journal of Scientific Research and Development 2 (12): 7-12, 2015 Available online at www.jsrad.org ISSN 1115-7569 2015 JSRAD Genetic algorithms for solving portfolio allocation models based on relative-entropy,
The CUSUM algorithm a small review. Pierre Granjon
The CUSUM algorithm a small review Pierre Granjon June, 1 Contents 1 The CUSUM algorithm 1.1 Algorithm............................... 1.1.1 The problem......................... 1.1. The different steps......................
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE YUAN TIAN This synopsis is designed merely for keep a record of the materials covered in lectures. Please refer to your own lecture notes for all proofs.
On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information
Finance 400 A. Penati - G. Pennacchi Notes on On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information by Sanford Grossman This model shows how the heterogeneous information
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images Małgorzata Charytanowicz, Jerzy Niewczas, Piotr A. Kowalski, Piotr Kulczycki, Szymon Łukasik, and Sławomir Żak Abstract Methods
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.
AUTOMATION OF ENERGY DEMAND FORECASTING by Sanzad Siddique, B.S. A Thesis submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment of the Requirements for the Degree
Working Papers. Cointegration Based Trading Strategy For Soft Commodities Market. Piotr Arendarski Łukasz Postek. No. 2/2012 (68)
Working Papers No. 2/2012 (68) Piotr Arendarski Łukasz Postek Cointegration Based Trading Strategy For Soft Commodities Market Warsaw 2012 Cointegration Based Trading Strategy For Soft Commodities Market
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Visualization of textual data: unfolding the Kohonen maps.
Visualization of textual data: unfolding the Kohonen maps. CNRS - GET - ENST 46 rue Barrault, 75013, Paris, France (e-mail: [email protected]) Ludovic Lebart Abstract. The Kohonen self organizing
Multivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
Econometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
System Identification for Acoustic Comms.:
System Identification for Acoustic Comms.: New Insights and Approaches for Tracking Sparse and Rapidly Fluctuating Channels Weichang Li and James Preisig Woods Hole Oceanographic Institution The demodulation
No-Arbitrage Condition of Option Implied Volatility and Bandwidth Selection
Kamla-Raj 2014 Anthropologist, 17(3): 751-755 (2014) No-Arbitrage Condition of Option Implied Volatility and Bandwidth Selection Milos Kopa 1 and Tomas Tichy 2 1 Institute of Information Theory and Automation
