Prediction of Stock Market Index Movement by Ten Data Mining Techniques

Size: px
Start display at page:

Download "Prediction of Stock Market Index Movement by Ten Data Mining Techniques"

Transcription

1 Vol. 3, o. Modern Appled Scence Predcton of Stoc Maret Index Movement by en Data Mnng echnques Phchhang Ou (Correspondng author) School of Busness, Unversty of Shangha for Scence and echnology Rm 0, Internatonal Exchange Center, o. 56, Jun Gong Road, Shangha 00093, Chna el: , Fax: E-mal: Hengshan Wang School of Busness, Unversty of Shangha for Scence and echnology Box 46, o. 56, Jun Gong Road, Shangha 00093, Chna el: E-mal: hs wor s supported by Shangha Leadng Academc Dscplne Proect, Proect umber: S Abstract Ablty to predct drecton of stoc/ndex prce accurately s crucal for maret dealers or nvestors to maxmze ther profts. Data mnng technques have been successfully shown to generate hgh forecastng accuracy of stoc prce movement. owadays, n stead of a sngle method, traders need to use varous forecastng technques to gan multple sgnals and more nformaton about the future of the marets. In ths paper, ten dfferent technques of data mnng are dscussed and appled to predct prce movement of Hang Seng ndex of Hong Kong stoc maret. he approaches nclude Lnear dscrmnant analyss (LDA), Quadratc dscrmnant analyss (QDA), K-nearest neghbor classfcaton, aïve Bayes based on ernel estmaton, Logt model, ree based classfcaton, neural networ, Bayesan classfcaton wth Gaussan process, Support vector machne (SVM) and Least squares support vector machne (LS-SVM). Expermental results show that the SVM and LS-SVM generate superor predctve performances among the other models. Specfcally, SVM s better than LS-SVM for n-sample predcton but LS-SVM s, n turn, better than the SVM for the out-of-sample forecasts n term of ht rate and error rate crtera. Keywords: Data mnng, Stoc prce movement predcton, SVM, LS-SVM,, Bayesan classfcaton wth Gaussan processes. Introducton Fnancal maret s a complex, nonstatonary, nosy, chaotc, nonlnear and dynamc system but t does not follow random wal process, (Lo & Macnlay, 988; Deng, 006). here are many factors that may cause the fluctuaton of fnancal maret movement. he man factors nclude economc condton, poltcal stuaton, traders expectatons, catastrophes and other unexpected events. herefore, predctons of stoc maret prce and ts drecton are qute dffcult. In response to such dffculty, data mnng (or machne learnng) technques have been ntroduced and appled for ths fnancal predcton. Most of the studes have focused on the accurate forecastng of the value of stoc prce. However, dfferent nvestors adopt dfferent tradng strateges; therefore, the forecastng model based on mnmzng the error between the actual values and the forecasts may not be sutable for them. In stead, accurate predcton of movement drecton of stoc ndex s crucal for them to mae effectve maret tradng strateges. Some recent studes have suggested that tradng strateges llustrated by the forecasts based on the drecton of stoc prce change may be more effectve and generate hgher proft. Specfcally, nvestors could effectvely hedge aganst potental maret rs and speculators as well as arbtrageurs could have opportunty of mang proft by tradng stoc ndex whenever they could obtan the accurate predcton of stoc prce drecton. hat s why there have been a number of studes loong at drecton or trend of movement of varous nds of fnancal nstruments (such as Wu & Zhang, 997; O connor et al, 997). But, these studes do not use data mnng based classfcaton technques. Data mnng technques have been 8

2 Modern Appled Scence December, 009 ntroduced for predcton of movement sgn of stoc maret ndex snce the results of Leung et al (000) and Chen et al (00), where LDA, Logt and Probt and eural networ were proposed and compared wth parametrc models, GMM-Kalman flter. Km (003) appled newly and powerful technques of data mnng, SVM and eural networ, to forecast the drecton of stoc ndex prce based on economc ndcators. o obtan more profts from the stoc maret, more and more best forecastng technques are used by dfferent traders. In stead of a sngle method, the traders need to use varous forecastng technques to gan multple sgnals and more nformaton about the future of the marets. Kumar & henmozh (006) collected fve dfferent approaches ncludng SVM, Random forecast, eural networ, Logt and LDA to predct Indan stoc ndex movement based on economc varable ndcators. From the comparson, the SVM outperformed the others n forecastng S&P CX IFY ndex drecton as the model does not requre any pror assumptons on data property and ts algorthm results global optmal soluton whch s unque. Huang et al (005) also forecasted the movement drecton of Japanese stoc maret (IKKEI 5 ndex) by varous technques such as SVM, LDA, QDA, and the all-n-one combned approach. he SVM approach also gves better predctve capablty than other models: LDA, QDA and, followng the out-performance of the combned model. In the study, they defned the movement of the IKKEI 5 ndex based on two man factors ncludng Amercan stoc maret, S&P 500 ndex, whch s the most nfluence on the world stoc marets ncludng Japanese maret, and the currency exchange rate between Japanese Yen and US dollar. In our study, ten dfferent technques of data mnng are dscussed and appled to predct prce movement of Hang Seng ndex of Hong Kong stoc maret. he approaches nclude Lnear dscrmnant analyss (LDA), Quadratc dscrmnant analyss (QDA), K-nearest neghbor classfcaton, aïve Bayes based on ernel estmaton, Logt model, ree based classfcaton, neural networ, Bayesan classfcaton wth Gaussan process, Support vector machne (SVM) and Least squares support vector machne (LS-SVM). he man goal s to explore the predctve ablty of the ten data-mnng technques n forecastng movement drecton of Hang Seng Index based on fve factors, ncludng ts open prce, hgh prce, low prce, S&P 500 ndex, and currency exchange rate between HK dollar and US dollar. he general model of stoc prce movement s defned as D t f ( Ot, Ht, Lt, S & P500t, FX t ), where Dt s the drecton of HSI movement at tme t and s defned as a categorcal value f the closng prce at tme t s greater than the closng prce at tme t and as 0, otherwse. he functon f (.) can be lnear or nonlnear and t s estmated by the ten data-mnng algorthms. O denotes the open prce of HSI at tme t ; t H s the hgh prce of HSI at tme t ; t L s the low prce n a day of HSI at tme t ; t S & P500 s the closng prce of S&P 500 ndex at tme t. t FX t s the currency exchange rate between HK dollar and US dollar. All the nputs are transformed nto log return to remove any trend pattern. he remanng of the paper s organzed as follow. ext sectons descrbe the data and predcton evaluaton. Secton 3 brefly dscusses the ten dfferent algorthms. he fnal secton s for concluson.. Data descrpton We examne the daly change of closng prces of Hang Seng ndex based on fve predctors, Open prce, Hgh prce, Low prce, S&P 500 ndex prce, and Exchange rate USD aganst HKD. he stoc prces are downloaded from the Yahoo fnance and the foregn exchange rate s taen from webste of Federal Reserve Ban of S. Lous. he sample perod s from Jan to Dec so that the whole sample s of 73 tradng days. he data s dvded nto two sub-samples where the n-sample or tranng data spans from Jan to Dec wth 48 tradng days. he whole year 006 from Jan 006 to Dec of sze 50 tradng days are reserved for out-of-sample or test data. Fgure dsplays the actual movement of HSI closng prces for the whole sample. Fgure plots S&P 500 prce and ts log return and Fgure 3 shows the plots of prce and log return of exchange rate of HKD aganst USD. o measure the predctve performances by dfferent models, Ht rate and Error rate are employed and defned as Ht rate = m m I[ A P ] and Error rate = I[ A ] m P m where A s the actual output for th tradng day and P s the predcted value for th tradng day, obtaned from each model. Here m s the number of the out-of-sample. R software wth related pacages s used to conduct the whole experment; for example Karatzoglou (004, 006) llustrates the R commands for SVM, LSSVM, and Bayesan classfcaton wth Gaussan Processes. 9

3 Vol. 3, o. Modern Appled Scence 3. Data-mnng methods Let the tranng data be { ( x, y),, ( x n, yn ) }where X ( X,, X p ) denote real-valued random nput vector and the response Y s categorcal,.e. Y {,,, K }. he goal s to form a predctor G (x) to predct Y based on X. So G(x) dvdes the nput space (or feature vector space) nto a collecton of regons where each labeled by one class. For bnary class problem, Y {, }, the decson boundary between the two classes s a hyperplane n the feature p space. A hyperplane n the p- dmensonal nput space s the set: x : x 0 }. () p { 0 { 0 p { 0 he two regons separated by the hyperplane are x : x 0 } and x : x 0 }. ow we defne Bayes classfcaton rule. Suppose the tranng data { ( x, y),, ( x n, yn ) }are ndependent samples from the ont dstrbuton of X and Y, f X, Y ( x, y) py ( y) f X / Y ( x / Y y) and the loss functon of classfyng Y as G( X ) Yˆ s L ( Y ˆ, Y ), where the margnal dstrbuton of Y s specfed by the pmf p Y (y) and f X / Y ( x / Y y) s the condtonal dstrbuton of X gven Y y. he goal of classfcaton s to mnmze the expected loss defned as EX, Y L( G( X ), Y ) EX [ EY / X L( G( X ), Y )]. o mnmze the left hand sde of the expected loss, t suffces to mnmze E Y / X L( G( X ), Y ) for each X. Hence the optmal classfer s G( X ) arg mn EY / X xl( y, Y ). For 0- loss y functon L( y, y) 0 for y y and 0 otherwse, we have EY / X xl( y, Y ) Pr( Y y / X x). herefore, the classfcaton rule, called Bayes rule, becomes the rule of maxmum a posteror probablty: G( x) arg mn E / L( y, Y ) = arg max Pr( Y y / X x). y Y X x y We consder ten algorthms for classfcaton where some of them attempt to estmate Pr( Y y / X x) and then apply the Bayes rule G( x) arg max Pr( Y y / X x).he algorthms nclude lnear dscrmnant analyss (LDA), y Quadratc dscrmnant analyss (QDA), aïve Bayes based on ernel estmaton, and Logt model. Other types of data mnng technques are K-nearest neghbor classfcaton, ree based classfcaton, neural networ, Bayesan classfcaton wth Gaussan process, Support vector machne (SVM) and Least squares support vector machne (LS-SVM). he last three models tae more advantage va ernel based methods. 3. Lnear Dscrmnant Analyss he goal here s to obtan class posterors Pr( Y / X ) for optmal classfcaton. Suppose f (x) s the class-condtonal K densty of X n class Y K and let be the pror probablty of class K wth. By Bayes theorem, K Pr( Y / X x) f ( x) / p / f ( x). Suppose each class densty s from Gaussan dstrbuton defned / as f ( x) ( ) exp( ( x ) ( ) x and the classes are assumed to have a common covarance matrx. Consderng the log rato of the two classes and l posterors log [Pr( Y / X x) / Pr( Y l / X x) ] = log [ f ( x) / fl ( x) log [ / l ] = log [ / l ] ( ) ( ) ( ) l l x l whch s a lnear equaton n x n p dmensonal hyperplane defned n (), the lnear dscrmnant functons are obtaned ( x) x log and the LDA classfer s G( x) arg max ( x). From the tranng data, we can estmate the Gaussan dstrbuton parameters as ˆ / where s the number of class observatons; ˆ y x / and ˆ x ˆ K ( ˆ )( ˆ y x x ) /( K). For two classes {, }, the LDA rule classfes to class f ( ˆ ˆ ) ˆ ˆ ˆ ˆ ˆ ˆ log( / ) log( / ) and class otherwse. () From the expermental result, we have X ( X, X 5) wth dmenson of p =5 so that fve coeffcents of the lnear dscrmnant n () are obtaned as the followng: V = , V = , V3 = , V4 = , V5 = Pror probabltes of groups: 0 and are and respectvely. For the n-sample data, the ht rate s and the error rate s 0.607, whle n the out-of-sample data, the ht rate and error rate are and respectvely. 30

4 Modern Appled Scence December, Quadratc dscrmnant analyss For the QDA, the above are not assumed equal for each, then ( x) log ( x ) ( x ) log. he decson boundary between two classes and l s the quadratc equaton { x : ( x) l ( x) }. he estmates for QDA are smlar to those LDA except that separate covarance matrx must be estmated for each class. Here QDA needs ( K ){ p( p 3) / } parameters. See McLachlan (99) and Duda et al (000) for comprehensve dscusson on dscrmnant analyss. From the predcton results, the ht rate and error rate for tranng data by QDA are and respectvely. For the test data, the ht rate and error rate are and 0.50 respectvely. 3.3 K-nearest neghborhood method K-nearest neghbor method s one of the smplest machne learnng algorthms used for classfyng obects based on closest tranng examples n the feature space. An obect s classfed by a maorty beng assgned to the class most common amongst ts nearest neghbors. Formally, the -nearest neghbor approach uses the tranng data set { ( x, y),, ( x n, yn ) }closest n nput space to x to form Yˆ. Specfcally, the -nearest neghbor ft for Yˆ s defned as yˆ x ( ) x x y ( ) where (x) s the neghborhood of x defned by closest ponts x n the tranng example. hat s, we fnd the observatons wth x closest to x n nput space and average ther responses. s estmated by cross-valdaton technque. he algorthm starts wth the determnaton of the optmal based on RMSE done by cross valdaton technque, then calculate the dstance between the query dstance and all the tranng samples. After sortng the dstance and determnaton of the nearest neghbors based on the th mnmum dstance, gather the Y of the nearest neghbors. Fnally, use smple maorty of the category Y of nearest neghbors as the predcton value of the query dstance. otceably, the -nearest neghbor approach does not rely on pror probabltes le LDA and QDA. Results: able dsplays the process of choosng by cross-valdaton technque n the experment. We consder = 30 as an ntal range and then select optmal n the range. he best = 0 s obtaned correspondng to smallest error (0.708). [Insert able around here]. he performance results are gven as the followng. For the tranng data, the ht rate s 0.83 and the error rate s and for the test data, the ht rate s and the error rate s aïve Bayes classfcaton method hs s a well establshed Bayesan method prmarly formulated for performng classfcaton tass. Gven ts smplcty,.e., the assumpton that the ndependent varables are statstcally ndependent, ave Bayes models are effectve classfcaton tools that are easy to use and nterpret. ave Bayes s partcularly approprate when the dmensonalty of the feature space (.e., number of nput varables) s hgh (a problem nown as the curse of dmensonalty). Mathemcally, aïve Bayes model requres an assumpton that gven a class Y, the features X are ndependent so p that f ( X ) f ( X ). he estmates of f (.) s from the tranng data va ernel smoothng. aïve Bayes classfcaton s G( x) arg max f ( X ) and s estmated by the sample proportons. See Mtchell (005) for precse explanaton. From the result obtaned n the experment, the ht rate and error rate for n-sample data are and 0.64 respectvely. For the out-of-sample, the ht rate s and the error rate s Logt model Logstc regresson refers to methods for descrbng the relatonshp between a categorcal response varable and a set of predctor varables. It can be used to predct a dependent varable on the bass of ndependents and to determne the percent of varance n the dependent varable explaned by the ndependents. he logstc regresson apples maxmum lelhood estmaton after transformng the dependent nto a logt varable. In ths way, logstc regresson estmates the probablty of a certan event occurrng. ote that logstc regresson calculates changes n the log odds of the dependent, not changes n the dependent tself. From the Fredman et al (008), the model for logstc regresson s gven p exp( 0 X ) as: ( x) Pr( Y / X x) for two classes of output Y. We obtan s usng the p exp( 0 X ) maxmum lelhood approach. ( x) Pr( Y / X x) p Logt s gven by: G( x) log log 0 X. he curve of (x) are called ( x) Pr( Y 0 / X x) sgmod because they are S-shape and therefore nonlnear. Statstcans have chosen the logstc dstrbuton to model a a bnary data because of ts flexblty and nterpretablty. he mnmum for (x) s attaned at lm e /( e ) 0, and the maxmum for (x) s obtaned at lm e /( e ). a a a a 3

5 Vol. 3, o. Modern Appled Scence From the experment, the coeffcents of s estmated by maxmum lelhood are obtaned as , -.385,.6697, 3.49, and he performance results show that the ht rate and error rate for n-sample data s and 0.56 respectvely whle the ht rate and error rate for out-of-sample data are and respectvely. 3.6 ree based classfcaton Classfcaton tree s one of the man technques used n Data Mnng. Classfcaton trees are used to predct membershp of obects n the classes of a categorcal dependent varable from ther measurements on one or more predctor varables. he goal of classfcaton trees s to predct or explan responses on a categorcal dependent varable, and as such, the avalable technques have much n common wth the technques used n the more tradtonal methods of Dscrmnant Analyss, Cluster Analyss, onparametrc Statstcs, and onlnear Estmaton. he flexblty of classfcaton trees maes them a very attractve analyss opton as t does not requre any assumpton on the dstrbuton le tradtonal statstcal methods. echncally, the ree-based methods partton the feature space nto a set of rectangles, and then ft a smple model n each one. Startng from the root of a tree, the feature space contanng all examples s splt recursvely nto subsets usually two at a tme. Each splt depends on the value of only a unque varable of nput x. If x s categorcal, the splt s of the form x A or x A where A s subset of. he goodness of splt s measured by an mpurty functon defned for each node. he basc dea s to choose a splt such that the chld nodes are purer than ther parent node. he splt contnues tll the end subsets (leaf nodes) are pure ; that s tll one class domnates. For an mpurty functon, defne the mpurty measure by ( t) (Pr(/ t), Pr( / t),, Pr( K / t)) where Pr( / t) s the estmated probablty of class wthn node t. he goodness of a splt s for node t s ( s, t) I( s, t) ( t) pr( tr ) pl( tl ) where pr and pl are the proportons of the samples n node t that go to the rght node and left node respectvely. Possble mpurty functons: a. Entropy: ( t) Pr( / t) log(/ Pr( / t)) 3 K K b. Gn ndex: ( t) Pr( / t)( Pr( / t)) where Pr( / t) I[ x ], I [ ] ( t) x f x and 0 otherwse, xt and (t) s the total number of samples n node t. he crtera s to stop splttng a node t when max I( s, t), s the chosen threshold. It s not trval to choose as t leads to overfttng or underfttng problems for a new data predcton. o solve ths problem, we go for prunng approach. he dea s to obtan the subtree from the ntal large tree. One of the most popular technques s the Cost-complexty prunng. Let max be ntal large tree and let the ~ pruned subtree by. hen the cost complexty measure C ( ) s defned as C ( ) R( ), max ~ denotes the number of leaf nodes. R( ) Pr( t) r( t) s the error measure of, r( t) max Pr( / t) and ~ t 0 s the complexty parameter. Here C ( ) represents the tradeoff between the cost of a tree and ts complexty. he goal of cost-complexty prunng s for each choose a tree ( ) max such that C ( ) s mnmzed. he estmaton of s acheved by crossvaldaton. We choose ˆ that mnmzes the cross-valdaton sum of squares. hus the fnal tree s (ˆ ). ree based method s consdered one of top ten algorthm of data mnng technque (Wu et al, 008). We refer to Qunlan (986) and Fredman et al (008) for detaled dscusson on the tree based classfcaton. For llustratve applcaton of ree method n predctng stoc prce behavor s referred to Pearson (004). Expermental results: Denote varables actually used n tree constructon by V = Open prce, V = Low prce, V3=Hgh prce, V4= S&P500, V5 = FX, V6 = Close prce. he class s when the next prce s larger than the prevous prce, whle the class 0 s refers to when the next prce s smaller than the prevous prce. able show the process of prunng tree. From the able, we choose CP = correspondng to 7 splts based on the smallest value of the x_error: he ntal large tree s not dsplayed to reduce the space but pruned tree s llustrated n the appendx secton. he performance result shows that the ht rate s and error rate s 0.83 n the tranng data; whle the test data, the ht rate s 0.8 and the error rate s eural networ for classfcaton Hayn (994) defnes neural networ as a massvely parallel dstrbuted processor that has a natural propensty for storng experental nowledge and mang t avalable for use. It resembles the bran n two respects: () Knowledge s acqured by the networ through a learnng process, and () Interneuron connecton strengths nown as synaptc weghts are used to store the nowledge. he lterature on neural networ s enormous, and ts applcaton spreads over ss

6 Modern Appled Scence December, 009 many scentfc areas see Bshop (995) and Rpley (996) for detaled. Recently, the neural networ has been well nown for good capablty n forecastng stoc maret movement. Let s go drectly to ts formulaton. Let X be nput vector and Y be output tang value as categorcal. Followng notatons n Haste (996), the neural networ model can be represented as z ( 0 x),,, m yˆ f ( 0 z),,, q z where ( z) /( e ) actvaton functon called sgmod. he parameters l and are nown as weghts and v 0 and 0 are bas. Here we use f ( v) /( e ) the nverse logt for bnary classfcaton. o learn the neural networ, bac propagaton s used. Suppose we use least squares on a sample of tranng data to learn the weghts: ( y ) R(, ) yˆ, R ( y yˆ ) f ( z ) z, R l he th component s denoted by superscrpt. Gradent update at the (r+)st teraton, ( ) K ( y yˆ ) f ( z ) ( x ) xl. ( r) R r ( r) ( r ) ( r) R l l r. ( r) l Here r s the learnng rate. In the expermental analyss, l,, 5 ;, ;, so that a 5-- networ wth 8 weghts are obtaned wth the followng estmated weghts: From the predcton performances, the ht rate s and error rate s 0.59 n the n-sample data. For the out-of-sample, and are ht rate and error rate respectvely. 3.8 Bayesan classfcaton for Gaussan process Gaussan processes are based on the pror assumpton that adacent observatons should convey nformaton about each other. Partcularly, t s assumed that the observed varables follow normal dstrbuton and that the couplng between them taes place by covarance matrx of a normal dstrbuton. Usng the ernel matrx as the covarance matrx s a convenent way of extendng Bayesan modelng of lnear estmators to nonlnear stuatons. In regresson problem, the goal s to predct a real valued output based on a set of nput varables. It s possble to carry out nonparametrc regresson usng Gaussan process. Wth Gaussan pror and Gaussan nose model, the soluton of the regresson problem can be obtaned va Kernel functon placed on each tranng data; the coeffcents are determned by solvng a lnear system. If the parameter ndexed n the Gaussan process are unnown, Bayesan nference can be carred out for them. Gaussan process can be extended to classfcaton problems by defnng a Gaussan process over y, the nput to the sgmod functon. he goal s to predct Pr( Y / X x),,, K. For bnary case, 0,, Pr( Y / X x) s estmated by ( y( x)) where ( y) /( e ). he dea s to place the Gaussan pror on y(x) and combne t wth the tranng data D ( x, t ),,, n to obtan predctons for new x ponts. Bayesan treatment s mposed by ntegratng over uncertanty n y and n the parameters that control the Gaussan pror. hen y 33

7 Vol. 3, o. Modern Appled Scence the Laplace s approxmaton s employed to obtan the results of the ntegraton over y. Specfcally, let Pr(y) be the pror of y ( y( x ), y( x),, y( xn )) so that Pr( y, y) s the ont dstrbuton ncludng y. Gven new nput x, we want to predct y y( x ) based on the D ( x, t ),,, n. Let Pr( t / y) be the probablty of observng the partcular values t t, t,, ) gven actual values y (.e., nose model). hen we have ( t n Pr( y / t) Pr( y, y / t) dy Pr( t) Pr( y, y) Pr( y) Pr( t / y) dy Pr( y, y) Pr( y / t) dy. Hence the predctve dstrbuton for y s found from the margnalzaton of the product of the pror and the nose model. he ntegral terms are estmated by Laplace s approxmaton. Wllam et al (999) and Rasmussen et al (006) provded a comprehensve and detaled dscusson on the Bayesan classfcaton wth Gaussan processes. he followng shortly descrbes the expermental results obtaned from tranng and forecastng the movement of HSI by Bayesan classfcaton wth Gaussan processes. Problem type: classfcaton Gaussan Radal Bass ernel functon hyperparameter: sgma = umber of tranng nstances learned: 48 ran error: Cross valdaton error: In-sample data: Ht rate s and error rate s Out-of-sample data: Ht rate s and error rate s Support vector machne for classfcaton A popular machne learnng algorthm wth neural networ type s SVM, support vector machne, developed by Vapn (995). he SVM s a ernel based learnng approach le the above Gaussan processes for classfcaton. However, the SVM does not requre any assumptons on the data property le n Gaussan processes. he SVM has been successfully appled for varous areas of predctons; for nstance, n fnancal tme seres forecastng (Muheree et al, 997; ay and Cao, 00), maretng (Bend-Davd and Lndenbaum, 997), estmatng manufacturng yelds (Stoneng, 999), text categorzaton (Joachms, 00), face detecton usng mage (Osuna et al, 997), handwrtten dgt recognton (Burges and Schoopf, 997); Cortes and Vapn, 995), medcal dagnoss (arasseno et al, 995). he SVM formulaton can be started as the followng. Gven a tranng set ( x, y ),,,, wth nput data x R and correspondng bnary class label y {, }, SVM algorthm sees the separatng hyperplane wth largest margn. he problem can be formulated as follow: mn, w, b w w subect to y ( w x b),,. (4) Standard method to solve the problem (3)-(4) s convex programmng, where Lagrange method s appled to transfer prmal to dual problems of optmzaton. Specfcally, we frst construct Lagrangan LP w w [ y ( w x b) ] (5) are nonnegatve Lagrange multplers correspondng to (4). he soluton s acheved by a saddle pont of the Lagrangan whch has to be mnmzed wth respect to w and b and maxmzed wth respect to. Dfferentatng (5) and set the results equal to zero, LP w w y x (6) 0 LP y 0 b (7) he optmal soluton s obtaned from (6) as w y x (8) where denotes optmal values. ow substtutng (8) and (7) nto (5), n (3) 34

8 Modern Appled Scence December, 009 LD w w y y he dual problem s posed as quadratc programmng: maxmze LD 0 subect to y x x (9) 0 C,,,. he condtons [ y ( w x b ) ] 0,,, (0) mples that 0 only when constrant (4) actve. he vectors for whch 0 are called support vectors. By (0), we obtan b y w x for any support vector x. By lnearty of the nner product and (8), the decson functon for the lnear separable case s w x b sgn y ( x x f ( x) sgn b. ) For lnearly non-separable case, we ntroduce a new set of varables volaton of the constrants. hus (3) and (4) are modfed as w, b, { },,, that measure the amount of mn w w C (), subect to y w x b,, () 0,,, (3) where C and are predetermned parameters defnes the cost of constrants. he Lagrangan s constructed as L P w w [ y ( w x b) ] C where and to ths problem s determned by mnmzng are Lagrange multplers whch are assocated wth constrants () and (3) respectvely. he soluton. Dfferentatng (4) and settng equal to zero, LP w LP b L P w y x 0 0 LP wth respect to (4) w, and b and maxmzng wth respect to and (5) y (6) C C 0 or 0 0,, by denotng C From (5), w / Substtutng (8),(7),(6) nto (4), we obtan / for. (7) y x. (8) / LD y y x x ( ). / ( C) hs leads to the dual problem as maxmze LD / y y x x ( ) / ( C) 35

9 Vol. 3, o. Modern Appled Scence subect to y 0 0,,,. When, max L D subect to y y y 0 0 C,,,. Hence the classfer s ) f ( x) sgn y ( x x b where x b y w x for any support vector [ y ( w x b ) ] 0,,,. x x such that 0 C followng from For nonlnear classfer, we need to map the nput varable x nto a hgher dmensonal feature space and wor wth lnear classfcaton n that space,.e., x ( x) ( x),, n n( x),where n n are some real functons. n n he soluton of the SVM has the form sgn ( x) w sgn y ( x) ( x b f ( x) b ) ( o avod complex calculaton of scalar product x) ( x ), we ntroduce ernel functon: K( x, y) ( x) ( x) n nn( x) n( y) whch satsfes Mercer s condtons. Hence, f ( x) sgn y K( x, x b are some real numbers and ). In ths wor, Gaussan ernel or RBF(radal bass functon) s used as t tends to gve good performance under general smoothng assumptons. he ernel s defned as K( x, x) exp( x x ). he ernel and regularzed parameters ( C, ) wth range [, ] are tuned by grdsearch technque to avod overftng problem. By applyng the above algorthm of SVM to tran the data under study, tranng results are obtaned. able 3 llustrates the cross-valdaton error correspondng to the tunng parameters (C, ). From the able 3, we obtan the best hyperparameters C = 4 4 and (the Gaussan ernel functon parameter) wth the smallest error of ten fold cross-valdaton = Consderng the n-sample data, the ht rate s.000 and error rate s but for the out-of-sample, the ht rate s and the error rate s Least square support vector machne for classfcaton LSSVM s a new verson of SVM modfed by Suyens et al (999). LSSVM uses least square loss functon to obtan a set of lnear equatons n dual space so that learnng rate s faster and the complexty of calculaton n convex programmng (n SVM) s relaxed. In addton, the LSSVM avods the drawbac faced by SVM such as trade-off parameters ( C,, ) selecton; nstead the LSSVM requres only two hyper-parameters (, ) to tran the model. Accordng to Suyens et al (00), the equalty constrants of LSSVM can act as recurrent neural networ and nonlnear optmal control. Due to these nce propertes, LSSVM has been successfully appled for classfcaton and regresson problems, ncludng tme seres forecastng. Further applcaton can be found n Van Gestel et al (004) for detaled dscusson on classfcaton performance of LSSVM and Van Gestel et al (00) and Ye et al (004) for predctve capablty of LSSVM n chaotc tme seres predcton. Formally, gven a tranng set y {, }, ( x, y ),,, wth nput data, LSSVM formulaton s represented as follow: x R n 5 and correspondng bnary class label 5 36

10 Modern Appled Scence December, 009 J e subect to the equalty constrants y[ w ( x ) b] e,,,. hs mn ( w, e) w w w, b, e formulaton conssts of equalty nstead of nequalty constrants and taes nto account a squared error wth regularzaton term smlar to rdge regresson. he soluton s obtaned after constructng the Lagrangan: ( L w, b, e ; ) J ( w, b, e) { y [ w ( x ) b] e }, where are Lagrange multplers that can be postve or negatve n the LSSVM formulaton. From the condtons for optmalty, one obtans the Karush-Kuhn-ucer (KK) L 0 w y( x ) w L 0 y 0 b L 0 e,, e L 0 y[ w ( x ) b] e By elmnatng w and e, 0 y b 0 y I v 0,,,. wth y y,, y ], [,,] e e,, e,,, ]. [ v [ ] [ Matrx y y ( x ) ( x ) y y K( x, x ) for,,, satsfes Mercer s condton and the LS-SVM model for estmatng classfer s obtaned as ) y( x) sgn y K( x, x b. In ths wor, Gaussan ernel or RBF(radal bass functon) s used as t tends to gve good performance under general smoothng assumptons. he ernel s defned as K( x, x ) exp( x ). he ernel and regularzed x parameters (, ) are tuned by grdsearch technque to avod overftng problem. he expermental results show that the obtaned hyper-parameters: gamma s chosen from the range [ , 48.43] and sgma square s selected from the range [ ,.85]. he cost of ten fold cross-valdaton s from the forecastng performance, the n-sample data ht rate s and error rate s For the out-of-sample data, the ht rate s and error rate s o sum t up, able 4 below gves a summary result of predcton performances by the ten dfferent approaches. From the able 4, we can see that almost all the algorthms generate hgh ht rate (more than 80%) and low error rate (less than 0%). By comparson, LSSVM rans frst one as t outperforms the other model though t s not better than SVM for n-sample predcton. he superor performances of the SVM models n ths study also support the results n the lterature. Bayesan classfcaton for Gaussan process produces good predcton as neural networ, followng the SVM and LS-SVM. K-nearest neghbor approach gves the worst predctve ablty and then ree classfcaton s the next one. 4. Concluson In ths paper, we apply ten dfferent technques of data mnng to forecast movement drecton of Hang Seng ndex from Hong Kong stoc maret. All algorthms produce good predcton wth ht rate more than 80%. he LS-SVM and SVM outperform the other models snce theoretcally they don t requre any pror assumpton on data property and ther algorthms guarantee to effcently obtan global optmal soluton whch s unque. he other models may be relable for other marets, especally when the data fall nto each of ther propertes. As can be seen n the fgure -3 n appendx secton, dfferent stoc prces behave dfferently. herefore, all the approaches are recommended for forecasters of stoc ndex movement and the better models, SVM and LS-SVM, are more preferred. 37

11 Vol. 3, o. Modern Appled Scence References Ben-Davd, S., & Lndenbaum, M. (997). Learnng dstrbutons by ther densty levels: A paradgm for learnng wthout a teacher, Journal of Computer and System Scences 55, 7-8. Bshop, C. (995). eural etwors for Pattern Recognton, Clarendon Press, Oxford. Burges, C.J.C., & Schoopf, B. (997). Improvng the Accuracy and Speed of Support Vector Machnes, Advances n eural Informaton Processng Systems, MI Press, Cambrdge, MA. pp Cao, L.J., & ay, F. (00). Applcaton of support vector machnes n fnancal tme seres forecastng, Internatonal Journal of Management Scence, pp Chen, An-Sng, Daou, Hazem & Leung, Mar. Applcaton of eural etwors to an Emergng Fnancal Maret: Forecastng and radng the awan Stoc Index. [Onlne] Avalable: or DOI: 0.39/ssrn (July 00) Cortes, C., & Vapn, V.. (995). Support Vector etwors, Machne Learnng 0, Deng, Mn. Pattern and Forecast of Movement n Stoc Prce. [Onlne] Avalable: =7048 or DOI: 0.39/ssrn.7048 (ovember 8, 006) Duda, R., Hart, P., & Stor, D. (000). Pattern Classfcaton (Second Edton), Wley, ew Yor. McLachlan, G. J. (99), Dscrmnant Analyss and Statstcal Pattern Recognton, Wley, ew Yor. Fredman, J., Haste,., & bshran, R. (008). he Elements of Statstcal Learnng:Data Mnng, Inference and Predcton, Second Edton Sprnger. Haste,. (996). eural etwor, Encyclopeda of Bostatstcs, John Wley. Hayn, S. (994). eural etwors: A Comprehensve Foundaton, Macmllan, ew Yor. Huang, W., aamor, Y. & Wang, S.Y. (005). Forecastng stoc maret movement drecton wth support vector machne, Journal of Computers & Operatonal Research, pp Joachms,. (00). Learnng to Classfy ext usng Support Vector Machnes, Kluwer Academc Publshers, London. Karatzoglou, A., Mayer, D., & Horn, A. (006). Support Vector Machnes n R, Journal of Statstcal Software. Karatzoglou, A., Smola, A., Horn, A., & Zeles, A.. (004). An S4 Pacage for Kernel Methods n R, Journal of Statstcal Software. Km, K.J. (003). Fnancal tme seres forecastng usng support vector machnes, euralcomputng, 55, Kumar, Mansh and henmozh, M. Forecastng Stoc Index Movement: A Comparson of Support Vector Machnes and Random Forest, Indan Insttute of Captal Marets 9th Captal Marets Conference Paper. [Onlne] Avalable: 06, 006) Leung, Mar., Daou, Hazem, & Chen, An-Sng. Forecastng Stoc Indces: A Comparson of Classfcaton and Level Estmaton Models. [Onlne] Avalable: (March 999) Lo, A.W., & MacKnlay, A.C. (988). Stoc maret prces do not follow random wals: Evdence from a smple specfcaton test, Revew of Fnancal Studes, Mtchell,. M. (007). Machne learnng, McGraw Hll, USA. Muheree, S., Osuna, E., & Gros, F. (997). onlnear predcton of chaotc tme seres usng support vector machne. In: Proceedngs of the IEEE worshop on eural etwors for Sgnal Processng, Amela Island, FL., pp O Connor, M., Remus, W., & Grggs, K. (997). Gong up-gong down: How good are people at forecastng trends and changes n trends? Journal of Forecastng 6, Osuna, E.E, Freund, R., & Gros, F. (997). Support Vector Machnes: ranng and Applcatons, MI press, USA. Pearson, R.A.(004). ree structures for predctng stoc prce behavour, AZIAM J., Austral. Mathematcal Soc. 45, pp C950-C963 Qunlan, J.R. (986). Inducton of Decson rees, Journal of Machne Learnng, :8-06, Kluwer Academc Publshers. Rasmussen, C. E. & Wllams, C. K. I. (006). Gaussan Processes for Machne Learnng, the MI Press, ISB 06853X. Rpley, B.D. (996). Pattern recognton and neural networs, Cambrdge Unversty Press. Stoneng, D. (999). Improvng the manufacturablty of electronc desgns, IEEE Spectrum 36,

12 Modern Appled Scence December, 009 Suyens, J.A.K, & Vandewalle, J., (999). Least squares support vector machne classfers. eural Processng Letters (9) Suyens, J.A.K., Vandewalle, J., & De Moor, B. (00). Optmal control by least squares support vector machnes. eural etwors (4) arasseno, L., Hayton, P., Cerneaz,., & Brady, M. (995). ovelty detecton for the dentfcaton of masses n mammograms. In: Proceedngs fourth IEE Internatonal Conference on Artfcal eural etwors, Cambrdge, pp Van Gestel,., Suyens, J.A.K, Baesens, B., Vaene, S., Vanthenen, J., Dedene, G., Moor B.D., & Vandewalle, G. (004). Benchmarng Least Squares Support Vector Machne Classfers, Machne Learnng, (54) 5-3. Van Gestel,.V., Suyens, J.A.K, Baestaens, D.E., Lambrechts, A., Lancret G., Vandaele B., De Moor B., & Vandewalle, J.(00). Fnancal me Seres Predcton usng Least Squares Support Vector Machnes wthn the Evdence Framewor, IEEE ransactons on eural etwors, () Vapn, V.. (995). he nature of statstcal learnng theory, Sprnger-Verlag, ew Yor. Wllams, C. K. I. & Barber, D. (999). Bayesan classfcaton wth Gaussan Processes, Pattern Analyss and Machne Intellgence, 0() 34-35, IEEE. Wu, Y., & Zhang, H. (997). Forward premums as unbased predctors of future currency deprecaton: A non-parametrc analyss, Journal of Internatonal Money and Fnance 6, Wu, X.D., Kumar V., Qunlan J.R., Ghosh J., Yang Q., Motoda, H., McLachlan G.J., g A., Lu B., Yu P.S., Zhou Z.H., Stenbach, M., Hand, D.J., & Stenberg, D. (008). op 0 algorthms n data mnng, Knowledge Informaton System, Sprnger-Verlag London. Ye, M.Y., & Wang, X.D. (004). Chaotc tme seres predcton usng least squares support vector machne, J. Chnese Physcs, IOP Publshng Ltd. IP address: Appendces: able. Selecton of by cross-valdaton technque = = = he table dsplays the mean square error and correspondng value of. he optmal s 0 wth smallest MSE

13 Vol. 3, o. Modern Appled Scence able. Process of prunng tree CP n.splt rel error x_error x_std Denote varables actually used n tree constructon by V = Open prce, V = Low prce, V3=Hgh prce, V4= S&p500, V5 = FX, V6 = Close prce. he class s when the next prce s larger than the prevous prce, whle the class 0 s refers to when the next prce s smaller than the prevous prce. We choose CP = correspondng to 7 splts based on the smallest value of the x_error : to obtan the followng pruned tree: Pruned ree V3< < V< V>= V>= V< V3< V3< V>= V4< V< V< V>=.005V5>= V4>= V3< V4< V>= V< V<

14 Modern Appled Scence December, 009 able 3. Hyperparamters selecton based on 0 fold cross-valdaton for SVM C / he table llustrates the cross-valdaton error correspondng to the tunng parameters (C, ). he tranng result shows 4 4 that (C, )=(, ) whch corresponds to the smallest tranng error = able 4. Summary result of predcton performances by ten dfferent approaches In-sample Out-of-sample Predctors Ht Rate Error Rate Ht Rate Error Rate Ran LDA QDA K aïve Bayes Logt model ree classfcaton eural networ Gaussan process SVM LS-SVM

15 Vol. 3, o. Modern Appled Scence. x 04 HSI prce Jan 000 Jan 00 Jan 004 Dec 006 Fgure. Daly closng prces of Hang Seng Index 600 S & P 500 ndex prce 6 S & P 500 log return Jan 000 Jan 00 Jan 004 Dec Jan 000 Jan 00 Jan 004 Dec 006 Fgure. Daly closng prces (left) and Log return (rght) of S & P 500 ndex 7.84 HKDvsUSD exchange rate 0.3 Log return of HKDvsUSD exchange rate Jan 000 Jan 00 Jan 004 Dec Jan 000 Jan 00 Jan 004 Dec 006 Fgure 3. Daly prces (left) and Log return (rght) of currency exchange rate between HKD and USD 4

Forecasting the Direction and Strength of Stock Market Movement

Forecasting the Direction and Strength of Stock Market Movement Forecastng the Drecton and Strength of Stock Market Movement Jngwe Chen Mng Chen Nan Ye cjngwe@stanford.edu mchen5@stanford.edu nanye@stanford.edu Abstract - Stock market s one of the most complcated systems

More information

Logistic Regression. Lecture 4: More classifiers and classes. Logistic regression. Adaboost. Optimization. Multiple class classification

Logistic Regression. Lecture 4: More classifiers and classes. Logistic regression. Adaboost. Optimization. Multiple class classification Lecture 4: More classfers and classes C4B Machne Learnng Hlary 20 A. Zsserman Logstc regresson Loss functons revsted Adaboost Loss functons revsted Optmzaton Multple class classfcaton Logstc Regresson

More information

The Development of Web Log Mining Based on Improve-K-Means Clustering Analysis

The Development of Web Log Mining Based on Improve-K-Means Clustering Analysis The Development of Web Log Mnng Based on Improve-K-Means Clusterng Analyss TngZhong Wang * College of Informaton Technology, Luoyang Normal Unversty, Luoyang, 471022, Chna wangtngzhong2@sna.cn Abstract.

More information

Forecasting the Demand of Emergency Supplies: Based on the CBR Theory and BP Neural Network

Forecasting the Demand of Emergency Supplies: Based on the CBR Theory and BP Neural Network 700 Proceedngs of the 8th Internatonal Conference on Innovaton & Management Forecastng the Demand of Emergency Supples: Based on the CBR Theory and BP Neural Network Fu Deqang, Lu Yun, L Changbng School

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Max Wellng Department of Computer Scence Unversty of Toronto 10 Kng s College Road Toronto, M5S 3G5 Canada wellng@cs.toronto.edu Abstract Ths s a note to explan support vector machnes.

More information

What is Candidate Sampling

What is Candidate Sampling What s Canddate Samplng Say we have a multclass or mult label problem where each tranng example ( x, T ) conssts of a context x a small (mult)set of target classes T out of a large unverse L of possble

More information

CS 2750 Machine Learning. Lecture 3. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 3. Density estimation. CS 2750 Machine Learning. Announcements Lecture 3 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Next lecture: Matlab tutoral Announcements Rules for attendng the class: Regstered for credt Regstered for audt (only f there

More information

Lecture 5,6 Linear Methods for Classification. Summary

Lecture 5,6 Linear Methods for Classification. Summary Lecture 5,6 Lnear Methods for Classfcaton Rce ELEC 697 Farnaz Koushanfar Fall 2006 Summary Bayes Classfers Lnear Classfers Lnear regresson of an ndcator matrx Lnear dscrmnant analyss (LDA) Logstc regresson

More information

An Alternative Way to Measure Private Equity Performance

An Alternative Way to Measure Private Equity Performance An Alternatve Way to Measure Prvate Equty Performance Peter Todd Parlux Investment Technology LLC Summary Internal Rate of Return (IRR) s probably the most common way to measure the performance of prvate

More information

NPAR TESTS. One-Sample Chi-Square Test. Cell Specification. Observed Frequencies 1O i 6. Expected Frequencies 1EXP i 6

NPAR TESTS. One-Sample Chi-Square Test. Cell Specification. Observed Frequencies 1O i 6. Expected Frequencies 1EXP i 6 PAR TESTS If a WEIGHT varable s specfed, t s used to replcate a case as many tmes as ndcated by the weght value rounded to the nearest nteger. If the workspace requrements are exceeded and samplng has

More information

Support Vector Machine Model for Currency Crisis Discrimination. Arindam Chaudhuri 1. Abstract

Support Vector Machine Model for Currency Crisis Discrimination. Arindam Chaudhuri 1. Abstract Support Vector Machne Model for Currency Crss Dscrmnaton Arndam Chaudhur Abstract Support Vector Machne (SVM) s powerful classfcaton technque based on the dea of structural rsk mnmzaton. Use of kernel

More information

Single and multiple stage classifiers implementing logistic discrimination

Single and multiple stage classifiers implementing logistic discrimination Sngle and multple stage classfers mplementng logstc dscrmnaton Hélo Radke Bttencourt 1 Dens Alter de Olvera Moraes 2 Vctor Haertel 2 1 Pontfíca Unversdade Católca do Ro Grande do Sul - PUCRS Av. Ipranga,

More information

Causal, Explanatory Forecasting. Analysis. Regression Analysis. Simple Linear Regression. Which is Independent? Forecasting

Causal, Explanatory Forecasting. Analysis. Regression Analysis. Simple Linear Regression. Which is Independent? Forecasting Causal, Explanatory Forecastng Assumes cause-and-effect relatonshp between system nputs and ts output Forecastng wth Regresson Analyss Rchard S. Barr Inputs System Cause + Effect Relatonshp The job of

More information

How Sets of Coherent Probabilities May Serve as Models for Degrees of Incoherence

How Sets of Coherent Probabilities May Serve as Models for Degrees of Incoherence 1 st Internatonal Symposum on Imprecse Probabltes and Ther Applcatons, Ghent, Belgum, 29 June 2 July 1999 How Sets of Coherent Probabltes May Serve as Models for Degrees of Incoherence Mar J. Schervsh

More information

Can Auto Liability Insurance Purchases Signal Risk Attitude?

Can Auto Liability Insurance Purchases Signal Risk Attitude? Internatonal Journal of Busness and Economcs, 2011, Vol. 10, No. 2, 159-164 Can Auto Lablty Insurance Purchases Sgnal Rsk Atttude? Chu-Shu L Department of Internatonal Busness, Asa Unversty, Tawan Sheng-Chang

More information

Statistical Methods to Develop Rating Models

Statistical Methods to Develop Rating Models Statstcal Methods to Develop Ratng Models [Evelyn Hayden and Danel Porath, Österrechsche Natonalbank and Unversty of Appled Scences at Manz] Source: The Basel II Rsk Parameters Estmaton, Valdaton, and

More information

How To Calculate The Accountng Perod Of Nequalty

How To Calculate The Accountng Perod Of Nequalty Inequalty and The Accountng Perod Quentn Wodon and Shlomo Ytzha World Ban and Hebrew Unversty September Abstract Income nequalty typcally declnes wth the length of tme taen nto account for measurement.

More information

benefit is 2, paid if the policyholder dies within the year, and probability of death within the year is ).

benefit is 2, paid if the policyholder dies within the year, and probability of death within the year is ). REVIEW OF RISK MANAGEMENT CONCEPTS LOSS DISTRIBUTIONS AND INSURANCE Loss and nsurance: When someone s subject to the rsk of ncurrng a fnancal loss, the loss s generally modeled usng a random varable or

More information

Discussion Papers. Support Vector Machines (SVM) as a Technique for Solvency Analysis. Laura Auria Rouslan A. Moro. Berlin, August 2008

Discussion Papers. Support Vector Machines (SVM) as a Technique for Solvency Analysis. Laura Auria Rouslan A. Moro. Berlin, August 2008 Deutsches Insttut für Wrtschaftsforschung www.dw.de Dscusson Papers 8 Laura Aura Rouslan A. Moro Support Vector Machnes (SVM) as a Technque for Solvency Analyss Berln, August 2008 Opnons expressed n ths

More information

Project Networks With Mixed-Time Constraints

Project Networks With Mixed-Time Constraints Project Networs Wth Mxed-Tme Constrants L Caccetta and B Wattananon Western Australan Centre of Excellence n Industral Optmsaton (WACEIO) Curtn Unversty of Technology GPO Box U1987 Perth Western Australa

More information

BERNSTEIN POLYNOMIALS

BERNSTEIN POLYNOMIALS On-Lne Geometrc Modelng Notes BERNSTEIN POLYNOMIALS Kenneth I. Joy Vsualzaton and Graphcs Research Group Department of Computer Scence Unversty of Calforna, Davs Overvew Polynomals are ncredbly useful

More information

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module LOSSLESS IMAGE COMPRESSION SYSTEMS Lesson 3 Lossless Compresson: Huffman Codng Instructonal Objectves At the end of ths lesson, the students should be able to:. Defne and measure source entropy..

More information

Institute of Informatics, Faculty of Business and Management, Brno University of Technology,Czech Republic

Institute of Informatics, Faculty of Business and Management, Brno University of Technology,Czech Republic Lagrange Multplers as Quanttatve Indcators n Economcs Ivan Mezník Insttute of Informatcs, Faculty of Busness and Management, Brno Unversty of TechnologCzech Republc Abstract The quanttatve role of Lagrange

More information

Forecasting the Prices of TAIEX Options by Using Genetic Programming and Support Vector Regression

Forecasting the Prices of TAIEX Options by Using Genetic Programming and Support Vector Regression Proceedngs of the Internatonal MultConference of Engneers and Computer Scentsts 2015 Vol I, Forecastng the Prces of AIEX Optons by Usng Genetc Programmng and Support Vector Regresson Chh-Mng Hsu, Yng-Ch

More information

A hybrid global optimization algorithm based on parallel chaos optimization and outlook algorithm

A hybrid global optimization algorithm based on parallel chaos optimization and outlook algorithm Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(7):1884-1889 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 A hybrd global optmzaton algorthm based on parallel

More information

8 Algorithm for Binary Searching in Trees

8 Algorithm for Binary Searching in Trees 8 Algorthm for Bnary Searchng n Trees In ths secton we present our algorthm for bnary searchng n trees. A crucal observaton employed by the algorthm s that ths problem can be effcently solved when the

More information

Recurrence. 1 Definitions and main statements

Recurrence. 1 Definitions and main statements Recurrence 1 Defntons and man statements Let X n, n = 0, 1, 2,... be a MC wth the state space S = (1, 2,...), transton probabltes p j = P {X n+1 = j X n = }, and the transton matrx P = (p j ),j S def.

More information

Lecture 2: Single Layer Perceptrons Kevin Swingler

Lecture 2: Single Layer Perceptrons Kevin Swingler Lecture 2: Sngle Layer Perceptrons Kevn Sngler kms@cs.str.ac.uk Recap: McCulloch-Ptts Neuron Ths vastly smplfed model of real neurons s also knon as a Threshold Logc Unt: W 2 A Y 3 n W n. A set of synapses

More information

Logistic Regression. Steve Kroon

Logistic Regression. Steve Kroon Logstc Regresson Steve Kroon Course notes sectons: 24.3-24.4 Dsclamer: these notes do not explctly ndcate whether values are vectors or scalars, but expects the reader to dscern ths from the context. Scenaro

More information

Performance Analysis and Coding Strategy of ECOC SVMs

Performance Analysis and Coding Strategy of ECOC SVMs Internatonal Journal of Grd and Dstrbuted Computng Vol.7, No. (04), pp.67-76 http://dx.do.org/0.457/jgdc.04.7..07 Performance Analyss and Codng Strategy of ECOC SVMs Zhgang Yan, and Yuanxuan Yang, School

More information

Descriptive Models. Cluster Analysis. Example. General Applications of Clustering. Examples of Clustering Applications

Descriptive Models. Cluster Analysis. Example. General Applications of Clustering. Examples of Clustering Applications CMSC828G Prncples of Data Mnng Lecture #9 Today s Readng: HMS, chapter 9 Today s Lecture: Descrptve Modelng Clusterng Algorthms Descrptve Models model presents the man features of the data, a global summary

More information

An Evaluation of the Extended Logistic, Simple Logistic, and Gompertz Models for Forecasting Short Lifecycle Products and Services

An Evaluation of the Extended Logistic, Simple Logistic, and Gompertz Models for Forecasting Short Lifecycle Products and Services An Evaluaton of the Extended Logstc, Smple Logstc, and Gompertz Models for Forecastng Short Lfecycle Products and Servces Charles V. Trappey a,1, Hsn-yng Wu b a Professor (Management Scence), Natonal Chao

More information

Feature selection for intrusion detection. Slobodan Petrović NISlab, Gjøvik University College

Feature selection for intrusion detection. Slobodan Petrović NISlab, Gjøvik University College Feature selecton for ntruson detecton Slobodan Petrovć NISlab, Gjøvk Unversty College Contents The feature selecton problem Intruson detecton Traffc features relevant for IDS The CFS measure The mrmr measure

More information

Financial market forecasting using a two-step kernel learning method for the support vector regression

Financial market forecasting using a two-step kernel learning method for the support vector regression Ann Oper Res (2010) 174: 103 120 DOI 10.1007/s10479-008-0357-7 Fnancal market forecastng usng a two-step kernel learnng method for the support vector regresson L Wang J Zhu Publshed onlne: 28 May 2008

More information

Improved SVM in Cloud Computing Information Mining

Improved SVM in Cloud Computing Information Mining Internatonal Journal of Grd Dstrbuton Computng Vol.8, No.1 (015), pp.33-40 http://dx.do.org/10.1457/jgdc.015.8.1.04 Improved n Cloud Computng Informaton Mnng Lvshuhong (ZhengDe polytechnc college JangSu

More information

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. Sequential Optimizing Investing Strategy with Neural Networks

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. Sequential Optimizing Investing Strategy with Neural Networks MATHEMATICAL ENGINEERING TECHNICAL REPORTS Sequental Optmzng Investng Strategy wth Neural Networks Ryo ADACHI and Akmch TAKEMURA METR 2010 03 February 2010 DEPARTMENT OF MATHEMATICAL INFORMATICS GRADUATE

More information

L10: Linear discriminants analysis

L10: Linear discriminants analysis L0: Lnear dscrmnants analyss Lnear dscrmnant analyss, two classes Lnear dscrmnant analyss, C classes LDA vs. PCA Lmtatons of LDA Varants of LDA Other dmensonalty reducton methods CSCE 666 Pattern Analyss

More information

ECE544NA Final Project: Robust Machine Learning Hardware via Classifier Ensemble

ECE544NA Final Project: Robust Machine Learning Hardware via Classifier Ensemble 1 ECE544NA Fnal Project: Robust Machne Learnng Hardware va Classfer Ensemble Sa Zhang, szhang12@llnos.edu Dept. of Electr. & Comput. Eng., Unv. of Illnos at Urbana-Champagn, Urbana, IL, USA Abstract In

More information

A COLLABORATIVE TRADING MODEL BY SUPPORT VECTOR REGRESSION AND TS FUZZY RULE FOR DAILY STOCK TURNING POINTS DETECTION

A COLLABORATIVE TRADING MODEL BY SUPPORT VECTOR REGRESSION AND TS FUZZY RULE FOR DAILY STOCK TURNING POINTS DETECTION A COLLABORATIVE TRADING MODEL BY SUPPORT VECTOR REGRESSION AND TS FUZZY RULE FOR DAILY STOCK TURNING POINTS DETECTION JHENG-LONG WU, PEI-CHANN CHANG, KAI-TING CHANG Department of Informaton Management,

More information

A study on the ability of Support Vector Regression and Neural Networks to Forecast Basic Time Series Patterns

A study on the ability of Support Vector Regression and Neural Networks to Forecast Basic Time Series Patterns A study on the ablty of Support Vector Regresson and Neural Networks to Forecast Basc Tme Seres Patterns Sven F. Crone, Jose Guajardo 2, and Rchard Weber 2 Lancaster Unversty, Department of Management

More information

Credit Limit Optimization (CLO) for Credit Cards

Credit Limit Optimization (CLO) for Credit Cards Credt Lmt Optmzaton (CLO) for Credt Cards Vay S. Desa CSCC IX, Ednburgh September 8, 2005 Copyrght 2003, SAS Insttute Inc. All rghts reserved. SAS Propretary Agenda Background Tradtonal approaches to credt

More information

Searching for Interacting Features for Spam Filtering

Searching for Interacting Features for Spam Filtering Searchng for Interactng Features for Spam Flterng Chuanlang Chen 1, Yun-Chao Gong 2, Rongfang Be 1,, and X. Z. Gao 3 1 Department of Computer Scence, Bejng Normal Unversty, Bejng 100875, Chna 2 Software

More information

Fault tolerance in cloud technologies presented as a service

Fault tolerance in cloud technologies presented as a service Internatonal Scentfc Conference Computer Scence 2015 Pavel Dzhunev, PhD student Fault tolerance n cloud technologes presented as a servce INTRODUCTION Improvements n technques for vrtualzaton and performance

More information

Data Visualization by Pairwise Distortion Minimization

Data Visualization by Pairwise Distortion Minimization Communcatons n Statstcs, Theory and Methods 34 (6), 005 Data Vsualzaton by Parwse Dstorton Mnmzaton By Marc Sobel, and Longn Jan Lateck* Department of Statstcs and Department of Computer and Informaton

More information

Gender Classification for Real-Time Audience Analysis System

Gender Classification for Real-Time Audience Analysis System Gender Classfcaton for Real-Tme Audence Analyss System Vladmr Khryashchev, Lev Shmaglt, Andrey Shemyakov, Anton Lebedev Yaroslavl State Unversty Yaroslavl, Russa vhr@yandex.ru, shmaglt_lev@yahoo.com, andrey.shemakov@gmal.com,

More information

New Approaches to Support Vector Ordinal Regression

New Approaches to Support Vector Ordinal Regression New Approaches to Support Vector Ordnal Regresson We Chu chuwe@gatsby.ucl.ac.uk Gatsby Computatonal Neuroscence Unt, Unversty College London, London, WCN 3AR, UK S. Sathya Keerth selvarak@yahoo-nc.com

More information

LSSVM-ABC Algorithm for Stock Price prediction Osman Hegazy 1, Omar S. Soliman 2 and Mustafa Abdul Salam 3

LSSVM-ABC Algorithm for Stock Price prediction Osman Hegazy 1, Omar S. Soliman 2 and Mustafa Abdul Salam 3 LSSVM-ABC Algorthm for Stock Prce predcton Osman Hegazy 1, Omar S. Solman 2 and Mustafa Abdul Salam 3 1, 2 (Faculty of Computers and Informatcs, Caro Unversty, Egypt) 3 (Hgher echnologcal Insttute (H..I),

More information

On the Optimal Control of a Cascade of Hydro-Electric Power Stations

On the Optimal Control of a Cascade of Hydro-Electric Power Stations On the Optmal Control of a Cascade of Hydro-Electrc Power Statons M.C.M. Guedes a, A.F. Rbero a, G.V. Smrnov b and S. Vlela c a Department of Mathematcs, School of Scences, Unversty of Porto, Portugal;

More information

Face Verification Problem. Face Recognition Problem. Application: Access Control. Biometric Authentication. Face Verification (1:1 matching)

Face Verification Problem. Face Recognition Problem. Application: Access Control. Biometric Authentication. Face Verification (1:1 matching) Face Recognton Problem Face Verfcaton Problem Face Verfcaton (1:1 matchng) Querymage face query Face Recognton (1:N matchng) database Applcaton: Access Control www.vsage.com www.vsoncs.com Bometrc Authentcaton

More information

The Greedy Method. Introduction. 0/1 Knapsack Problem

The Greedy Method. Introduction. 0/1 Knapsack Problem The Greedy Method Introducton We have completed data structures. We now are gong to look at algorthm desgn methods. Often we are lookng at optmzaton problems whose performance s exponental. For an optmzaton

More information

Georey E. Hinton. University oftoronto. Email: zoubin@cs.toronto.edu. Technical Report CRG-TR-96-1. May 21, 1996 (revised Feb 27, 1997) Abstract

Georey E. Hinton. University oftoronto. Email: zoubin@cs.toronto.edu. Technical Report CRG-TR-96-1. May 21, 1996 (revised Feb 27, 1997) Abstract The EM Algorthm for Mxtures of Factor Analyzers Zoubn Ghahraman Georey E. Hnton Department of Computer Scence Unversty oftoronto 6 Kng's College Road Toronto, Canada M5S A4 Emal: zoubn@cs.toronto.edu Techncal

More information

CHAPTER 14 MORE ABOUT REGRESSION

CHAPTER 14 MORE ABOUT REGRESSION CHAPTER 14 MORE ABOUT REGRESSION We learned n Chapter 5 that often a straght lne descrbes the pattern of a relatonshp between two quanttatve varables. For nstance, n Example 5.1 we explored the relatonshp

More information

SVM Tutorial: Classification, Regression, and Ranking

SVM Tutorial: Classification, Regression, and Ranking SVM Tutoral: Classfcaton, Regresson, and Rankng Hwanjo Yu and Sungchul Km 1 Introducton Support Vector Machnes(SVMs) have been extensvely researched n the data mnng and machne learnng communtes for the

More information

Binomial Link Functions. Lori Murray, Phil Munz

Binomial Link Functions. Lori Murray, Phil Munz Bnomal Lnk Functons Lor Murray, Phl Munz Bnomal Lnk Functons Logt Lnk functon: ( p) p ln 1 p Probt Lnk functon: ( p) 1 ( p) Complentary Log Log functon: ( p) ln( ln(1 p)) Motvatng Example A researcher

More information

On-Line Fault Detection in Wind Turbine Transmission System using Adaptive Filter and Robust Statistical Features

On-Line Fault Detection in Wind Turbine Transmission System using Adaptive Filter and Robust Statistical Features On-Lne Fault Detecton n Wnd Turbne Transmsson System usng Adaptve Flter and Robust Statstcal Features Ruoyu L Remote Dagnostcs Center SKF USA Inc. 3443 N. Sam Houston Pkwy., Houston TX 77086 Emal: ruoyu.l@skf.com

More information

J. Parallel Distrib. Comput.

J. Parallel Distrib. Comput. J. Parallel Dstrb. Comput. 71 (2011) 62 76 Contents lsts avalable at ScenceDrect J. Parallel Dstrb. Comput. journal homepage: www.elsever.com/locate/jpdc Optmzng server placement n dstrbuted systems n

More information

Scale Dependence of Overconfidence in Stock Market Volatility Forecasts

Scale Dependence of Overconfidence in Stock Market Volatility Forecasts Scale Dependence of Overconfdence n Stoc Maret Volatlty Forecasts Marus Glaser, Thomas Langer, Jens Reynders, Martn Weber* June 7, 007 Abstract In ths study, we analyze whether volatlty forecasts (judgmental

More information

Power-of-Two Policies for Single- Warehouse Multi-Retailer Inventory Systems with Order Frequency Discounts

Power-of-Two Policies for Single- Warehouse Multi-Retailer Inventory Systems with Order Frequency Discounts Power-of-wo Polces for Sngle- Warehouse Mult-Retaler Inventory Systems wth Order Frequency Dscounts José A. Ventura Pennsylvana State Unversty (USA) Yale. Herer echnon Israel Insttute of echnology (Israel)

More information

An Enhanced Super-Resolution System with Improved Image Registration, Automatic Image Selection, and Image Enhancement

An Enhanced Super-Resolution System with Improved Image Registration, Automatic Image Selection, and Image Enhancement An Enhanced Super-Resoluton System wth Improved Image Regstraton, Automatc Image Selecton, and Image Enhancement Yu-Chuan Kuo ( ), Chen-Yu Chen ( ), and Chou-Shann Fuh ( ) Department of Computer Scence

More information

PSYCHOLOGICAL RESEARCH (PYC 304-C) Lecture 12

PSYCHOLOGICAL RESEARCH (PYC 304-C) Lecture 12 14 The Ch-squared dstrbuton PSYCHOLOGICAL RESEARCH (PYC 304-C) Lecture 1 If a normal varable X, havng mean µ and varance σ, s standardsed, the new varable Z has a mean 0 and varance 1. When ths standardsed

More information

THE DISTRIBUTION OF LOAN PORTFOLIO VALUE * Oldrich Alfons Vasicek

THE DISTRIBUTION OF LOAN PORTFOLIO VALUE * Oldrich Alfons Vasicek HE DISRIBUION OF LOAN PORFOLIO VALUE * Oldrch Alfons Vascek he amount of captal necessary to support a portfolo of debt securtes depends on the probablty dstrbuton of the portfolo loss. Consder a portfolo

More information

Vision Mouse. Saurabh Sarkar a* University of Cincinnati, Cincinnati, USA ABSTRACT 1. INTRODUCTION

Vision Mouse. Saurabh Sarkar a* University of Cincinnati, Cincinnati, USA ABSTRACT 1. INTRODUCTION Vson Mouse Saurabh Sarkar a* a Unversty of Cncnnat, Cncnnat, USA ABSTRACT The report dscusses a vson based approach towards trackng of eyes and fngers. The report descrbes the process of locatng the possble

More information

THE APPLICATION OF DATA MINING TECHNIQUES AND MULTIPLE CLASSIFIERS TO MARKETING DECISION

THE APPLICATION OF DATA MINING TECHNIQUES AND MULTIPLE CLASSIFIERS TO MARKETING DECISION Internatonal Journal of Electronc Busness Management, Vol. 3, No. 4, pp. 30-30 (2005) 30 THE APPLICATION OF DATA MINING TECHNIQUES AND MULTIPLE CLASSIFIERS TO MARKETING DECISION Yu-Mn Chang *, Yu-Cheh

More information

1. Measuring association using correlation and regression

1. Measuring association using correlation and regression How to measure assocaton I: Correlaton. 1. Measurng assocaton usng correlaton and regresson We often would lke to know how one varable, such as a mother's weght, s related to another varable, such as a

More information

A Genetic Programming Based Stock Price Predictor together with Mean-Variance Based Sell/Buy Actions

A Genetic Programming Based Stock Price Predictor together with Mean-Variance Based Sell/Buy Actions Proceedngs of the World Congress on Engneerng 28 Vol II WCE 28, July 2-4, 28, London, U.K. A Genetc Programmng Based Stock Prce Predctor together wth Mean-Varance Based Sell/Buy Actons Ramn Rajaboun and

More information

Bayesian Network Based Causal Relationship Identification and Funding Success Prediction in P2P Lending

Bayesian Network Based Causal Relationship Identification and Funding Success Prediction in P2P Lending Proceedngs of 2012 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 25 (2012) (2012) IACSIT Press, Sngapore Bayesan Network Based Causal Relatonshp Identfcaton and Fundng Success

More information

Realistic Image Synthesis

Realistic Image Synthesis Realstc Image Synthess - Combned Samplng and Path Tracng - Phlpp Slusallek Karol Myszkowsk Vncent Pegoraro Overvew: Today Combned Samplng (Multple Importance Samplng) Renderng and Measurng Equaton Random

More information

1 Example 1: Axis-aligned rectangles

1 Example 1: Axis-aligned rectangles COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture # 6 Scrbe: Aaron Schld February 21, 2013 Last class, we dscussed an analogue for Occam s Razor for nfnte hypothess spaces that, n conjuncton

More information

Exhaustive Regression. An Exploration of Regression-Based Data Mining Techniques Using Super Computation

Exhaustive Regression. An Exploration of Regression-Based Data Mining Techniques Using Super Computation Exhaustve Regresson An Exploraton of Regresson-Based Data Mnng Technques Usng Super Computaton Antony Daves, Ph.D. Assocate Professor of Economcs Duquesne Unversty Pttsburgh, PA 58 Research Fellow The

More information

Calculation of Sampling Weights

Calculation of Sampling Weights Perre Foy Statstcs Canada 4 Calculaton of Samplng Weghts 4.1 OVERVIEW The basc sample desgn used n TIMSS Populatons 1 and 2 was a two-stage stratfed cluster desgn. 1 The frst stage conssted of a sample

More information

Chapter 6. Classification and Prediction

Chapter 6. Classification and Prediction Chapter 6. Classfcaton and Predcton What s classfcaton? What s Lazy learners (or learnng from predcton? your neghbors) Issues regardng classfcaton and Frequent-pattern-based predcton classfcaton Classfcaton

More information

A DATA MINING APPLICATION IN A STUDENT DATABASE

A DATA MINING APPLICATION IN A STUDENT DATABASE JOURNAL OF AERONAUTICS AND SPACE TECHNOLOGIES JULY 005 VOLUME NUMBER (53-57) A DATA MINING APPLICATION IN A STUDENT DATABASE Şenol Zafer ERDOĞAN Maltepe Ünversty Faculty of Engneerng Büyükbakkalköy-Istanbul

More information

Luby s Alg. for Maximal Independent Sets using Pairwise Independence

Luby s Alg. for Maximal Independent Sets using Pairwise Independence Lecture Notes for Randomzed Algorthms Luby s Alg. for Maxmal Independent Sets usng Parwse Independence Last Updated by Erc Vgoda on February, 006 8. Maxmal Independent Sets For a graph G = (V, E), an ndependent

More information

NEURO-FUZZY INFERENCE SYSTEM FOR E-COMMERCE WEBSITE EVALUATION

NEURO-FUZZY INFERENCE SYSTEM FOR E-COMMERCE WEBSITE EVALUATION NEURO-FUZZY INFERENE SYSTEM FOR E-OMMERE WEBSITE EVALUATION Huan Lu, School of Software, Harbn Unversty of Scence and Technology, Harbn, hna Faculty of Appled Mathematcs and omputer Scence, Belarusan State

More information

ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING

ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING Matthew J. Lberatore, Department of Management and Operatons, Vllanova Unversty, Vllanova, PA 19085, 610-519-4390,

More information

A Novel Methodology of Working Capital Management for Large. Public Constructions by Using Fuzzy S-curve Regression

A Novel Methodology of Working Capital Management for Large. Public Constructions by Using Fuzzy S-curve Regression Novel Methodology of Workng Captal Management for Large Publc Constructons by Usng Fuzzy S-curve Regresson Cheng-Wu Chen, Morrs H. L. Wang and Tng-Ya Hseh Department of Cvl Engneerng, Natonal Central Unversty,

More information

NON-CONSTANT SUM RED-AND-BLACK GAMES WITH BET-DEPENDENT WIN PROBABILITY FUNCTION LAURA PONTIGGIA, University of the Sciences in Philadelphia

NON-CONSTANT SUM RED-AND-BLACK GAMES WITH BET-DEPENDENT WIN PROBABILITY FUNCTION LAURA PONTIGGIA, University of the Sciences in Philadelphia To appear n Journal o Appled Probablty June 2007 O-COSTAT SUM RED-AD-BLACK GAMES WITH BET-DEPEDET WI PROBABILITY FUCTIO LAURA POTIGGIA, Unversty o the Scences n Phladelpha Abstract In ths paper we nvestgate

More information

How To Understand The Results Of The German Meris Cloud And Water Vapour Product

How To Understand The Results Of The German Meris Cloud And Water Vapour Product Ttel: Project: Doc. No.: MERIS level 3 cloud and water vapour products MAPP MAPP-ATBD-ClWVL3 Issue: 1 Revson: 0 Date: 9.12.1998 Functon Name Organsaton Sgnature Date Author: Bennartz FUB Preusker FUB Schüller

More information

APPLICATION OF PROBE DATA COLLECTED VIA INFRARED BEACONS TO TRAFFIC MANEGEMENT

APPLICATION OF PROBE DATA COLLECTED VIA INFRARED BEACONS TO TRAFFIC MANEGEMENT APPLICATION OF PROBE DATA COLLECTED VIA INFRARED BEACONS TO TRAFFIC MANEGEMENT Toshhko Oda (1), Kochro Iwaoka (2) (1), (2) Infrastructure Systems Busness Unt, Panasonc System Networks Co., Ltd. Saedo-cho

More information

An Interest-Oriented Network Evolution Mechanism for Online Communities

An Interest-Oriented Network Evolution Mechanism for Online Communities An Interest-Orented Network Evoluton Mechansm for Onlne Communtes Cahong Sun and Xaopng Yang School of Informaton, Renmn Unversty of Chna, Bejng 100872, P.R. Chna {chsun,yang}@ruc.edu.cn Abstract. Onlne

More information

Modelling of Web Domain Visits by Radial Basis Function Neural Networks and Support Vector Machine Regression

Modelling of Web Domain Visits by Radial Basis Function Neural Networks and Support Vector Machine Regression Modellng of Web Doman Vsts by Radal Bass Functon Neural Networks and Support Vector Machne Regresson Vladmír Olej, Jana Flpová Insttute of System Engneerng and Informatcs Faculty of Economcs and Admnstraton,

More information

Course outline. Financial Time Series Analysis. Overview. Data analysis. Predictive signal. Trading strategy

Course outline. Financial Time Series Analysis. Overview. Data analysis. Predictive signal. Trading strategy Fnancal Tme Seres Analyss Patrck McSharry patrck@mcsharry.net www.mcsharry.net Trnty Term 2014 Mathematcal Insttute Unversty of Oxford Course outlne 1. Data analyss, probablty, correlatons, vsualsaton

More information

Survival analysis methods in Insurance Applications in car insurance contracts

Survival analysis methods in Insurance Applications in car insurance contracts Survval analyss methods n Insurance Applcatons n car nsurance contracts Abder OULIDI 1 Jean-Mare MARION 2 Hervé GANACHAUD 3 Abstract In ths wor, we are nterested n survval models and ther applcatons on

More information

Regression Models for a Binary Response Using EXCEL and JMP

Regression Models for a Binary Response Using EXCEL and JMP SEMATECH 997 Statstcal Methods Symposum Austn Regresson Models for a Bnary Response Usng EXCEL and JMP Davd C. Trndade, Ph.D. STAT-TECH Consultng and Tranng n Appled Statstcs San Jose, CA Topcs Practcal

More information

Support vector domain description

Support vector domain description Pattern Recognton Letters 20 (1999) 1191±1199 www.elsever.nl/locate/patrec Support vector doman descrpton Davd M.J. Tax *,1, Robert P.W. Dun Pattern Recognton Group, Faculty of Appled Scence, Delft Unversty

More information

Intelligent stock trading system by turning point confirming and probabilistic reasoning

Intelligent stock trading system by turning point confirming and probabilistic reasoning Expert Systems wth Applcatons Expert Systems wth Applcatons 34 (2008) 620 627 www.elsever.com/locate/eswa Intellgent stock tradng system by turnng pont confrmng and probablstc reasonng Depe Bao *, Zehong

More information

Fast Fuzzy Clustering of Web Page Collections

Fast Fuzzy Clustering of Web Page Collections Fast Fuzzy Clusterng of Web Page Collectons Chrstan Borgelt and Andreas Nürnberger Dept. of Knowledge Processng and Language Engneerng Otto-von-Guercke-Unversty of Magdeburg Unverstätsplatz, D-396 Magdeburg,

More information

Design of Output Codes for Fast Covering Learning using Basic Decomposition Techniques

Design of Output Codes for Fast Covering Learning using Basic Decomposition Techniques Journal of Computer Scence (7): 565-57, 6 ISSN 59-66 6 Scence Publcatons Desgn of Output Codes for Fast Coverng Learnng usng Basc Decomposton Technques Aruna Twar and Narendra S. Chaudhar, Faculty of Computer

More information

Research Article Integrated Model of Multiple Kernel Learning and Differential Evolution for EUR/USD Trading

Research Article Integrated Model of Multiple Kernel Learning and Differential Evolution for EUR/USD Trading Hndaw Publshng Corporaton e Scentfc World Journal, Artcle ID 914641, 12 pages http://dx.do.org/10.1155/2014/914641 Research Artcle Integrated Model of Multple Kernel Learnng and Dfferental Evoluton for

More information

The Application of Fractional Brownian Motion in Option Pricing

The Application of Fractional Brownian Motion in Option Pricing Vol. 0, No. (05), pp. 73-8 http://dx.do.org/0.457/jmue.05.0..6 The Applcaton of Fractonal Brownan Moton n Opton Prcng Qng-xn Zhou School of Basc Scence,arbn Unversty of Commerce,arbn zhouqngxn98@6.com

More information

Calendar Corrected Chaotic Forecast of Financial Time Series

Calendar Corrected Chaotic Forecast of Financial Time Series INTERNATIONAL JOURNAL OF BUSINESS, 11(4), 2006 ISSN: 1083 4346 Calendar Corrected Chaotc Forecast of Fnancal Tme Seres Alexandros Leonttss a and Costas Sropoulos b a Center for Research and Applcatons

More information

Figure 1. Training and Test data sets for Nasdaq-100 Index (b) NIFTY index

Figure 1. Training and Test data sets for Nasdaq-100 Index (b) NIFTY index Modelng Chaotc Behavor of Stock Indces Usng Intellgent Paradgms Ajth Abraham, Nnan Sajth Phlp and P. Saratchandran Department of Computer Scence, Oklahoma State Unversty, ulsa, Oklahoma 746, USA, Emal:

More information

A novel Method for Data Mining and Classification based on

A novel Method for Data Mining and Classification based on A novel Method for Data Mnng and Classfcaton based on Ensemble Learnng 1 1, Frst Author Nejang Normal Unversty;Schuan Nejang 641112,Chna, E-mal: lhan-gege@126.com Abstract Data mnng has been attached great

More information

Application of Quasi Monte Carlo methods and Global Sensitivity Analysis in finance

Application of Quasi Monte Carlo methods and Global Sensitivity Analysis in finance Applcaton of Quas Monte Carlo methods and Global Senstvty Analyss n fnance Serge Kucherenko, Nlay Shah Imperal College London, UK skucherenko@mperalacuk Daro Czraky Barclays Captal DaroCzraky@barclayscaptalcom

More information

Probability and Optimization Models for Racing

Probability and Optimization Models for Racing 1 Probablty and Optmzaton Models for Racng Vctor S. Y. Lo Unversty of Brtsh Columba Fdelty Investments Dsclamer: Ths presentaton does not reflect the opnons of Fdelty Investments. The work here was completed

More information

Learning from Large Distributed Data: A Scaling Down Sampling Scheme for Efficient Data Processing

Learning from Large Distributed Data: A Scaling Down Sampling Scheme for Efficient Data Processing Internatonal Journal of Machne Learnng and Computng, Vol. 4, No. 3, June 04 Learnng from Large Dstrbuted Data: A Scalng Down Samplng Scheme for Effcent Data Processng Che Ngufor and Janusz Wojtusak part

More information

Prediction of Disability Frequencies in Life Insurance

Prediction of Disability Frequencies in Life Insurance Predcton of Dsablty Frequences n Lfe Insurance Bernhard Köng Fran Weber Maro V. Wüthrch October 28, 2011 Abstract For the predcton of dsablty frequences, not only the observed, but also the ncurred but

More information

Latent Class Regression. Statistics for Psychosocial Research II: Structural Models December 4 and 6, 2006

Latent Class Regression. Statistics for Psychosocial Research II: Structural Models December 4 and 6, 2006 Latent Class Regresson Statstcs for Psychosocal Research II: Structural Models December 4 and 6, 2006 Latent Class Regresson (LCR) What s t and when do we use t? Recall the standard latent class model

More information

Comparison of support-vector machines and back propagation neural networks in forecasting the six major Asian stock markets

Comparison of support-vector machines and back propagation neural networks in forecasting the six major Asian stock markets Int. J. Electronc Fnance, Vol., No., 006 49 Comparson of support-vector machnes and back propagaton neural networks n forecastng the sx major Asan stock markets Wun-Hua Chen and Jen-Yng Shh Graduate Insttute

More information

Efficient Project Portfolio as a tool for Enterprise Risk Management

Efficient Project Portfolio as a tool for Enterprise Risk Management Effcent Proect Portfolo as a tool for Enterprse Rsk Management Valentn O. Nkonov Ural State Techncal Unversty Growth Traectory Consultng Company January 5, 27 Effcent Proect Portfolo as a tool for Enterprse

More information