On Autoencoders and Score Matching for Energy Based Models

Size: px
Start display at page:

Download "On Autoencoders and Score Matching for Energy Based Models"

Transcription

1 On Autoencoders and Score Matchng for Energy Based Models Kevn Swersky* Marc Aurelo Ranzato Davd Buchman* Benjamn M. Marln* Nando de Fretas* *Department of Computer Scence, Uersty of Brtsh Columba, Vancouver, BC V6T Z4, Canada Department of Computer Scence, Uersty of Toronto, Toronto, ON M5S G4, Canada Abstract We consder estmaton methods for the class of contnuous-data energy based models EBMs). Our man result shows that estmatng the parameters of an EBM usng score matchng when the condtonal dstrbuton over the vsble unts s Gaussan corresponds to tranng a partcular form of regularzed autoencoder. We show how dfferent Gaussan EBMs lead to dfferent autoencoder archtectures, provdng deep lnks between these two famles of models. We compare the score matchng estmator for the mpot model, a partcular Gaussan EBM, to several other tranng methods on a varety of tasks ncludng mage denosng and unsupervsed feature extracton. We show that the regularzaton functon nduced by score matchng leads to superor classfcaton performance relatve to a standard autoencoder. We also show that score matchng yelds classfcaton results that are ndstngushable from better-known stochastc approxmaton maxmum lkelhood estmators.. Introducton In ths work, we consder a rch class of probablstc models called energy based models EBMs) LeCun et al., 006; Teh et al., 003; Hnton, 00). These models defne a probablty dstrbuton though an exponentated energy functon. Markov Random Felds MRFs) and Restrcted Boltzmann Machnes RBMs) are the most common nstance of such models and have Appearng n Proceedngs of the 8 th Internatonal Conference on Machne Learnng, Bellevue, WA, USA, 0. Copyrght 0 by the authors)/owners). a long hstory n partcular applcaton areas ncludng modelng natural mages. Recently, more sophstcated latent varable EBMs for contnuous data ncludng the PoT Wellng et al., 003), mpot Ranzato et al., 00b), mcrbm Ranzato & Hnton, 00), FoE Schmdt et al., 00) and others have become popular models for learnng representatons of natural mages as well as other sources of real-valued data. Such models, also called gated MRFs, leverage latent varables to represent hgher order nteractons between the nput varables. In the very actve research area of deep learnng Hnton et al., 006), these models been employed as elementary buldng blocks to construct herarchcal models that acheve very promsng performance on several perceptual tasks Ranzato & Hnton, 00; Bengo, 009). Maxmum lkelhood estmaton s the default parameter estmaton approach for probablstc models due to ts optmal theoretcal propertes. Unfortunately, maxmum lkelhood estmaton s computatonally nfeasble n many EBM models due to the presence of an ntractable normalzaton term the partton functon) n the model probablty. Ths term arses n EBMs because the exponentated energes do not automatcally ntegrate to unty, unlke drected models parameterzed by products of locally normalzed condtonal dstrbutons Bayesan networks). Several alternatve methods have been proposed to estmate the parameters of an EBM wthout the need for computng the partton functon. One partcularly nterestng method s called score matchng SM) Hyvärnen, 005). The score matchng objectve functon s constructed from an L loss on the dfference between the dervatves of the log of the model and emprcal dstrbuton functons wth respect to the nputs. Hyvärnen 005) showed that ths results n a cancellaton of the

2 Autoencoders and Score Matchng partton functon. Further manpulaton yelds an estmator that can be computed analytcally and s provably consstent. Autoencoder neural networks are another class of models that are often used to model hgh-dmensonal realvalued data Hnton & Zemel, 994; Vncent et al., 008; Vncent, 0; Kngma & LeCun, 00). Both EBMs and autoencoders are unsupervsed models that can be thought of as learnng to re-represent nput data n a latent space. In contrast to probablstc EBMs, autoencoders are determnstc and feed-forward. As a result, autoencoders can be traned to reconstruct ther nput through one or more hdden layers, they have fast feed-forward nference for hdden layer states, and all common tranng losses lead to computatonally tractable model estmaton methods. In order to learn better representatons, autoencoders are often modfed by tyng the weghts between the nput and output layers to reduce the number of parameters, ncludng addtonal terms n the objectve to bas learnng toward sparse hdden unt actvatons, and addng nose to nput data to ncrease robustness Vncent et al., 008; Vncent, 0). Interestngly, Vncent 0) showed that a partcular knd of denosng autoencoder traned to mnmze an L reconstructon error can be nterpreted as Gaussan RBM traned usng Hyvärnen s score matchng estmator. In ths paper, we apply score matchng to a number of latent varable EBMs where the condtonal dstrbuton of the vsble unts gven the hdden unts s Gaussan. We show that the resultng estmaton algorthms can be nterpreted as mnmzng a regularzed L reconstructon error on the vsble unts. For Gaussan-bnary RBMs, the reconstructon term corresponds to a standard autoencoder wth ted weghts. For the mpot and mcrbm models, the reconstructon terms correspond to new autoencoder archtectures that take nto account the covarance structure of the nputs. Ths suggests a new way to derve novel autoencoder tranng crtera by applyng score matchng to the free energy of an EBM. We further generalze score matchng to arbtrary EBMs wth real-valued nput unts and show that ths vew leads to an ntutve nterpretaton for the regularzaton terms that appear n the score matchng objectve functon.. Score Matchng for Latent Energy Based Models A latent varable energy based model defnes a probablty dstrbuton over real valued data vectors v V R as follows: P v, h; θ) = exp E θv, h)), ) Zθ) where h H R n h are the latent varables, E θ v, h) s an energy functon parameterzed by θ Θ, and Zθ) s the partton functon. We refer to these models as latent energy based models. Ths general latent energy based model subsumes many specfc models for real-valued data such as Boltzmann machnes, exponental-famly harmonums Wellng et al., 005), factored RBMs and Product of Student s T PoT) models Memsevc & Hnton, 009; Ranzato & Hnton, 00; Ranzato et al., 00a;b). The margnal dstrbuton n terms of the free energy F θ v) s obtaned by ntegratng out the hdden varables as seen below. Typcally, but not always, ths margnalzaton can be carred out analytcally. P v; θ) = exp F θv)). ) Zθ) Maxmum lkelhood parameter estmaton s dffcult when Zθ) s ntractable. In EBMs the ntractablty of Zθ) arses due to the fact that t s a very hgh-dmensonal ntegral that often lacks a closed form soluton. In such cases, stochastc algorthms can be appled to approxmately maxmze the lkelhood and a varety of algorthms have been descrbed and evaluated Swersky et al., 00; Marln et al., 00) n the lterature ncludng contrastve dvergence CD) Hnton, 00), persstent contrastve dvergence PCD) Younes, 989; Teleman, 008), and fast persstent contrastve dvergence FPCD) Teleman & Hnton, 009). However, these methods often requre very careful hand-tunng of optmzaton-related parameters lke step sze, momentum, batch sze and weght decay, whch s complcated by the fact that the objectve functon can not be computed. The score matchng estmator was proposed by Hyvärnen 005) to overcome the ntractablty of Zθ) when dealng wth contnuous data. The score matchng objectve functon s defned through a score functon appled to the emprcal pv) and model p θ v) dstrbutons. The score functon for a generc dstrbuton pv) s gven by ψ pv)) = log pv) v = F θv) v = h E θv,h) v p θ h v)dh. The full objectve functon s gven below. Jθ) = E pv) ψ pv)) ψ p θ v))). 3)

3 Autoencoders and Score Matchng The beneft of optmzng Jθ) s that Zθ) cancels off n the dervatve of log p θ v) snce t s constant wth respect to each v. However, n the above form, Jθ) s stll ntractable due to the dependence on pv). Hyvärnen, shows that under weak regularty condtons Jθ) can be expressed n the followng form, whch can be tractably approxmated by replacng the expectaton over the emprcal dstrbuton by an emprcal average over the tranng set: Jθ) = E pv) ψ p θ v))) + ψ p θ v)) v. 4) In theoretcal stuatons where the regularty condtons on the dervatves of the emprcal dstrbuton are not satsfed, or n practcal stuatons where a fnte sample approxmaton to the expectaton over the emprcal dstrbuton s used, a smoothed verson of the score matchng estmator may be of nterest. Consder smoothng pv) usng a probablstc kernel q β v v ) wth bandwdth parameter β > 0. We obtan a new dstrbuton q β v) = q β v v ) pv )dv. Vncent 0) showed that applyng score matchng to q β v) s equvalent to the followng objectve functon where q β v, v ) = q β v v ) pv ): Qθ) =E qβ v,v ) ψ q β v v )) ψ p θ v))). 5) For the case where q β v v ) = N v v, β ).e. a Gaussan smoothng kernel wth varance β, ths s equvalent to the regularzed score matchng objectve proposed n Kngma & LeCun, 00). We refer to the objectve gven by Equaton 5 as denosng score matchng SMD). Although SMD s ntractable to evaluate analytcally, we can agan replace the ntegral over v by an emprcal average over a fnte sample of tranng data. We can then replace the ntegral over v by an emprcal average over samples v, whch can be easly drawn from q β v v ) for each tranng sample v. Compared to PCD and CD, SM and SMD gve tractable objectve functons that can be used to montor tranng progress. Whle SMD s not consstent, t does have sgnfcant computatonal advantages relatve to SM Vncent, 0). 3. Applyng and Interpretng Score Matchng For Latent EBMs We now derve score matchng objectves for several commonly used EBMs. In order to apply score matchng to a partcular EBM, one smply needs an expresson for the correspondng free energy. Example Score Matchng for Gaussanbnary RBMs: Here, the energy E θ v, h) s gven by: n v n h v σ W j h j n h b j h j + n v c v ), 6) where the parameters are θ = W, σ, b, c) and h j {0, }. Ths leads to the free energy F θ v): n v c v ) n h log + exp σ σ v σ W j + b j The correspondng score matchng objectve s: Jθ) = N σ N n v v n σ n= n h + Wj σ where ĥjn := sgm +exp x). c σ n h W j σ ĥ jn )) 7) ĥ jn ĥjn), 8) ) v n σ W j + b j and sgmx) := For a standardzed Normal model, wth c = 0 and σ =, ths objectve reduces to: Jθ) = N N n v n h v n W j ĥ jn n= n h + W jĥjn ĥjn), 9) The frst term corresponds to the quadratc reconstructon error of an autoencoder wth ted weghts. From ths we can see that ths type of of autoencoder, whch researchers have prevously treated as a dfferent model, can n fact be explaned by the applcaton of the score matchng estmaton prncple to Gaussan RBMs. Example Score matchng for mcrbm: The energy E θ v, h m, h c ) of the mcrbm model for each data pont ncludes mean Bernoull hdden unts h m j {0, } and covarance Bernoull hdden unts h c k {0, }. The latter allow one to model correlatons n the data v Ranzato & Hnton, 00; Ranzato et al., 00a). To ease the notaton, we wll gnore the ndex,

4 Autoencoders and Score Matchng n over the data. The energy for ths model s: n f n hc n v n hm n v P fk h c k C f v ) W j h m j v n hm f= n hc n v b m j h m j b c kh c k b v v + n v v, 0) where θ = b v, b m, b c, P, W, C). Ths leads to the free energy F θ v): n hc log + e φc k ) n hm n v log + e φm j ) b v v + n v v, ) where φ c k = nf f= P fk n v C f v ) + b c k and φm j = W jv + b m j. The correspondng score matchng objectve s: Jθ) = ψ p θ v)) n hc + n hm + n f ρĥc k )D k + ĥc k K k hm ˆ j h ˆ ) m j )W j ) ) n hc n ψ p θ v)) = ĥ c k D hm k + hˆ m j W j + b v v K k = P fk Cf f= n f ) ) D k = P fk C f v C f f= ĥ c k =sgm φc k) hˆ m j =sgm φ m ) j ρx) :=x x). = Example 3 Score matchng for mpot The energy E θ v, h m, h c ) of the mpot model s: n hc h c k + n v C k v ) ) + γ) logh c k) n v + v n v n hm b v v n hm h m j W j v b m j h m j, 3) where θ = γ, W, C, b v, b m ) and h c s a vector of Gamma covarance latent varables, C s a flter bank and γ s a scalar parameter. energy F θ v): n hc γ log + φc k) ) n hm Ths leads to the free n v log + e φm j ) b v v + n v v, 4) where φ c k = n v C kv and φ m j = n v W jv + b m j. The correspondng score matchng objectve Jθ) s equvalent to the objectve gven n Equaton wth the followng redefnton of terms: where I nhc P = I nhc ĥ c k =γϕφc k) 5) ˆ h m j =sgmφ m j ) 6) ϕx) := + x) ρx) :=x, s the n hc n hc dentty matrx. In each of these examples, we see that an objectve emerges whch seeks to mnmze a form of regularzed reconstructon error, and that the forms of these regularzers can end up beng qute dfferent. Rather than tryng to nterpret score matchng on a case by case bass, we provde a general theorem for all latent EBMs on whch score matchng can be appled: Theorem The score matchng objectve, Equaton 4), for a latent energy based model can be expressed succnctly n terms of ether the free energy or expectatons of the energy wth respect to the condtonal dstrbuton ph v). Specfcally, Jθ) =E pv) =E pv) + var pθ h v) ψ p θ v))) + ψ p θ v)) v ) Eθ v, h) E pθ h v) v Eθ v, h) E θ v, h) E pθ h v) v v Corollary If the energy functon of a latent EBM E θ v, h) takes the followng form: E θ v, h) = v µh))t Ωh)v µh)) + gh), where µh) s an arbtrary vector-valued functon of length n v, gh) s an arbtrary scalar functon, and.

5 Autoencoders and Score Matchng Ωh) s an n v n v postve-defnte matrx-valued functon, then the vector-valued score functon ψp θ v)) wll be: E pθ h v) Ωh)v µh)). As a result, the score matchng objectve can be expressed as: ) Jθ) =E pv) Epθ h v) Ωh)v µh)) +var pθ h v) Ωh)v µh)) E pθ h v) Ωh). The proofs of Theorem and Corollary are straghtforward, and can be found n an onlne appendx to ths paper. Corollary states that score matchng appled to a Gaussan latent EBM wll always result n a quadratc reconstructon term wth penaltes to mnmze the varance of the reconstructon and to maxmze the expected curvature of the energy wth respect to v. Ths shows that we can develop new autoencoder archtectures n a prncpled way by smply startng wth an EBM and applyng score matchng. One further connecton between the two models s that one step of gradent descent on the free energy F θ v) of an EBM corresponds to one feed-forward step of an autoencoder. To see ths, consder the mpot model. If we start at some vsble confguraton v and update a sngle dmenson : v t+) = v t) = v t) n hm + η F θv) v nhc + η ĥ c k D k hˆ m j W j + b v v t). Then settng η =, the v t) terms cancel and we get: n hc n v t+) = ĥ c k D hm k + hˆ m j W j + b v. 7) Ths corresponds to the reconstructon produced by mpot n ts score matchng objectve. In general, an autoencoder reconstructon can be produced by takng a sngle step of gradent descent along the free energy of ts correspondng EBM. smpaper-appendx.pdf 4. Experments In ths secton, we study several estmaton methods appled to the mpot model ncludng SM, SMD, CD, PCD, and FPCD wth the goal of uncoverng dfferences n the characterstcs of traned models due to varatons n tranng methods. For our experments, we used two datasets of mages. The frst dataset conssts of 8,000 color mage patches of sze 6x6 pxels randomly extracted from the Berkeley segmentaton dataset. We subtracted the per-patch means and appled PCA whtenng. We retaned 99% of the varance, correspondng to 05 egeectors. All estmaton methods were appled to the mpot model by tranng on mn-batches of sze 8 for 00 epochs of stochastc gradent descent. The second dataset, named CIFAR 0 Krzhevsky, 009), conssts of color mages of sze 3x3 pxels belongng to one of 0 categores. The task s to classfy a set of 0,000 test mages. CIFAR 0 s a subset of a larger dataset of tny mages Torralba et al., 008). Usng a protocol establshed n prevous work Krzhevsky, 009; Ranzato & Hnton, 00) we bult a tranng dataset of 8x8 color mage patches from ths larger dataset, ensurng there was no overlap wth CIFAR 0. The preprocessng of the data s exactly the same as for the Berkeley dataset, but here we use approxmately 800,000 mage patches and perform only 0 epochs of tranng. For our experments, we used the Theano package 3, and mpot 4 code from Ranzato et al., 00b). 4.. Objectve Functon Analyss From Corollary, we know that we can nterpret score matchng for mpot as tradng off reconstructon error, reconstructon varance and the expected curvature of the energy functon wth respect to the vsble unts. Ths experment, usng the Berkeley dataset, s desgned to determne how these terms evolve over the course of tranng and to what degree ther changes mpact the fnal model. Fgures a) and b) show the values of the three terms usng non-nosy nputs on each tranng epoch, as well as the overall objectve functon the sum of the 3 terms). Surprsngly, these results show that most of the tranng s olved wth maxmzng the expected curvature correspondng to a lower negatve curvature). In SM, each pont groupng/segbench/ publcatons/mpot/mpot.html

6 Autoencoders and Score Matchng 0.5 Total Recon Var Curve 00 0 Total Recon Var Curve Total Recon Var Curve Value Value Value a) SM terms 0 4 ) b) SMD terms c) Autoencoder terms Free energy dfference FPCD PCD CD SM SMD MSE FPCD PCD CD SM SMD MSE FPCD PCD CD SM SMD d) Free energy dfference e) Mean-feld denosng f) Bayesan denosng Fgure. a), b), c) Expected reconstructon error, reconstructon varance, and energy curvature for SM, SMD, and AE. Total represents the sum of these terms. d) Dfference of free energy between nosy and test mages. e) MSE of denosed test mages usng mean-feld. f) MSE of denosed test mages usng Bayesan MAP. s relatvely solated n v-space meanng that the objectve wll try to make the dstrbuton very peaked. In SMD, each pont exsts near a cloud of ponts and so the dstrbuton must be broader. From ths perspectve, SMD can be seen as a regularzed verson of SM that puts less emphass on changng the expected curvature. Ths also seems to gve SMD some room to reduce the reconstructon error. To examne the mpact of regularzaton, we traned an autoencoder AE) based on the mpot model usng the reconstructon gven by Equaton 7, whch corresponds to SM wthout the varance and curvature terms. Fgure c) shows that smply optmzng the reconstructon leaves the curvature almost arant, whch agrees wth the fndngs of Ranzato et al., 007). 4.. Denosng In our next set of experments, we compare models learned by each of the score matchng estmators wth models learned by the more commonly used stochastc estmators. For these experments, we traned mpot models correspondng to SM, SMD, FPCD, PCD, and CD. We compare the models n terms of the average free energy dfference between natural mage patches and patches corrupted by Gaussan nose. consder denosng natural mage patches. 5 We also Durng tranng, we hope that the probablty of natural mages wll ncrease whle that of other mages decreases. The free energy dfference between natural and other mages s equvalent to the log of ther probablty rato, so we expect the free energy dfference to ncrease durng tranng as well. Fgure d) shows the dfference n free energy between a test set of 0,000 mage patches from the Berkeley dataset, and the energy of the same mages corrupted by nose. For most estmators, the free energy dfference mproves as tranng proceeds, as expected. Interestngly, SM and SMD exhbt completely opposte behavors. SM seems to sgnfcantly ncrease the free energy dfference relatve to nearby nosy mages, correspondng to a dstrbuton that s peaked around natural mages. SMD, on the other hand, actually decreases the free energy dfference relatve to nearby nosy mages. In the next experment, we consder an mage denosng task. We take an mage patch v and add Gaussan whte nose, obtanng a nosy patch v. We then ap- 5 Note that for coenence, both tasks were performed n the PCA doman. We use a standard devaton of for the Gaussan nose n all cases.

7 Autoencoders and Score Matchng a) Mean flters b) Covarance flters Fgure. mpot flters learned usng dfferent estmaton methods: a) mean flters, b) covarance flters. ply each model to denose each patch v, obtanng a reconstructon ˆv. The frst denosng method, shown n Fgure e), computes a reconstructon ˆv by smulatng one step of a Markov chan usng a mean-feld approxmaton. That s, we frst compute h c k and hm j by Equatons 5 and 6 usng v as the nput. The reconstructon s the expectaton of the condtonal dstrbuton P θ v h c k, hm j ). The second method, shown n Fgure f), s the Bayesan MAP estmator: ˆv = arg mn F θ v) + λ v v v, 8) where λ s a scalar representng how close the reconstructon should reman to the nosy nput. We select λ by cross-valdaton. The results show that score matchng acheves the mnmum error usng both denosng approaches, however t quckly overfts as tranng proceeds. FPCD and PCD do not match the mnmum error of SM and also overft, albet to a lesser extent. CD and SMD do not appear to overft. However, we note that the mnmum error obtaned by SMD s sgnfcantly hgher than the mnmum error obtaned by SM usng both denosng methods. Ths s qute ntutve snce SMD s equvalent to estmatng the model usng a smoothed tranng dstrbuton that shfts mass onto nearby nosy mages Feature Extracton and Classfcaton One of the prmary uses for latent EBMs s to generate dscrmnatve features. Table shows the result of usng each method to extract features on the benchmark CIFAR 0 dataset. We follow the protocol of Ranzato & Hnton, 00) wth early stoppng. We use a valdaton set to select regularzaton parame- Table. Recognton accuracy on CIFAR 0. CD PCD FPCD SM SMD AE 64.6% 64.7% 65.5% 65.0% 64.7% 57.6% ters. Wth the excepton of AE, all methods appear to do well and the dfferences between them are not statstcally sgnfcant. AE, on the other hand, does sgnfcantly worse. Fnally, we show examples of flters learned by each method. Fgure a) shows a random subset of mean flters correspondng to the columns of W, whle Fgure b) shows a random subset of covarance flters correspondng to the columns of C. Interestngly, only FPCD and PCD show structure n the learned mean flters. In the covarance unts, all methods except AE learn localzed Gabor-lke flters. It s well known that obtanng nce lookng flters wll usually correlate wth good performance, but t s not always clear what leads to these flters. We have shown here that one way to obtan good qualtatve and quanttatve performance s to focus on approprately modelng the curvature of the energy wth respect to v. In ths context, the SM reconstructon and varance terms serve to ensure that the peaks of the dstrbuton occur around the tranng cases. 5. Concluson By applyng score matchng to the energy space of a latent EBM, as opposed to the free energy space, we gan an ntutve nterpretaton of the score matchng objectve. We can always break the objectve down nto three terms correspondng to expectatons under the condtonal dstrbuton of the hdden unts: reconstructon, reconstructon varance, and curvature. We have determned that for the Gaussan-bnary RBM, the reconstructon term wll always correspond to an autoencoder wth ted weghts. Whle autoencoders and RBMs were prevously consdered to be related, but separate models, ths analyss shows that they can be nterpreted as dfferent estmators appled to the same underlyng model. We also showed that one can derve novel autoencoders by applyng score matchng to more complex EBMs. Ths allows us to

8 Autoencoders and Score Matchng thnk about models n terms of EBMs before creatng a correspondng autoencoder to leverage fast nference. Furthermore, ths framework provdes gudance on selectng prncpled regularzaton functons for autoencoder tranng, leadng to mproved representatons. Our experments show that not only does score matchng yeld smlar performance to exstng estmaton methods when appled to classfcaton, but that shapng the curvature of the energy approprately may be mportant for generatng good features. Whle ths seems obvous for probablstc EBMs, t has prevously been dffcult to apply to autoencoders because they were not thought of as havng a correspondng energy functon. Now that we know whch statstcs may be mportant to montor durng tranng, t would be nterestng to see what happens when other heurstcs, such as sparsty, are appled to help generate nterpretable features. References Bengo, Y. Learnng deep archtectures for AI. Foundatons and Trends n Machne Learnng, ): 7, 009. Hnton, G.E. Tranng products of experts by mnmzng contrastve dvergence. Neural Computaton, 4:77 800, 00. Hnton, G.E. and Zemel, R.S. Autoencoders, mnmum descrpton length and Helmholtz free energy. In Advances n Neural Informaton Processng Systems, pp. 3 0, 994. Hnton, G.E., Osndero, S., and Teh, Y.W. A fast learnng algorthm for deep belef nets. Neural Computaton, 8 7):57 554, 006. Hyvärnen, A. Estmaton of non-normalzed statstcal models usng score matchng. Journal of Machne Learnng Research, 6: , 005. Kngma, D. and LeCun, Y. Regularzed estmaton of mage statstcs by score matchng. In Advances n Neural Informaton Processng Systems, 00. Krzhevsky, A. Learnng multple layers of features from tny mages, 009. MSc Thess, Dept. of Comp. Scence, U. of Toronto. LeCun, Y., Chopra, S., Hadsell, R., Ranzato, M., and Huang, F.J. A tutoral on energy-based learnng. In Predctng Structured Data. MIT Press, 006. Marln, B.M., Swersky, K., Chen, B., and de Fretas, N. Inductve prncples for restrcted Boltzmann machne learnng. In Artfcal Intellgence and Statstcs, pp , 00. Memsevc, R. and Hnton, G.E. Learnng to represent spatal transformatons wth factored hgher-order Boltzmann machnes. Neural Computaton, :473 49, 009. Ranzato, M. and Hnton, G.E. Modelng pxel means and covarances usng factorzed thrd-order Boltzmann machnes. In IEEE Computer Vson and Pattern Recognton, pp , 00. Ranzato, M., Boureau, Y.L., Chopra, S., and LeCun, Y. A unfed energy-based framework for unsupervsed learnng. In Artfcal Intellgence and Statstcs, 007. Ranzato, M., Krzhevsky, A., and Hnton, G.E. Factored 3- way restrcted Boltzmann machnes for modelng natural mages. In Artfcal Intellgence and Statstcs, pp. 6 68, 00a. Ranzato, M., Mnh, V., and Hnton, G.E. How to generate realstc mages usng gated MRF s. In Advances n Neural Informaton Processng Systems, pp , 00b. Schmdt, U., Gao, Q., and Roth, S. A generatve perspectve on MRFs n low-level vson. In IEEE Computer Vson and Pattern Recognton, 00. Swersky, K., Chen, B., Marln, B.M., and de Fretas, N. A tutoral on stochastc approxmaton algorthms for tranng restrcted Boltzmann machnes and deep belef nets. In Informaton Theory and Applcatons Workshop, pp. 0, 00. Teh, Y.W., Wellng, M., Osndero, S., and Hnton, G.E. Energy-based models for sparse overcomplete representatons. Journal of Machne Learnng Research, 4:35 60, 003. Teleman, T. Tranng restrcted Boltzmann machnes usng approxmatons to the lkelhood gradent. In Internatonal Conference on Machne Learnng, pp , 008. Teleman, T. and Hnton, G.E. Usng fast weghts to mprove persstent contrastve dvergence. In Internatonal Conference on Machne Learnng, 009. Torralba, A., Fergus, R., and Freeman, W.T. 80 mllon tny mages: A large dataset for non-parametrc object and scene recognton. IEEE Transactons on Pattern Analyss and Machne Intellgence, 30: , 008. Vncent, P. A connecton between score matchng and denosng autoencoders. Neural Computaton, To appear, 0. Vncent, P., Larochelle, H., Bengo, Y., and Manzagol, P.A. Extractng and composng robust features wth denosng autoencoders. In Internatonal Conference on Machne Learnng, pp , 008. Wellng, M., Hnton, G.E., and Osndero, S. Learnng sparse topographc representatons wth products of student-t dstrbutons. In Advances n Neural Informaton Processng Systems, 003. Wellng, M., Rosen-Zv, M., and Hnton, G.E. Exponental famly harmonums wth an applcaton to nformaton retreval. In Advances n Neural Informaton Processng Systems, 005. Younes, L. Parametrc nference for mperfectly observed Gbbsan felds. Probablty Theory and Related Felds, 84):65 645, 989.

What is Candidate Sampling

What is Candidate Sampling What s Canddate Samplng Say we have a multclass or mult label problem where each tranng example ( x, T ) conssts of a context x a small (mult)set of target classes T out of a large unverse L of possble

More information

CS 2750 Machine Learning. Lecture 3. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 3. Density estimation. CS 2750 Machine Learning. Announcements Lecture 3 Densty estmaton Mlos Hauskrecht [email protected] 5329 Sennott Square Next lecture: Matlab tutoral Announcements Rules for attendng the class: Regstered for credt Regstered for audt (only f there

More information

L10: Linear discriminants analysis

L10: Linear discriminants analysis L0: Lnear dscrmnants analyss Lnear dscrmnant analyss, two classes Lnear dscrmnant analyss, C classes LDA vs. PCA Lmtatons of LDA Varants of LDA Other dmensonalty reducton methods CSCE 666 Pattern Analyss

More information

How To Calculate The Accountng Perod Of Nequalty

How To Calculate The Accountng Perod Of Nequalty Inequalty and The Accountng Perod Quentn Wodon and Shlomo Ytzha World Ban and Hebrew Unversty September Abstract Income nequalty typcally declnes wth the length of tme taen nto account for measurement.

More information

Forecasting the Direction and Strength of Stock Market Movement

Forecasting the Direction and Strength of Stock Market Movement Forecastng the Drecton and Strength of Stock Market Movement Jngwe Chen Mng Chen Nan Ye [email protected] [email protected] [email protected] Abstract - Stock market s one of the most complcated systems

More information

Face Verification Problem. Face Recognition Problem. Application: Access Control. Biometric Authentication. Face Verification (1:1 matching)

Face Verification Problem. Face Recognition Problem. Application: Access Control. Biometric Authentication. Face Verification (1:1 matching) Face Recognton Problem Face Verfcaton Problem Face Verfcaton (1:1 matchng) Querymage face query Face Recognton (1:N matchng) database Applcaton: Access Control www.vsage.com www.vsoncs.com Bometrc Authentcaton

More information

Lecture 2: Single Layer Perceptrons Kevin Swingler

Lecture 2: Single Layer Perceptrons Kevin Swingler Lecture 2: Sngle Layer Perceptrons Kevn Sngler [email protected] Recap: McCulloch-Ptts Neuron Ths vastly smplfed model of real neurons s also knon as a Threshold Logc Unt: W 2 A Y 3 n W n. A set of synapses

More information

Logistic Regression. Steve Kroon

Logistic Regression. Steve Kroon Logstc Regresson Steve Kroon Course notes sectons: 24.3-24.4 Dsclamer: these notes do not explctly ndcate whether values are vectors or scalars, but expects the reader to dscern ths from the context. Scenaro

More information

Dropout: A Simple Way to Prevent Neural Networks from Overfitting

Dropout: A Simple Way to Prevent Neural Networks from Overfitting Journal of Machne Learnng Research 15 (2014) 1929-1958 Submtted 11/13; Publshed 6/14 Dropout: A Smple Way to Prevent Neural Networks from Overfttng Ntsh Srvastava Geoffrey Hnton Alex Krzhevsky Ilya Sutskever

More information

Logistic Regression. Lecture 4: More classifiers and classes. Logistic regression. Adaboost. Optimization. Multiple class classification

Logistic Regression. Lecture 4: More classifiers and classes. Logistic regression. Adaboost. Optimization. Multiple class classification Lecture 4: More classfers and classes C4B Machne Learnng Hlary 20 A. Zsserman Logstc regresson Loss functons revsted Adaboost Loss functons revsted Optmzaton Multple class classfcaton Logstc Regresson

More information

Feature selection for intrusion detection. Slobodan Petrović NISlab, Gjøvik University College

Feature selection for intrusion detection. Slobodan Petrović NISlab, Gjøvik University College Feature selecton for ntruson detecton Slobodan Petrovć NISlab, Gjøvk Unversty College Contents The feature selecton problem Intruson detecton Traffc features relevant for IDS The CFS measure The mrmr measure

More information

The Development of Web Log Mining Based on Improve-K-Means Clustering Analysis

The Development of Web Log Mining Based on Improve-K-Means Clustering Analysis The Development of Web Log Mnng Based on Improve-K-Means Clusterng Analyss TngZhong Wang * College of Informaton Technology, Luoyang Normal Unversty, Luoyang, 471022, Chna [email protected] Abstract.

More information

Forecasting the Demand of Emergency Supplies: Based on the CBR Theory and BP Neural Network

Forecasting the Demand of Emergency Supplies: Based on the CBR Theory and BP Neural Network 700 Proceedngs of the 8th Internatonal Conference on Innovaton & Management Forecastng the Demand of Emergency Supples: Based on the CBR Theory and BP Neural Network Fu Deqang, Lu Yun, L Changbng School

More information

benefit is 2, paid if the policyholder dies within the year, and probability of death within the year is ).

benefit is 2, paid if the policyholder dies within the year, and probability of death within the year is ). REVIEW OF RISK MANAGEMENT CONCEPTS LOSS DISTRIBUTIONS AND INSURANCE Loss and nsurance: When someone s subject to the rsk of ncurrng a fnancal loss, the loss s generally modeled usng a random varable or

More information

Support Vector Machines

Support Vector Machines Support Vector Machnes Max Wellng Department of Computer Scence Unversty of Toronto 10 Kng s College Road Toronto, M5S 3G5 Canada [email protected] Abstract Ths s a note to explan support vector machnes.

More information

Luby s Alg. for Maximal Independent Sets using Pairwise Independence

Luby s Alg. for Maximal Independent Sets using Pairwise Independence Lecture Notes for Randomzed Algorthms Luby s Alg. for Maxmal Independent Sets usng Parwse Independence Last Updated by Erc Vgoda on February, 006 8. Maxmal Independent Sets For a graph G = (V, E), an ndependent

More information

The OC Curve of Attribute Acceptance Plans

The OC Curve of Attribute Acceptance Plans The OC Curve of Attrbute Acceptance Plans The Operatng Characterstc (OC) curve descrbes the probablty of acceptng a lot as a functon of the lot s qualty. Fgure 1 shows a typcal OC Curve. 10 8 6 4 1 3 4

More information

v a 1 b 1 i, a 2 b 2 i,..., a n b n i.

v a 1 b 1 i, a 2 b 2 i,..., a n b n i. SECTION 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS 455 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS All the vector spaces we have studed thus far n the text are real vector spaces snce the scalars are

More information

Can Auto Liability Insurance Purchases Signal Risk Attitude?

Can Auto Liability Insurance Purchases Signal Risk Attitude? Internatonal Journal of Busness and Economcs, 2011, Vol. 10, No. 2, 159-164 Can Auto Lablty Insurance Purchases Sgnal Rsk Atttude? Chu-Shu L Department of Internatonal Busness, Asa Unversty, Tawan Sheng-Chang

More information

An Alternative Way to Measure Private Equity Performance

An Alternative Way to Measure Private Equity Performance An Alternatve Way to Measure Prvate Equty Performance Peter Todd Parlux Investment Technology LLC Summary Internal Rate of Return (IRR) s probably the most common way to measure the performance of prvate

More information

Recurrence. 1 Definitions and main statements

Recurrence. 1 Definitions and main statements Recurrence 1 Defntons and man statements Let X n, n = 0, 1, 2,... be a MC wth the state space S = (1, 2,...), transton probabltes p j = P {X n+1 = j X n = }, and the transton matrx P = (p j ),j S def.

More information

Data Visualization by Pairwise Distortion Minimization

Data Visualization by Pairwise Distortion Minimization Communcatons n Statstcs, Theory and Methods 34 (6), 005 Data Vsualzaton by Parwse Dstorton Mnmzaton By Marc Sobel, and Longn Jan Lateck* Department of Statstcs and Department of Computer and Informaton

More information

ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING

ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING Matthew J. Lberatore, Department of Management and Operatons, Vllanova Unversty, Vllanova, PA 19085, 610-519-4390,

More information

BERNSTEIN POLYNOMIALS

BERNSTEIN POLYNOMIALS On-Lne Geometrc Modelng Notes BERNSTEIN POLYNOMIALS Kenneth I. Joy Vsualzaton and Graphcs Research Group Department of Computer Scence Unversty of Calforna, Davs Overvew Polynomals are ncredbly useful

More information

The Application of Fractional Brownian Motion in Option Pricing

The Application of Fractional Brownian Motion in Option Pricing Vol. 0, No. (05), pp. 73-8 http://dx.do.org/0.457/jmue.05.0..6 The Applcaton of Fractonal Brownan Moton n Opton Prcng Qng-xn Zhou School of Basc Scence,arbn Unversty of Commerce,arbn [email protected]

More information

Single and multiple stage classifiers implementing logistic discrimination

Single and multiple stage classifiers implementing logistic discrimination Sngle and multple stage classfers mplementng logstc dscrmnaton Hélo Radke Bttencourt 1 Dens Alter de Olvera Moraes 2 Vctor Haertel 2 1 Pontfíca Unversdade Católca do Ro Grande do Sul - PUCRS Av. Ipranga,

More information

CHOLESTEROL REFERENCE METHOD LABORATORY NETWORK. Sample Stability Protocol

CHOLESTEROL REFERENCE METHOD LABORATORY NETWORK. Sample Stability Protocol CHOLESTEROL REFERENCE METHOD LABORATORY NETWORK Sample Stablty Protocol Background The Cholesterol Reference Method Laboratory Network (CRMLN) developed certfcaton protocols for total cholesterol, HDL

More information

A Multi-mode Image Tracking System Based on Distributed Fusion

A Multi-mode Image Tracking System Based on Distributed Fusion A Mult-mode Image Tracng System Based on Dstrbuted Fuson Ln zheng Chongzhao Han Dongguang Zuo Hongsen Yan School of Electroncs & nformaton engneerng, X an Jaotong Unversty X an, Shaanx, Chna [email protected]

More information

1 Example 1: Axis-aligned rectangles

1 Example 1: Axis-aligned rectangles COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture # 6 Scrbe: Aaron Schld February 21, 2013 Last class, we dscussed an analogue for Occam s Razor for nfnte hypothess spaces that, n conjuncton

More information

Quantization Effects in Digital Filters

Quantization Effects in Digital Filters Quantzaton Effects n Dgtal Flters Dstrbuton of Truncaton Errors In two's complement representaton an exact number would have nfntely many bts (n general). When we lmt the number of bts to some fnte value

More information

Vision Mouse. Saurabh Sarkar a* University of Cincinnati, Cincinnati, USA ABSTRACT 1. INTRODUCTION

Vision Mouse. Saurabh Sarkar a* University of Cincinnati, Cincinnati, USA ABSTRACT 1. INTRODUCTION Vson Mouse Saurabh Sarkar a* a Unversty of Cncnnat, Cncnnat, USA ABSTRACT The report dscusses a vson based approach towards trackng of eyes and fngers. The report descrbes the process of locatng the possble

More information

Title Language Model for Information Retrieval

Title Language Model for Information Retrieval Ttle Language Model for Informaton Retreval Rong Jn Language Technologes Insttute School of Computer Scence Carnege Mellon Unversty Alex G. Hauptmann Computer Scence Department School of Computer Scence

More information

Sketching Sampled Data Streams

Sketching Sampled Data Streams Sketchng Sampled Data Streams Florn Rusu, Aln Dobra CISE Department Unversty of Florda Ganesvlle, FL, USA [email protected] [email protected] Abstract Samplng s used as a unversal method to reduce the

More information

Learning from Multiple Outlooks

Learning from Multiple Outlooks Learnng from Multple Outlooks Maayan Harel Department of Electrcal Engneerng, Technon, Hafa, Israel She Mannor Department of Electrcal Engneerng, Technon, Hafa, Israel [email protected] [email protected]

More information

Statistical Methods to Develop Rating Models

Statistical Methods to Develop Rating Models Statstcal Methods to Develop Ratng Models [Evelyn Hayden and Danel Porath, Österrechsche Natonalbank and Unversty of Appled Scences at Manz] Source: The Basel II Rsk Parameters Estmaton, Valdaton, and

More information

ECE544NA Final Project: Robust Machine Learning Hardware via Classifier Ensemble

ECE544NA Final Project: Robust Machine Learning Hardware via Classifier Ensemble 1 ECE544NA Fnal Project: Robust Machne Learnng Hardware va Classfer Ensemble Sa Zhang, [email protected] Dept. of Electr. & Comput. Eng., Unv. of Illnos at Urbana-Champagn, Urbana, IL, USA Abstract In

More information

How Sets of Coherent Probabilities May Serve as Models for Degrees of Incoherence

How Sets of Coherent Probabilities May Serve as Models for Degrees of Incoherence 1 st Internatonal Symposum on Imprecse Probabltes and Ther Applcatons, Ghent, Belgum, 29 June 2 July 1999 How Sets of Coherent Probabltes May Serve as Models for Degrees of Incoherence Mar J. Schervsh

More information

A Continuous Restricted Boltzmann Machine with a Hardware-Amenable Learning Algorithm

A Continuous Restricted Boltzmann Machine with a Hardware-Amenable Learning Algorithm A Contnuous Restrcted Boltzmann Machne wth a Hardware-Amenable Learnng Algorthm Hsn Chen and Alan Murray Dept. of Electroncs and Electrcal Engneerng, Unversty of Ednburgh, Mayfeld Rd., Ednburgh, EH93JL,

More information

Institute of Informatics, Faculty of Business and Management, Brno University of Technology,Czech Republic

Institute of Informatics, Faculty of Business and Management, Brno University of Technology,Czech Republic Lagrange Multplers as Quanttatve Indcators n Economcs Ivan Mezník Insttute of Informatcs, Faculty of Busness and Management, Brno Unversty of TechnologCzech Republc Abstract The quanttatve role of Lagrange

More information

Stochastic Protocol Modeling for Anomaly Based Network Intrusion Detection

Stochastic Protocol Modeling for Anomaly Based Network Intrusion Detection Stochastc Protocol Modelng for Anomaly Based Network Intruson Detecton Juan M. Estevez-Tapador, Pedro Garca-Teodoro, and Jesus E. Daz-Verdejo Department of Electroncs and Computer Technology Unversty of

More information

Calculation of Sampling Weights

Calculation of Sampling Weights Perre Foy Statstcs Canada 4 Calculaton of Samplng Weghts 4.1 OVERVIEW The basc sample desgn used n TIMSS Populatons 1 and 2 was a two-stage stratfed cluster desgn. 1 The frst stage conssted of a sample

More information

How To Understand The Results Of The German Meris Cloud And Water Vapour Product

How To Understand The Results Of The German Meris Cloud And Water Vapour Product Ttel: Project: Doc. No.: MERIS level 3 cloud and water vapour products MAPP MAPP-ATBD-ClWVL3 Issue: 1 Revson: 0 Date: 9.12.1998 Functon Name Organsaton Sgnature Date Author: Bennartz FUB Preusker FUB Schüller

More information

CHAPTER 5 RELATIONSHIPS BETWEEN QUANTITATIVE VARIABLES

CHAPTER 5 RELATIONSHIPS BETWEEN QUANTITATIVE VARIABLES CHAPTER 5 RELATIONSHIPS BETWEEN QUANTITATIVE VARIABLES In ths chapter, we wll learn how to descrbe the relatonshp between two quanttatve varables. Remember (from Chapter 2) that the terms quanttatve varable

More information

Active Learning for Interactive Visualization

Active Learning for Interactive Visualization Actve Learnng for Interactve Vsualzaton Tomoharu Iwata Nel Houlsby Zoubn Ghahraman Unversty of Cambrdge Unversty of Cambrdge Unversty of Cambrdge Abstract Many automatc vsualzaton methods have been. However,

More information

Fast Fuzzy Clustering of Web Page Collections

Fast Fuzzy Clustering of Web Page Collections Fast Fuzzy Clusterng of Web Page Collectons Chrstan Borgelt and Andreas Nürnberger Dept. of Knowledge Processng and Language Engneerng Otto-von-Guercke-Unversty of Magdeburg Unverstätsplatz, D-396 Magdeburg,

More information

NPAR TESTS. One-Sample Chi-Square Test. Cell Specification. Observed Frequencies 1O i 6. Expected Frequencies 1EXP i 6

NPAR TESTS. One-Sample Chi-Square Test. Cell Specification. Observed Frequencies 1O i 6. Expected Frequencies 1EXP i 6 PAR TESTS If a WEIGHT varable s specfed, t s used to replcate a case as many tmes as ndcated by the weght value rounded to the nearest nteger. If the workspace requrements are exceeded and samplng has

More information

1. Measuring association using correlation and regression

1. Measuring association using correlation and regression How to measure assocaton I: Correlaton. 1. Measurng assocaton usng correlaton and regresson We often would lke to know how one varable, such as a mother's weght, s related to another varable, such as a

More information

Georey E. Hinton. University oftoronto. Email: [email protected]. Technical Report CRG-TR-96-1. May 21, 1996 (revised Feb 27, 1997) Abstract

Georey E. Hinton. University oftoronto. Email: zoubin@cs.toronto.edu. Technical Report CRG-TR-96-1. May 21, 1996 (revised Feb 27, 1997) Abstract The EM Algorthm for Mxtures of Factor Analyzers Zoubn Ghahraman Georey E. Hnton Department of Computer Scence Unversty oftoronto 6 Kng's College Road Toronto, Canada M5S A4 Emal: [email protected] Techncal

More information

CHAPTER 14 MORE ABOUT REGRESSION

CHAPTER 14 MORE ABOUT REGRESSION CHAPTER 14 MORE ABOUT REGRESSION We learned n Chapter 5 that often a straght lne descrbes the pattern of a relatonshp between two quanttatve varables. For nstance, n Example 5.1 we explored the relatonshp

More information

An Algorithm for Data-Driven Bandwidth Selection

An Algorithm for Data-Driven Bandwidth Selection IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 25, NO. 2, FEBRUARY 2003 An Algorthm for Data-Drven Bandwdth Selecton Dorn Comancu, Member, IEEE Abstract The analyss of a feature space

More information

A hybrid global optimization algorithm based on parallel chaos optimization and outlook algorithm

A hybrid global optimization algorithm based on parallel chaos optimization and outlook algorithm Avalable onlne www.ocpr.com Journal of Chemcal and Pharmaceutcal Research, 2014, 6(7):1884-1889 Research Artcle ISSN : 0975-7384 CODEN(USA) : JCPRC5 A hybrd global optmzaton algorthm based on parallel

More information

HÜCKEL MOLECULAR ORBITAL THEORY

HÜCKEL MOLECULAR ORBITAL THEORY 1 HÜCKEL MOLECULAR ORBITAL THEORY In general, the vast maorty polyatomc molecules can be thought of as consstng of a collecton of two electron bonds between pars of atoms. So the qualtatve pcture of σ

More information

An Interest-Oriented Network Evolution Mechanism for Online Communities

An Interest-Oriented Network Evolution Mechanism for Online Communities An Interest-Orented Network Evoluton Mechansm for Onlne Communtes Cahong Sun and Xaopng Yang School of Informaton, Renmn Unversty of Chna, Bejng 100872, P.R. Chna {chsun,yang}@ruc.edu.cn Abstract. Onlne

More information

IMPACT ANALYSIS OF A CELLULAR PHONE

IMPACT ANALYSIS OF A CELLULAR PHONE 4 th ASA & μeta Internatonal Conference IMPACT AALYSIS OF A CELLULAR PHOE We Lu, 2 Hongy L Bejng FEAonlne Engneerng Co.,Ltd. Bejng, Chna ABSTRACT Drop test smulaton plays an mportant role n nvestgatng

More information

Causal, Explanatory Forecasting. Analysis. Regression Analysis. Simple Linear Regression. Which is Independent? Forecasting

Causal, Explanatory Forecasting. Analysis. Regression Analysis. Simple Linear Regression. Which is Independent? Forecasting Causal, Explanatory Forecastng Assumes cause-and-effect relatonshp between system nputs and ts output Forecastng wth Regresson Analyss Rchard S. Barr Inputs System Cause + Effect Relatonshp The job of

More information

8.5 UNITARY AND HERMITIAN MATRICES. The conjugate transpose of a complex matrix A, denoted by A*, is given by

8.5 UNITARY AND HERMITIAN MATRICES. The conjugate transpose of a complex matrix A, denoted by A*, is given by 6 CHAPTER 8 COMPLEX VECTOR SPACES 5. Fnd the kernel of the lnear transformaton gven n Exercse 5. In Exercses 55 and 56, fnd the mage of v, for the ndcated composton, where and are gven by the followng

More information

How To Know The Components Of Mean Squared Error Of Herarchcal Estmator S

How To Know The Components Of Mean Squared Error Of Herarchcal Estmator S S C H E D A E I N F O R M A T I C A E VOLUME 0 0 On Mean Squared Error of Herarchcal Estmator Stans law Brodowsk Faculty of Physcs, Astronomy, and Appled Computer Scence, Jagellonan Unversty, Reymonta

More information

An artificial Neural Network approach to monitor and diagnose multi-attribute quality control processes. S. T. A. Niaki*

An artificial Neural Network approach to monitor and diagnose multi-attribute quality control processes. S. T. A. Niaki* Journal of Industral Engneerng Internatonal July 008, Vol. 4, No. 7, 04 Islamc Azad Unversty, South Tehran Branch An artfcal Neural Network approach to montor and dagnose multattrbute qualty control processes

More information

The Greedy Method. Introduction. 0/1 Knapsack Problem

The Greedy Method. Introduction. 0/1 Knapsack Problem The Greedy Method Introducton We have completed data structures. We now are gong to look at algorthm desgn methods. Often we are lookng at optmzaton problems whose performance s exponental. For an optmzaton

More information

Conversion between the vector and raster data structures using Fuzzy Geographical Entities

Conversion between the vector and raster data structures using Fuzzy Geographical Entities Converson between the vector and raster data structures usng Fuzzy Geographcal Enttes Cdála Fonte Department of Mathematcs Faculty of Scences and Technology Unversty of Combra, Apartado 38, 3 454 Combra,

More information

Ring structure of splines on triangulations

Ring structure of splines on triangulations www.oeaw.ac.at Rng structure of splnes on trangulatons N. Vllamzar RICAM-Report 2014-48 www.rcam.oeaw.ac.at RING STRUCTURE OF SPLINES ON TRIANGULATIONS NELLY VILLAMIZAR Introducton For a trangulated regon

More information

Risk-based Fatigue Estimate of Deep Water Risers -- Course Project for EM388F: Fracture Mechanics, Spring 2008

Risk-based Fatigue Estimate of Deep Water Risers -- Course Project for EM388F: Fracture Mechanics, Spring 2008 Rsk-based Fatgue Estmate of Deep Water Rsers -- Course Project for EM388F: Fracture Mechancs, Sprng 2008 Chen Sh Department of Cvl, Archtectural, and Envronmental Engneerng The Unversty of Texas at Austn

More information

Descriptive Models. Cluster Analysis. Example. General Applications of Clustering. Examples of Clustering Applications

Descriptive Models. Cluster Analysis. Example. General Applications of Clustering. Examples of Clustering Applications CMSC828G Prncples of Data Mnng Lecture #9 Today s Readng: HMS, chapter 9 Today s Lecture: Descrptve Modelng Clusterng Algorthms Descrptve Models model presents the man features of the data, a global summary

More information

An Enhanced Super-Resolution System with Improved Image Registration, Automatic Image Selection, and Image Enhancement

An Enhanced Super-Resolution System with Improved Image Registration, Automatic Image Selection, and Image Enhancement An Enhanced Super-Resoluton System wth Improved Image Regstraton, Automatc Image Selecton, and Image Enhancement Yu-Chuan Kuo ( ), Chen-Yu Chen ( ), and Chou-Shann Fuh ( ) Department of Computer Scence

More information

Realistic Image Synthesis

Realistic Image Synthesis Realstc Image Synthess - Combned Samplng and Path Tracng - Phlpp Slusallek Karol Myszkowsk Vncent Pegoraro Overvew: Today Combned Samplng (Multple Importance Samplng) Renderng and Measurng Equaton Random

More information

Inter-Ing 2007. INTERDISCIPLINARITY IN ENGINEERING SCIENTIFIC INTERNATIONAL CONFERENCE, TG. MUREŞ ROMÂNIA, 15-16 November 2007.

Inter-Ing 2007. INTERDISCIPLINARITY IN ENGINEERING SCIENTIFIC INTERNATIONAL CONFERENCE, TG. MUREŞ ROMÂNIA, 15-16 November 2007. Inter-Ing 2007 INTERDISCIPLINARITY IN ENGINEERING SCIENTIFIC INTERNATIONAL CONFERENCE, TG. MUREŞ ROMÂNIA, 15-16 November 2007. UNCERTAINTY REGION SIMULATION FOR A SERIAL ROBOT STRUCTURE MARIUS SEBASTIAN

More information

INVESTIGATION OF VEHICULAR USERS FAIRNESS IN CDMA-HDR NETWORKS

INVESTIGATION OF VEHICULAR USERS FAIRNESS IN CDMA-HDR NETWORKS 21 22 September 2007, BULGARIA 119 Proceedngs of the Internatonal Conference on Informaton Technologes (InfoTech-2007) 21 st 22 nd September 2007, Bulgara vol. 2 INVESTIGATION OF VEHICULAR USERS FAIRNESS

More information

Latent Class Regression. Statistics for Psychosocial Research II: Structural Models December 4 and 6, 2006

Latent Class Regression. Statistics for Psychosocial Research II: Structural Models December 4 and 6, 2006 Latent Class Regresson Statstcs for Psychosocal Research II: Structural Models December 4 and 6, 2006 Latent Class Regresson (LCR) What s t and when do we use t? Recall the standard latent class model

More information

Characterization of Assembly. Variation Analysis Methods. A Thesis. Presented to the. Department of Mechanical Engineering. Brigham Young University

Characterization of Assembly. Variation Analysis Methods. A Thesis. Presented to the. Department of Mechanical Engineering. Brigham Young University Characterzaton of Assembly Varaton Analyss Methods A Thess Presented to the Department of Mechancal Engneerng Brgham Young Unversty In Partal Fulfllment of the Requrements for the Degree Master of Scence

More information

Out-of-Sample Extensions for LLE, Isomap, MDS, Eigenmaps, and Spectral Clustering

Out-of-Sample Extensions for LLE, Isomap, MDS, Eigenmaps, and Spectral Clustering Out-of-Sample Extensons for LLE, Isomap, MDS, Egenmaps, and Spectral Clusterng Yoshua Bengo, Jean-Franços Paement, Pascal Vncent Olver Delalleau, Ncolas Le Roux and Mare Oumet Département d Informatque

More information

A study on the ability of Support Vector Regression and Neural Networks to Forecast Basic Time Series Patterns

A study on the ability of Support Vector Regression and Neural Networks to Forecast Basic Time Series Patterns A study on the ablty of Support Vector Regresson and Neural Networks to Forecast Basc Tme Seres Patterns Sven F. Crone, Jose Guajardo 2, and Rchard Weber 2 Lancaster Unversty, Department of Management

More information

Damage detection in composite laminates using coin-tap method

Damage detection in composite laminates using coin-tap method Damage detecton n composte lamnates usng con-tap method S.J. Km Korea Aerospace Research Insttute, 45 Eoeun-Dong, Youseong-Gu, 35-333 Daejeon, Republc of Korea [email protected] 45 The con-tap test has the

More information

Joint Scheduling of Processing and Shuffle Phases in MapReduce Systems

Joint Scheduling of Processing and Shuffle Phases in MapReduce Systems Jont Schedulng of Processng and Shuffle Phases n MapReduce Systems Fangfe Chen, Mural Kodalam, T. V. Lakshman Department of Computer Scence and Engneerng, The Penn State Unversty Bell Laboratores, Alcatel-Lucent

More information

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module LOSSLESS IMAGE COMPRESSION SYSTEMS Lesson 3 Lossless Compresson: Huffman Codng Instructonal Objectves At the end of ths lesson, the students should be able to:. Defne and measure source entropy..

More information

Implementation of Deutsch's Algorithm Using Mathcad

Implementation of Deutsch's Algorithm Using Mathcad Implementaton of Deutsch's Algorthm Usng Mathcad Frank Roux The followng s a Mathcad mplementaton of Davd Deutsch's quantum computer prototype as presented on pages - n "Machnes, Logc and Quantum Physcs"

More information

Trade Adjustment and Productivity in Large Crises. Online Appendix May 2013. Appendix A: Derivation of Equations for Productivity

Trade Adjustment and Productivity in Large Crises. Online Appendix May 2013. Appendix A: Derivation of Equations for Productivity Trade Adjustment Productvty n Large Crses Gta Gopnath Department of Economcs Harvard Unversty NBER Brent Neman Booth School of Busness Unversty of Chcago NBER Onlne Appendx May 2013 Appendx A: Dervaton

More information

THE DISTRIBUTION OF LOAN PORTFOLIO VALUE * Oldrich Alfons Vasicek

THE DISTRIBUTION OF LOAN PORTFOLIO VALUE * Oldrich Alfons Vasicek HE DISRIBUION OF LOAN PORFOLIO VALUE * Oldrch Alfons Vascek he amount of captal necessary to support a portfolo of debt securtes depends on the probablty dstrbuton of the portfolo loss. Consder a portfolo

More information

Mining Feature Importance: Applying Evolutionary Algorithms within a Web-based Educational System

Mining Feature Importance: Applying Evolutionary Algorithms within a Web-based Educational System Mnng Feature Importance: Applyng Evolutonary Algorthms wthn a Web-based Educatonal System Behrouz MINAEI-BIDGOLI 1, and Gerd KORTEMEYER 2, and Wllam F. PUNCH 1 1 Genetc Algorthms Research and Applcatons

More information

On-Line Fault Detection in Wind Turbine Transmission System using Adaptive Filter and Robust Statistical Features

On-Line Fault Detection in Wind Turbine Transmission System using Adaptive Filter and Robust Statistical Features On-Lne Fault Detecton n Wnd Turbne Transmsson System usng Adaptve Flter and Robust Statstcal Features Ruoyu L Remote Dagnostcs Center SKF USA Inc. 3443 N. Sam Houston Pkwy., Houston TX 77086 Emal: [email protected]

More information

Review of Hierarchical Models for Data Clustering and Visualization

Review of Hierarchical Models for Data Clustering and Visualization Revew of Herarchcal Models for Data Clusterng and Vsualzaton Lola Vcente & Alfredo Velldo Grup de Soft Computng Seccó d Intel lgènca Artfcal Departament de Llenguatges Sstemes Informàtcs Unverstat Poltècnca

More information

THE METHOD OF LEAST SQUARES THE METHOD OF LEAST SQUARES

THE METHOD OF LEAST SQUARES THE METHOD OF LEAST SQUARES The goal: to measure (determne) an unknown quantty x (the value of a RV X) Realsaton: n results: y 1, y 2,..., y j,..., y n, (the measured values of Y 1, Y 2,..., Y j,..., Y n ) every result s encumbered

More information

Machine Learning and Data Mining Lecture Notes

Machine Learning and Data Mining Lecture Notes Machne Learnng and Data Mnng Lecture Notes CSC 411/D11 Computer Scence Department Unversty of Toronto Verson: February 6, 2012 Copyrght c 2010 Aaron Hertzmann and Davd Fleet CONTENTS Contents Conventons

More information

Risk Model of Long-Term Production Scheduling in Open Pit Gold Mining

Risk Model of Long-Term Production Scheduling in Open Pit Gold Mining Rsk Model of Long-Term Producton Schedulng n Open Pt Gold Mnng R Halatchev 1 and P Lever 2 ABSTRACT Open pt gold mnng s an mportant sector of the Australan mnng ndustry. It uses large amounts of nvestments,

More information

A Secure Password-Authenticated Key Agreement Using Smart Cards

A Secure Password-Authenticated Key Agreement Using Smart Cards A Secure Password-Authentcated Key Agreement Usng Smart Cards Ka Chan 1, Wen-Chung Kuo 2 and Jn-Chou Cheng 3 1 Department of Computer and Informaton Scence, R.O.C. Mltary Academy, Kaohsung 83059, Tawan,

More information

Lecture 5,6 Linear Methods for Classification. Summary

Lecture 5,6 Linear Methods for Classification. Summary Lecture 5,6 Lnear Methods for Classfcaton Rce ELEC 697 Farnaz Koushanfar Fall 2006 Summary Bayes Classfers Lnear Classfers Lnear regresson of an ndcator matrx Lnear dscrmnant analyss (LDA) Logstc regresson

More information

Support vector domain description

Support vector domain description Pattern Recognton Letters 20 (1999) 1191±1199 www.elsever.nl/locate/patrec Support vector doman descrpton Davd M.J. Tax *,1, Robert P.W. Dun Pattern Recognton Group, Faculty of Appled Scence, Delft Unversty

More information

An interactive system for structure-based ASCII art creation

An interactive system for structure-based ASCII art creation An nteractve system for structure-based ASCII art creaton Katsunor Myake Henry Johan Tomoyuk Nshta The Unversty of Tokyo Nanyang Technologcal Unversty Abstract Non-Photorealstc Renderng (NPR), whose am

More information

Study on Model of Risks Assessment of Standard Operation in Rural Power Network

Study on Model of Risks Assessment of Standard Operation in Rural Power Network Study on Model of Rsks Assessment of Standard Operaton n Rural Power Network Qngj L 1, Tao Yang 2 1 Qngj L, College of Informaton and Electrcal Engneerng, Shenyang Agrculture Unversty, Shenyang 110866,

More information

Regression Models for a Binary Response Using EXCEL and JMP

Regression Models for a Binary Response Using EXCEL and JMP SEMATECH 997 Statstcal Methods Symposum Austn Regresson Models for a Bnary Response Usng EXCEL and JMP Davd C. Trndade, Ph.D. STAT-TECH Consultng and Tranng n Appled Statstcs San Jose, CA Topcs Practcal

More information

Bayesian Network Based Causal Relationship Identification and Funding Success Prediction in P2P Lending

Bayesian Network Based Causal Relationship Identification and Funding Success Prediction in P2P Lending Proceedngs of 2012 4th Internatonal Conference on Machne Learnng and Computng IPCSIT vol. 25 (2012) (2012) IACSIT Press, Sngapore Bayesan Network Based Causal Relatonshp Identfcaton and Fundng Success

More information

A Probabilistic Theory of Coherence

A Probabilistic Theory of Coherence A Probablstc Theory of Coherence BRANDEN FITELSON. The Coherence Measure C Let E be a set of n propostons E,..., E n. We seek a probablstc measure C(E) of the degree of coherence of E. Intutvely, we want

More information

Imperial College London

Imperial College London F. Fang 1, C.C. Pan 1, I.M. Navon 2, M.D. Pggott 1, G.J. Gorman 1, P.A. Allson 1 and A.J.H. Goddard 1 1 Appled Modellng and Computaton Group Department of Earth Scence and Engneerng Imperal College London,

More information

Tracking with Non-Linear Dynamic Models

Tracking with Non-Linear Dynamic Models CHAPTER 2 Trackng wth Non-Lnear Dynamc Models In a lnear dynamc model wth lnear measurements, there s always only one peak n the posteror; very small non-lneartes n dynamc models can lead to a substantal

More information

Multi-View Regression via Canonical Correlation Analysis

Multi-View Regression via Canonical Correlation Analysis Mult-Vew Regresson va Canoncal Correlaton Analyss Sham M. Kakade 1 and Dean P. Foster 2 1 Toyota Technologcal Insttute at Chcago Chcago, IL 60637 2 Unversty of Pennsylvana Phladelpha, PA 19104 Abstract.

More information