Simple and efficient online algorithms for real world applications


 Scot Simon
 2 years ago
 Views:
Transcription
1 Simple and efficient online algorithms for real world applications Università degli Studi di Milano Milano, Italy Centro de Visión por Computador
2 Something about me PhD in Robotics at LIRALab, University of Genova PostDoc in Machine Learning
3 Something about me PhD in Robotics at LIRALab, University of Genova PostDoc in Machine Learning I like theoretical motivated algorithms
4 Something about me PhD in Robotics at LIRALab, University of Genova PostDoc in Machine Learning I like theoretical motivated algorithms But they must work well too! ;)
5 Outline 1 2 OM2 Projectron BBQ
6 Outline 1 2 OM2 Projectron BBQ
7 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training
8 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training Shifting distributions: the Learner must track the best classifier
9 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training Shifting distributions: the Learner must track the best classifier Learning with limited feedback: selecting publicity banners
10 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training Shifting distributions: the Learner must track the best classifier Learning with limited feedback: selecting publicity banners Active learning: asking for specific samples to label
11 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training Shifting distributions: the Learner must track the best classifier Learning with limited feedback: selecting publicity banners Active learning: asking for specific samples to label Interactive learning: the agent and the human are learning at the same time
12 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training Shifting distributions: the Learner must track the best classifier Learning with limited feedback: selecting publicity banners Active learning: asking for specific samples to label Interactive learning: the agent and the human are learning at the same time These problems cannot be solved with standard batch learning!
13 What do they have in common? Spam filtering: minimize the number of wrongly classified mails, while training Shifting distributions: the Learner must track the best classifier Learning with limited feedback: selecting publicity banners Active learning: asking for specific samples to label Interactive learning: the agent and the human are learning at the same time These problems cannot be solved with standard batch learning! But they can in the online learning framework!
14 Learning in an artificial agent We have the Learner and the Teacher. The Learner observes examples in a sequence of rounds, and constructs the classification function incrementally.
15 Learning in an artificial agent Given an input, the Learner predicts its label.
16 Learning in an artificial agent Then the Teacher reveals the true label.
17 Learning in an artificial agent The Learner compares its prediction with the true label and update its knowledge. The aim of the learner is to minimize the number of mistakes.
18 Machine learning point of view on online learning The Teacher is a black box It can chose the examples arbitrarily  in the worst case the choice can be adversarial! There is no IID assumption on the data! Useful to model interaction between the user and the algorithm or nonstationary data Performance is measured while training : no separate testing set Update after each sample: efficiency of the update is important See [CesaBianchi & Lugosi, 2006] for a full introduction to the theory of online learning. universitylogo
19  Formal setting Learning goes on in a sequence of T rounds Instances: x t X Images, sounds, etc. Labels: y t Y Labels, numbers, structured output, etc. Prediction rule, f t (x) = ŷ for binary classification ŷ = sign( w, x ) Loss, l(ŷ, y) R +
20 Regret bounds The learner must minimize its loss on the T observed samples T l(ŷ t, y t ) t=1
21 Regret bounds The learner must minimize its loss on the T observed samples, compared to the loss of the best fixed predictor T l(ŷ t, y t ) min f t=1 T l (f (x t ), y t ) + R T t=1
22 Regret bounds The learner must minimize its loss on the T observed samples, compared to the loss of the best fixed predictor T l(ŷ t, y t ) min f t=1 T l (f (x t ), y t ) + R T t=1 We want the regret, R T, to be small: a small regret indicates that the performance of the learner is not too far from the one of the best fixed classifier
23 A loss for everyone Mistake, l 01 (w, x, y) := 1(y sign( w, x )) Hinge, l hinge (w, x, y) := max (0, 1 y w, x ) Logistic regression, l logreg (w, x, y) := log (1 + exp ( y w, x )) exponential loss, etc.
24 Let s start from the Perceptron for t = 1, 2,..., T do Receive new instance x t Predict ŷ t = sign( w, x t ) Receive label y t if y t ŷ t then w = w + y t x t end if end for
25 The mistake bound of the Perceptron Let (x 1, y 1 ),, (x T, y T ) be any sequence of instancelabel pairs, y t { 1, +1}, and x t R. The number of mistakes of the Perceptron is bounded by where L = T i l hinge (u, x i, y i ) min L + u 2 R 2 + u R L u } {{ } R T
26 The mistake bound of the Perceptron Let (x 1, y 1 ),, (x T, y T ) be any sequence of instancelabel pairs, y t { 1, +1}, and x t R. The number of mistakes of the Perceptron is bounded by where L = T i l hinge (u, x i, y i ) min L + u 2 R 2 + u R L u } {{ } R T If the problem is linearly separable the maximum number of mistakes is R 2 u 2, regardless of the ordering of the samples!
27 The mistake bound of the Perceptron Let (x 1, y 1 ),, (x T, y T ) be any sequence of instancelabel pairs, y t { 1, +1}, and x t R. The number of mistakes of the Perceptron is bounded by where L = T i l hinge (u, x i, y i ) min L + u 2 R 2 + u R L u } {{ } R T If the problem is linearly separable the maximum number of mistakes is R 2 u 2, regardless of the ordering of the samples! Video universitylogo
28 Pros and Cons of the Perceptron Pros :) Very efficient! It can be easily trained with huge datasets Theoretical guarantee on the maximum number of mistakes
29 Pros and Cons of the Perceptron Pros :) Cons :( Very efficient! It can be easily trained with huge datasets Theoretical guarantee on the maximum number of mistakes Linear hyperplane only Binary classification only
30 Pros and Cons of the Perceptron Pros :) Cons :( Very efficient! It can be easily trained with huge datasets Theoretical guarantee on the maximum number of mistakes Linear hyperplane only Binary classification only Let s generalize the Perceptron!
31 Nonlinear classifiers using Kernels! Suppose to transform the inputs through a nonlinear transform φ(x), to the feature space A linear classifier in the feature space will result in a nonlinear classifier in the input space We can do even better: the algorithms just need to access to φ(x 1 ), φ(x 2 ), so if we have such a function, K (x 1, x 2 ), we do not need φ()!
32 Kernel Perceptron for t = 1, 2,..., T do Receive new instance ( x t ) Predict ŷ t = sign x i S y ik (x i, x t ) Receive label y t if y t ŷ t then Add x t to the support set S end if end for
33 Follow The Regularized Leader algorithm is just a particular case of a more general algorithm: the follow the regularized leader (FTRL) algorithm In FTRL, at each time step we predict with the approximate batch solution using a linearization of the loss, instead of the true loss functions Regret bounds will come almost for free! See [Kakade et al., 2009] for FTRL and [Orabona&Crammer, 2010] for an even more general algorithm
34 The general algorithm for t = 1, 2,..., T do Receive new instance x t Predict ŷ t Receive label y t θ = θ η t l(w, x t, y t ) w = g t (θ) end for F. Orabona, and K. Crammer. New Adaptive for Online Classification. Accepted in Neural Information Processing Systems (NIPS) 2010
35 The general algorithm for t = 1, 2,..., T do Receive new instance x t Predict ŷ t Receive label y t θ = θ η t l(w, x t, y t ) w = g t (θ) end for Suppose g t (θ) = 1 2 θ 2 g t (θ) = θ Let s use the hinge loss l(w, x t, y t ) = y t x t η t = 1 on mistakes, 0 otherwise We recovered the Perceptron algorithm! F. Orabona, and K. Crammer. New Adaptive for Online Classification. Accepted in Neural Information Processing Systems (NIPS) 2010
36 The Recipe Create the online algorithm that best suits your needs! Define a task this will define a (convex) loss function. Define which charateristic you would like your solution to have this will define a (strongly convex) regularizer. Compute the gradient of the loss. Compute the gradient of the fenchel dual of the regularizer. Just code it!
37 Some examples Multiclassmultilabel classification, M classes, M different classifiers w i. l( w, x, Y) = max(1 + max y / Y w y, x min y Y w y, x, 0) Do you prefer sparse classifiers? Use the regularizer w 2 p with 1 < p 2. K different kernels? Use the regularizer [ w 1 2,..., w K 2 ] 2 with 1 < p 2. p
38 Batch Solutions with Online What most of the persons do when they want a batch solution: they stop the online algorithm and use the last solution found. This is wrong: the last solution can be arbitrarly bad! If the data are IID you can simply use the averaged solution [CesaBianchi et al., 2004]. Alternative way: several epochs of training, until convergence.
39 Batch Solutions with Online What most of the persons do when they want a batch solution: they stop the online algorithm and use the last solution found. This is wrong: the last solution can be arbitrarly bad! If the data are IID you can simply use the averaged solution [CesaBianchi et al., 2004]. Alternative way: several epochs of training, until convergence. Video
40 Outline OM2 Projectron BBQ 1 2 OM2 Projectron BBQ
41 Multi Kernel Learning online? OM2 Projectron BBQ Cue 1 w 1 Cue 2 w 2... Σ Prediction Cue F w F We have F different features, e.g. color, shape, etc.
42 Multi Kernel Learning online? OM2 Projectron BBQ Cue 1 w 1 Cue 2 w 2... Σ Prediction Cue F w F We have F different features, e.g. color, shape, etc. Solution: Online MultiClass MultiKernel learning algorithm L. Jie, F. Orabona, M. Fornoni, B. Caputo, and N. CesaBianchi. OM2: An Online Multiclass Multikernel Learning Algorithm. In Proc. of the 4th IEEE for Computer Vision Workshop (in CVPR10) universitylogo
43 OM2: Pseudocode OM2 Projectron BBQ Input: q Initialize: θ 1 = 0, w 1 = 0 for t = 1, 2,..., T do Receive new instance x t Predict ŷ t = argmax y=1,...,m Receive label y t z t = φ(x t, y t ) φ(x t, ŷ t ) if l( w t, x t, y t ) > 0 then η t = min { 2 w t zt 1, 1 z t 2 2,q else η t = 0 θ t+1 = θ t + η t z ( t w j t+1 = 1 q end for w t φ(x t, y) } ) q 2 θ j t+1 2 θ j θ t+1 2,q t+1, j = 1,, F Code available! universitylogo
44 Experiments OM2 Projectron BBQ We compared OM2 to OMCL [Jie et al. ACCV09], and to PAI [Crammer et al. JMLR06] using the best feature and the sum of the kernels. We also used SILP [Sonnenburg et al. JMLR06], a stateoftheart MKL batch solver. We used the Caltech101 with 39 different kernels, as in [Gehler and Nowozin ICCV09].
45 OM2 Projectron BBQ Caltech101: online performance 100 Caltech 101 (5 splits, 30 training examples per category) Online average training error rate OM 2, p=1.01 OMCL PA I (average kernel) PA I (best kernel) number of training samples Best results with p = OM2 achieves the best performance among the online algorithms.
46 OM2 Projectron BBQ Caltech101: batch performance classification rate on test data Caltech 101 (5 splits, 30 training examples per category) OM 2, p=1.01 OMCL PA I (average kernel) PA I (best kernel) SILP number of epochs Matlab implementation of OM2 takes 45 mins, SILP more than 2 hours. The performance advantage of OM2 over SILP is due the fact that OM2 is based on a native multiclass formulation. universitylogo
47 OM2 Projectron BBQ Learning with bounded complexity Perceptronlike algorithms will never stop updating the solution if the problem is not linearly separable.
48 OM2 Projectron BBQ Learning with bounded complexity Perceptronlike algorithms will never stop updating the solution if the problem is not linearly separable. On the other hand, if we use kernels, sooner or later the memory of the computer will finish...
49 OM2 Projectron BBQ Learning with bounded complexity Perceptronlike algorithms will never stop updating the solution if the problem is not linearly separable. On the other hand, if we use kernels, sooner or later the memory of the computer will finish... Is it possible to bound the complexity of the learner?
50 OM2 Projectron BBQ Learning with bounded complexity Perceptronlike algorithms will never stop updating the solution if the problem is not linearly separable. On the other hand, if we use kernels, sooner or later the memory of the computer will finish... Is it possible to bound the complexity of the learner? Yes! Just update with a noisy version of the gradient If the new vector can be well approximated with the old ones, just update the coefficients of the old ones. F. Orabona, J. Keshet, and B. Caputo. Bounded KernelBased. Journal of Machine Learning Research, 2009 universitylogo
51 The Projectron Algorithm OM2 Projectron BBQ for t = 1, 2,..., T do Receive new instance x t Predict ŷ t = sign( w, x t ) Receive label y t if y t ŷ t then w = w + y t x t w = w + y t P(x t ) if δ t = w w η then w = w else w = w end if end if end for It is possible to calculate the projection even using Kernels. The algorithm has a mistake bound and a bounded memory growth. Code available! universitylogo
52 OM2 Projectron BBQ Average Online Error vs Budget Size Adult9, samples, 123 features, Gaussian Kernel Number of Mistakes (%) A9A Gaussian Kernel RBP Forgetron Plain Perceptron PA I Projectron Size of the support set = (36.9) Size of the Support Set universitylogo
53 OM2 Projectron BBQ Robot Navigation & Phoneme Recognition Place Recognition IDOL2 database 5 rooms CRFH features Phoneme recognition Subset of the TIMIT corpus 55 phonemes MFCC + + features
54 Place recognition OM2 Projectron BBQ 1 IDOL2 exp χ 2 kernel 0.9 Accuracy on Test Set (%) Perceptron Projectron++ η 1 Projectron++ η 2 Projectron++ η Training step
55 Place recognition OM2 Projectron BBQ Size of the Support Set Perceptron Projectron++ η 1 Projectron++ η 2 Projectron++ η 3 IDOL2 exp χ 2 kernel Training step
56 Phoneme recognition OM2 Projectron BBQ Phoneme recognition Gaussian kernel Projectron++ PA I 0.31 Number of Mistakes (%) Size of the Support Set x 10 4
57 Phoneme recognition OM2 Projectron BBQ Phoneme recognition Gaussian kernel Projectron++ PA I Error on Test Set (%) Size of the Support Set x 10 4
58 OM2 Projectron BBQ Semisupervised online learning We are given N training samples and a regression problem, {x i, y i } N i=1, y i [ 1, 1] Obtaining the output for a given sample can be expensive, so we want to ask for as few as possible labels. We assume the outputs y t are realizations of random variables Y t such that E Y t = u x t for all t, where u R d is a fixed and unknown vector such that u = 1. The order of the data is still adversarial, but the outputs are generated by a stochastic source
59 OM2 Projectron BBQ Semisupervised online learning We are given N training samples and a regression problem, {x i, y i } N i=1, y i [ 1, 1] Obtaining the output for a given sample can be expensive, so we want to ask for as few as possible labels. We assume the outputs y t are realizations of random variables Y t such that E Y t = u x t for all t, where u R d is a fixed and unknown vector such that u = 1. The order of the data is still adversarial, but the outputs are generated by a stochastic source Solution: Bound on Bias Query algorithm (BBQ) N. CesaBianchi, C. Gentile and F. Orabona. Robust Bounds for Classification via Selective Sampling. In Proc. of the International Conference on Machine Learning (ICML), 2009 universitylogo
60 OM2 Projectron BBQ The Parametric BBQ Algorithm Parameters: 0 < ε, δ < 1 for t = 1, 2,..., T do Receive new instance x t Predict ŷ t = w x ( t ) r = x t I + S t 1 St 1 + x t x 1 t xt ( ) q = St 1 I + S t 1 St 1 + x t x 1 t xt ( ) s = I + S t 1 St 1 + x t x 1 t xt if [ ε r s ] + < q t(t + 1) 2 ln then 2δ Query label y t Update w with a Regularized Least Square S t = [S t, x t ] end if end for The algorithm follows the general framework Every time a query is not issued the predicted output is far from the correct one at most by ε
61 OM2 Projectron BBQ Regret Bound of Parametric BBQ Theorem If Parametric BBQ is run with input ε, δ (0, 1) then: with probability at least 1 δ, t t ε holds on all time steps t when no query is issued; the number N T of queries issued after any number T of steps is bounded as ( ( d N T = O ε 2 ln T ) ) ln(t /δ) ln. δ ε
62 Synthetic Experiment OM2 Projectron BBQ Maximum error Theoretical maximum error Measured maximum error Fraction of queried labels ε Fraction of queried labels We tested Parametric BBQ. 10,000 random examples on the unit circle in R 2. The labels were generated according to our noise model using a randomly selected hyperplane u with unit norm. universitylogo
63 Real World Experiments OM2 Projectron BBQ F measure SOLE Random Parametric BBQ Fraction of queried labels F measure SOLE 0.48 Random Parametric BBQ Fraction of queried labels Fmeasure and fraction of queried labels for different algorithms on Adult9 dataset (left)(gaussian Kernel) and RCV1 (right)(linear kernel). universitylogo
64 Summary OM2 Projectron BBQ Many real world problem cannot be solved using the standard batch framework The online learning framework offers a useful tool in these cases A general algorithm that covers many of the previous known online learning algorithms has been presented The framework allows to easily design algorithms for specific problems
65 OM2 Projectron BBQ Thanks for your attention Code: My website:
DUOL: A Double Updating Approach for Online Learning
: A Double Updating Approach for Online Learning Peilin Zhao School of Comp. Eng. Nanyang Tech. University Singapore 69798 zhao6@ntu.edu.sg Steven C.H. Hoi School of Comp. Eng. Nanyang Tech. University
More informationIntroduction to Online Learning Theory
Introduction to Online Learning Theory Wojciech Kot lowski Institute of Computing Science, Poznań University of Technology IDSS, 04.06.2013 1 / 53 Outline 1 Example: Online (Stochastic) Gradient Descent
More informationFoundations of Machine Learning OnLine Learning. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu
Foundations of Machine Learning OnLine Learning Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Motivation PAC learning: distribution fixed over time (training and test). IID assumption.
More informationList of Publications by Claudio Gentile
List of Publications by Claudio Gentile Claudio Gentile DiSTA, University of Insubria, Italy claudio.gentile@uninsubria.it November 6, 2013 Abstract Contains the list of publications by Claudio Gentile,
More informationA Potentialbased Framework for Online Multiclass Learning with Partial Feedback
A Potentialbased Framework for Online Multiclass Learning with Partial Feedback Shijun Wang Rong Jin Hamed Valizadegan Radiology and Imaging Sciences Computer Science and Engineering Computer Science
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationOnline SemiSupervised Learning
Online SemiSupervised Learning Andrew B. Goldberg, Ming Li, Xiaojin Zhu jerryzhu@cs.wisc.edu Computer Sciences University of Wisconsin Madison Xiaojin Zhu (Univ. WisconsinMadison) Online SemiSupervised
More informationDirect Loss Minimization for Structured Prediction
Direct Loss Minimization for Structured Prediction David McAllester TTIChicago mcallester@ttic.edu Tamir Hazan TTIChicago tamir@ttic.edu Joseph Keshet TTIChicago jkeshet@ttic.edu Abstract In discriminative
More informationSteven C.H. Hoi. School of Computer Engineering Nanyang Technological University Singapore
Steven C.H. Hoi School of Computer Engineering Nanyang Technological University Singapore Acknowledgments: Peilin Zhao, Jialei Wang, Hao Xia, Jing Lu, Rong Jin, Pengcheng Wu, Dayong Wang, etc. 2 Agenda
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationSemiSupervised Support Vector Machines and Application to Spam Filtering
SemiSupervised Support Vector Machines and Application to Spam Filtering Alexander Zien Empirical Inference Department, Bernhard Schölkopf Max Planck Institute for Biological Cybernetics ECML 2006 Discovery
More informationData Mining. Nonlinear Classification
Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Nonlinear Classification Classes may not be separable by a linear boundary Suppose we randomly generate a data set as follows: X has range between 0 to 15
More informationEnsemble Methods. Knowledge Discovery and Data Mining 2 (VU) (707.004) Roman Kern. KTI, TU Graz 20150305
Ensemble Methods Knowledge Discovery and Data Mining 2 (VU) (707004) Roman Kern KTI, TU Graz 20150305 Roman Kern (KTI, TU Graz) Ensemble Methods 20150305 1 / 38 Outline 1 Introduction 2 Classification
More informationSparse Online Learning via Truncated Gradient
Sparse Online Learning via Truncated Gradient John Langford Yahoo! Research jl@yahooinc.com Lihong Li Department of Computer Science Rutgers University lihong@cs.rutgers.edu Tong Zhang Department of Statistics
More informationFrom N to N+1: Multiclass Transfer Incremental Learning
From N to N+1: Multiclass Transfer Incremental Learning Ilja Kuzborskij 1,2, Francesco Orabona 3, and Barbara Caputo 1 1 Idiap Research Institute, Switzerland, 2 École Polytechnique Fédérale de Lausanne
More informationIntroduction to Machine Learning. Speaker: Harry Chao Advisor: J.J. Ding Date: 1/27/2011
Introduction to Machine Learning Speaker: Harry Chao Advisor: J.J. Ding Date: 1/27/2011 1 Outline 1. What is machine learning? 2. The basic of machine learning 3. Principles and effects of machine learning
More informationSupervised Learning (Big Data Analytics)
Supervised Learning (Big Data Analytics) Vibhav Gogate Department of Computer Science The University of Texas at Dallas Practical advice Goal of Big Data Analytics Uncover patterns in Data. Can be used
More informationOnline Feature Selection for Mining Big Data
Online Feature Selection for Mining Big Data Steven C.H. Hoi, Jialei Wang, Peilin Zhao, Rong Jin School of Computer Engineering, Nanyang Technological University, Singapore Department of Computer Science
More informationTable 1: Summary of the settings and parameters employed by the additive PA algorithm for classification, regression, and uniclass.
Online PassiveAggressive Algorithms Koby Crammer Ofer Dekel Shai ShalevShwartz Yoram Singer School of Computer Science & Engineering The Hebrew University, Jerusalem 91904, Israel {kobics,oferd,shais,singer}@cs.huji.ac.il
More informationAdaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
More informationThe Artificial Prediction Market
The Artificial Prediction Market Adrian Barbu Department of Statistics Florida State University Joint work with Nathan Lay, Siemens Corporate Research 1 Overview Main Contributions A mathematical theory
More informationWeek 1: Introduction to Online Learning
Week 1: Introduction to Online Learning 1 Introduction This is written based on Prediction, Learning, and Games (ISBN: 2184189 / 2184189 CesaBianchi, Nicolo; Lugosi, Gabor 1.1 A Gentle Start Consider
More informationMultiArmed Bandit Problems
Machine Learning Coms4771 MultiArmed Bandit Problems Lecture 20 Multiarmed Bandit Problems The Setting: K arms (or actions) Each time t, each arm i pays off a bounded realvalued reward x i (t), say
More informationAn Introduction to Gametheoretic Online Learning Tim van Erven
General Mathematics Colloquium Leiden, December 4, 2014 An Introduction to Gametheoretic Online Learning Tim van Erven Statistics Bayesian Statistics Frequentist Statistics Statistics Bayesian Statistics
More informationLinear smoother. ŷ = S y. where s ij = s ij (x) e.g. s ij = diag(l i (x)) To go the other way, you need to diagonalize S
Linear smoother ŷ = S y where s ij = s ij (x) e.g. s ij = diag(l i (x)) To go the other way, you need to diagonalize S 2 Online Learning: LMS and Perceptrons Partially adapted from slides by Ryan Gabbard
More informationBig Data Analytics CSCI 4030
High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering data streams SVM Recommen der systems Clustering Community Detection Web advertising
More informationOnline Kernel Selection: Algorithms and Evaluations
Proceedings of the TwentySixth AAAI Conference on Artificial Intelligence Online Kernel Selection: Algorithms and Evaluations Tianbao Yang 1, Mehrdad Mahdavi 1, Rong Jin 1, Jinfeng Yi 1, Steven C. H.
More informationInteractive Machine Learning. MariaFlorina Balcan
Interactive Machine Learning MariaFlorina Balcan Machine Learning Image Classification Document Categorization Speech Recognition Protein Classification Branch Prediction Fraud Detection Spam Detection
More informationSteven C.H. Hoi. Methods and Applications. School of Computer Engineering Nanyang Technological University Singapore 4 May, 2013
Methods and Applications Steven C.H. Hoi School of Computer Engineering Nanyang Technological University Singapore 4 May, 2013 http://libol.stevenhoi.org/ Acknowledgements Peilin Zhao Jialei Wang Rong
More informationData Mining Practical Machine Learning Tools and Techniques
Ensemble learning Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 8 of Data Mining by I. H. Witten, E. Frank and M. A. Hall Combining multiple models Bagging The basic idea
More informationMachine Learning in Spam Filtering
Machine Learning in Spam Filtering A Crash Course in ML Konstantin Tretyakov kt@ut.ee Institute of Computer Science, University of Tartu Overview Spam is Evil ML for Spam Filtering: General Idea, Problems.
More informationLargeScale Similarity and Distance Metric Learning
LargeScale Similarity and Distance Metric Learning Aurélien Bellet Télécom ParisTech Joint work with K. Liu, Y. Shi and F. Sha (USC), S. Clémençon and I. Colin (Télécom ParisTech) Séminaire Criteo March
More informationCOS 511: Foundations of Machine Learning. Rob Schapire Lecture #14 Scribe: Qian Xi March 30, 2006
COS 511: Foundations of Machine Learning Rob Schapire Lecture #14 Scribe: Qian Xi March 30, 006 In the previous lecture, we introduced a new learning model, the Online Learning Model. After seeing a concrete
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More informationOnline PassiveAggressive Algorithms on a Budget
Zhuang Wang Dept. of Computer and Information Sciences Temple University, USA zhuang@temple.edu Slobodan Vucetic Dept. of Computer and Information Sciences Temple University, USA vucetic@temple.edu Abstract
More informationOnline Algorithms: Learning & Optimization with No Regret.
Online Algorithms: Learning & Optimization with No Regret. Daniel Golovin 1 The Setup Optimization: Model the problem (objective, constraints) Pick best decision from a feasible set. Learning: Model the
More informationIntroduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multiclass classification.
More informationIntroduction to Machine Learning Lecture 1. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu
Introduction to Machine Learning Lecture 1 Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Introduction Logistics Prerequisites: basics concepts needed in probability and statistics
More informationDistributed Machine Learning and Big Data
Distributed Machine Learning and Big Data Sourangshu Bhattacharya Dept. of Computer Science and Engineering, IIT Kharagpur. http://cse.iitkgp.ac.in/~sourangshu/ August 21, 2015 Sourangshu Bhattacharya
More information1 Introduction. 2 Prediction with Expert Advice. Online Learning 9.520 Lecture 09
1 Introduction Most of the course is concerned with the batch learning problem. In this lecture, however, we look at a different model, called online. Let us first compare and contrast the two. In batch
More informationCS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott
More informationTrading regret rate for computational efficiency in online learning with limited feedback
Trading regret rate for computational efficiency in online learning with limited feedback Shai ShalevShwartz TTIC Hebrew University Online Learning with Limited Feedback Workshop, 2009 June 2009 Shai
More informationCSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
More informationBDUOL: Double Updating Online Learning on a Fixed Budget
BDUOL: Double Updating Online Learning on a Fixed Budget Peilin Zhao and Steven C.H. Hoi School of Computer Engineering, Nanyang Technological University, Singapore Email: {zhao0106,chhoi}@ntu.edu.sg
More informationINTRODUCTION TO NEURAL NETWORKS
INTRODUCTION TO NEURAL NETWORKS Pictures are taken from http://www.cs.cmu.edu/~tom/mlbookchapterslides.html http://research.microsoft.com/~cmbishop/prml/index.htm By Nobel Khandaker Neural Networks An
More informationThe pnorm generalization of the LMS algorithm for adaptive filtering
The pnorm generalization of the LMS algorithm for adaptive filtering Jyrki Kivinen University of Helsinki Manfred Warmuth University of California, Santa Cruz Babak Hassibi California Institute of Technology
More informationOnline Learning for Big Data Analytics
Online Learning for Big Data Analytics Irwin King and Haiqin Yang Dept. of Computer Science and Engineering The Chinese University of Hong Kong 1 Outline Introduction Big data: definition and history Online
More informationNeural Networks. Introduction to Artificial Intelligence CSE 150 May 29, 2007
Neural Networks Introduction to Artificial Intelligence CSE 150 May 29, 2007 Administration Last programming assignment has been posted! Final Exam: Tuesday, June 12, 11:302:30 Last Lecture Naïve Bayes
More informationCheng Soon Ong & Christfried Webers. Canberra February June 2016
c Cheng Soon Ong & Christfried Webers Research Group and College of Engineering and Computer Science Canberra February June (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 31 c Part I
More informationAn Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia Alex.Smola@nicta.com.au Tata Institute, Pune,
More informationOnline Classification on a Budget
Online Classification on a Budget Koby Crammer Computer Sci. & Eng. Hebrew University Jerusalem 91904, Israel kobics@cs.huji.ac.il Jaz Kandola Royal Holloway, University of London Egham, UK jaz@cs.rhul.ac.uk
More informationProgramming Exercise 3: Multiclass Classification and Neural Networks
Programming Exercise 3: Multiclass Classification and Neural Networks Machine Learning November 4, 2011 Introduction In this exercise, you will implement onevsall logistic regression and neural networks
More informationChapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
More informationLogistic Regression for Spam Filtering
Logistic Regression for Spam Filtering Nikhila Arkalgud February 14, 28 Abstract The goal of the spam filtering problem is to identify an email as a spam or not spam. One of the classic techniques used
More informationIntroduction to Machine Learning
Introduction to Machine Learning Prof. Alexander Ihler Prof. Max Welling icamp Tutorial July 22 What is machine learning? The ability of a machine to improve its performance based on previous results:
More informationClassification using intersection kernel SVMs is efficient
Classification using intersection kernel SVMs is efficient Jitendra Malik UC Berkeley Joint work with Subhransu Maji and Alex Berg Fast intersection kernel SVMs and other generalizations of linear SVMs
More informationData Mining  Evaluation of Classifiers
Data Mining  Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
More informationThese slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
Music and Machine Learning (IFT6080 Winter 08) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher
More information17.3.1 Follow the Perturbed Leader
CS787: Advanced Algorithms Topic: Online Learning Presenters: David He, Chris Hopman 17.3.1 Follow the Perturbed Leader 17.3.1.1 Prediction Problem Recall the prediction problem that we discussed in class.
More informationStochastic Optimization for Big Data Analytics: Algorithms and Libraries
Stochastic Optimization for Big Data Analytics: Algorithms and Libraries Tianbao Yang SDM 2014, Philadelphia, Pennsylvania collaborators: Rong Jin, Shenghuo Zhu NEC Laboratories America, Michigan State
More informationMachine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
More informationData Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression
Data Mining and Data Warehousing Henryk Maciejewski Data Mining Predictive modelling: regression Algorithms for Predictive Modelling Contents Regression Classification Auxiliary topics: Estimation of prediction
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
More informationBig learning: challenges and opportunities
Big learning: challenges and opportunities Francis Bach SIERRA Projectteam, INRIA  Ecole Normale Supérieure December 2013 Omnipresent digital media Scientific context Big data Multimedia, sensors, indicators,
More informationLecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
More informationOnline Learning. 1 Online Learning and Mistake Bounds for Finite Hypothesis Classes
Advanced Course in Machine Learning Spring 2011 Online Learning Lecturer: Shai ShalevShwartz Scribe: Shai ShalevShwartz In this lecture we describe a different model of learning which is called online
More informationSteven C.H. Hoi School of Information Systems Singapore Management University Email: chhoi@smu.edu.sg
Steven C.H. Hoi School of Information Systems Singapore Management University Email: chhoi@smu.edu.sg Introduction http://stevenhoi.org/ Finance Recommender Systems Cyber Security Machine Learning Visual
More informationLinear Classification. Volker Tresp Summer 2015
Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong
More informationMaking Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University atran3@stanford.edu ginzberg@stanford.edu I. Introduction III. Model The goal of our research
More informationKarthik Sridharan. 424 Gates Hall Ithaca, Email: sridharan@cs.cornell.edu http://www.cs.cornell.edu/ sridharan/ Contact Information
Karthik Sridharan Contact Information 424 Gates Hall Ithaca, NY 148537501 USA Email: sridharan@cs.cornell.edu http://www.cs.cornell.edu/ sridharan/ Research Interests Machine Learning, Statistical Learning
More informationOnline Active Learning Methods for Fast LabelEfficient Spam Filtering
Online Active Learning Methods for Fast LabelEfficient Spam Filtering D. Sculley Department of Computer Science Tufts University, Medford, MA USA dsculley@cs.tufts.edu ABSTRACT Active learning methods
More informationNeural Networks. CAP5610 Machine Learning Instructor: GuoJun Qi
Neural Networks CAP5610 Machine Learning Instructor: GuoJun Qi Recap: linear classifier Logistic regression Maximizing the posterior distribution of class Y conditional on the input vector X Support vector
More informationArtificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network?  Perceptron learners  Multilayer networks What is a Support
More informationMachine Learning. CUNY Graduate Center, Spring 2013. Professor Liang Huang. huang@cs.qc.cuny.edu
Machine Learning CUNY Graduate Center, Spring 2013 Professor Liang Huang huang@cs.qc.cuny.edu http://acl.cs.qc.edu/~lhuang/teaching/machinelearning Logistics Lectures M 9:3011:30 am Room 4419 Personnel
More informationLearning. Artificial Intelligence. Learning. Types of Learning. Inductive Learning Method. Inductive Learning. Learning.
Learning Learning is essential for unknown environments, i.e., when designer lacks omniscience Artificial Intelligence Learning Chapter 8 Learning is useful as a system construction method, i.e., expose
More informationLecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
More informationTensor Methods for Machine Learning, Computer Vision, and Computer Graphics
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics Part I: Factorizations and Statistical Modeling/Inference Amnon Shashua School of Computer Science & Eng. The Hebrew University
More informationBig Data Analytics. Lucas Rego Drumond
Big Data Analytics Lucas Rego Drumond Information Systems and Machine Learning Lab (ISMLL) Institute of Computer Science University of Hildesheim, Germany Going For Large Scale Going For Large Scale 1
More informationCS 688 Pattern Recognition Lecture 4. Linear Models for Classification
CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(
More informationSupport Vector Machine. Tutorial. (and Statistical Learning Theory)
Support Vector Machine (and Statistical Learning Theory) Tutorial Jason Weston NEC Labs America 4 Independence Way, Princeton, USA. jasonw@neclabs.com 1 Support Vector Machines: history SVMs introduced
More informationMapReduce/Bigtable for Distributed Optimization
MapReduce/Bigtable for Distributed Optimization Keith B. Hall Google Inc. kbhall@google.com Scott Gilpin Google Inc. sgilpin@google.com Gideon Mann Google Inc. gmann@google.com Abstract With large data
More informationRecognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
More informationScalable Developments for Big Data Analytics in Remote Sensing
Scalable Developments for Big Data Analytics in Remote Sensing Federated Systems and Data Division Research Group High Productivity Data Processing Dr.Ing. Morris Riedel et al. Research Group Leader,
More informationlargescale machine learning revisited Léon Bottou Microsoft Research (NYC)
largescale machine learning revisited Léon Bottou Microsoft Research (NYC) 1 three frequent ideas in machine learning. independent and identically distributed data This experimental paradigm has driven
More informationBeyond stochastic gradient descent for largescale machine learning
Beyond stochastic gradient descent for largescale machine learning Francis Bach INRIA  Ecole Normale Supérieure, Paris, France Joint work with Eric Moulines, Nicolas Le Roux and Mark Schmidt  ECMLPKDD,
More informationChapter 6. The stacking ensemble approach
82 This chapter proposes the stacking ensemble approach for combining different data mining classifiers to get better performance. Other combination techniques like voting, bagging etc are also described
More informationQuestion 2 Naïve Bayes (16 points)
Question 2 Naïve Bayes (16 points) About 2/3 of your email is spam so you downloaded an open source spam filter based on word occurrences that uses the Naive Bayes classifier. Assume you collected the
More informationPredict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, lyzhou@stanford.edu Department of Electrical Engineering, Stanford University Abstract Given two persons
More informationBilinear Prediction Using LowRank Models
Bilinear Prediction Using LowRank Models Inderjit S. Dhillon Dept of Computer Science UT Austin 26th International Conference on Algorithmic Learning Theory Banff, Canada Oct 6, 2015 Joint work with CJ.
More informationCS 2750 Machine Learning. Lecture 1. Machine Learning. CS 2750 Machine Learning.
Lecture 1 Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott
More informationAnalecta Vol. 8, No. 2 ISSN 20647964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
More informationSupport Vector Machine (SVM)
Support Vector Machine (SVM) CE725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept HardMargin SVM SoftMargin SVM Dual Problems of HardMargin
More informationADVANCED MACHINE LEARNING. Introduction
1 1 Introduction Lecturer: Prof. Aude Billard (aude.billard@epfl.ch) Teaching Assistants: Guillaume de Chambrier, Nadia Figueroa, Denys Lamotte, Nicola Sommer 2 2 Course Format Alternate between: Lectures
More informationLecture 6: Logistic Regression
Lecture 6: CS 19410, Fall 2011 Laurent El Ghaoui EECS Department UC Berkeley September 13, 2011 Outline Outline Classification task Data : X = [x 1,..., x m]: a n m matrix of data points in R n. y { 1,
More informationMachine Learning Big Data using Map Reduce
Machine Learning Big Data using Map Reduce By Michael Bowles, PhD Where Does Big Data Come From? Web data (web logs, click histories) ecommerce applications (purchase histories) Retail purchase histories
More informationHT2015: SC4 Statistical Data Mining and Machine Learning
HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric
More informationMachine learning challenges for big data
Machine learning challenges for big data Francis Bach SIERRA Projectteam, INRIA  Ecole Normale Supérieure Joint work with R. Jenatton, J. Mairal, G. Obozinski, N. Le Roux, M. Schmidt  December 2012
More informationA Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Largemargin linear
More informationDifferential privacy in health care analytics and medical research An interactive tutorial
Differential privacy in health care analytics and medical research An interactive tutorial Speaker: Moritz Hardt Theory Group, IBM Almaden February 21, 2012 Overview 1. Releasing medical data: What could
More informationMachine Learning Logistic Regression
Machine Learning Logistic Regression Jeff Howbert Introduction to Machine Learning Winter 2012 1 Logistic regression Name is somewhat misleading. Really a technique for classification, not regression.
More information