Online Algorithms: Learning & Optimization with No Regret.
|
|
|
- Berenice Page
- 10 years ago
- Views:
Transcription
1 Online Algorithms: Learning & Optimization with No Regret. Daniel Golovin 1
2 The Setup Optimization: Model the problem (objective, constraints) Pick best decision from a feasible set. Learning: Model the problem (objective, hypothesis class) Pick best hypothesis from a feasible set. 2
3 Online Learning/Optimization Choose an action x t 2 X Get f t (x t ) and feedback f t : X! [0; 1] Same feasible set X in each round t Different Reward Models: Stochastic, Arbitrary but Oblivious, Adaptive and Arbitrary 3
4 Concrete Example: Commuting Pick a path x t from home to school. Pay cost f t (x t ) := P e2x t c t (e) Then see all edge costs for that round. Dealing with Limited Feedback: later in the course. 4
5 Other Applications Sequential decision problems Streaming algorithms for optimization/learning with large data sets Combining weak learners into strong ones ( boosting ) Fast approximate solvers for certain classes of convex programs Playing repeated games 5
6 Binary prediction with a perfect expert n hypotheses ( experts ) h 1 ; h 2 ; : : : ; h n Guaranteed that some hypothesis is perfect. Each round, get a data point pt and classifications h i (p t ) 2 f0; 1g Output binary prediction xt, observe correct label Minimize # mistakes Any Suggestions? 6
7 A Weighted Majority Algorithm Each expert votes for it's classification. Only votes from experts who have never been wrong are counted. Go with the majority # mistakes M log 2 (n) Weights w it = I(h i correct on rst t rounds ). W t = P i w it. W 0 = n, W T 1 Mistake on round t implies W t+1 W t =2 So 1 W T W 0 =2 M = n=2 M 7
8 Weighted Majority [Littlestone & Warmuth '89] What if there's no perfect expert? Each expert i has a weight w(i), votes for it's classification in {-1, 1}. Go with the weighted majority, predict sign( i w i x i ). Halve weights of wrong experts. Let m = # mistakes of best expert. How many mistakes M do we make? Weights w it = (1=2) ( # mistakes by i on rst t rounds). Let W t := P i w it. Note W 0 = n, W T (1=2) m Mistake on round t implies W t W t So (1=2) m W T W 0 (3=4) M = n (3=4) M Thus (4=3) M n 2 m and M 2:41(m + log 2 (n)). 8
9 Can we do better? M 2:41(m + log 2 (n)) Experts\Time e e 1 1 No deterministic algorithm can get M < 2m. What if there are more than 2 choices? 9
10 Notation: Define loss or cost functions ct and define the regret of x 1, x 2,..., x T as R T = TX c t (x t ) t=1 Regret Maybe all one can do is hope to end up with the right regrets. Arthur Miller TX t=1 c t (x ) where x = argmin x2x P T t=1 c t(x) A sequence has \no-regret" if R T = o(t ). Questions: How can we improve Weighted Majority? What is the lowest regret we can hope for? 10
11 The Hedge/WMR Algorithm* [Freund & Schapire '97] Hedge(²) p t (i) := w it = X w jt Initialize w i0 = 1 for all i. j In each round t: Choose expert e t from categorical distribution p t Select x t = x(e t ; t), the advice/prediction of e t. For each i, set w i;t+1 = w it (1 ²) c t(x(e i ;t)) How does this compare to WM? * Pedantic note: Hedge is often called Randomized Weighted Majority, and abbreviated WMR, though WMR was published in the context of binary classification, unlike Hedge. 11
12 The Hedge/WMR Algorithm Hedge(²) p t (i) := w it = X w jt Initialize w i0 = 1 for all i. j In each round t: Choose expert e t from categorical distribution p t Select x t = x(e t ; t), the advice/prediction of e t. For each i, set w i;t+1 = w it (1 ²) c t(x(e i ;t)) Randomization Influence shrinks exponentially with cumulative loss. Intuitively: Either we do well on a round, or total weight drops, and total weight can't drop too much unless every expert is lousy. 12
13 t=1 Hedge Performance Theorem: Let x 1 ; x 2 ; : : : be the choices of Hedge(²). Then " T # X µ 1 E c t (x t ) O PT T + ln(n) 1 ² ² P T where O PT T := min i t=1 c t(x(e i ; t)). ³ pln(n)=opt If ² =, the regret is ( p O PT ln(n)) 13
14 Hedge Analysis Intuitively: Either we do well on a round, or total weight drops, and total weight can't drop too much unless every expert is lousy. Let W t := P i w it. Then W 0 = n and W T +1 (1 ²) OPT. W t+1 = X i w it (1 ²) c t(x it ) (1) = X W t p t (i)(1 ²) c t(x it ) [def of pt(i)] (2) i X W t p t (i) (1 ² c t (x it )) [Bernoulli's ineq] (3) i If x > 1; r 2 (0; 1) then (1 + x) r 1 + rx = W t (1 ² E [c t (x t )]) (4) W t exp ( ² E [c t (x t )]) [1 x e x ] (5) 14
15 E t=1 Hedge Analysis W T +1 =W 0 exp W 0 =W T +1 exp ln(n) ² Ã Ã ² ² + OPT 1 ² TX t=1 TX t=1 E [c t (x t )] E [c t (x t )] Recall W 0 = n and W T +1 (1 ²) OPT. " X T # c t (x t ) 1 µ ² ln W0 ln(n) O PT ln(1 ²) ² ² W T +1!! 15
16 If ² = Lower Bound ³ pln(n)=opt, the regret is ( p O PT ln(n)) Can we do better? Let c t (x)» Bernoulli(1/2) for all x and t. Let Z i := P T t=1 c t(x(e i ; t)). Then Z i» Bin(T; 1=2) is roughly normally distributed, with ¾ = 2p 1 T. P [Z i ¹ k¾] = exp (k 2 ) We get about ¹ = T=2, best choice is likely to get ¹ ( p T ln(n)) = ¹ ( p O PT ln(n)). 16
17 What have we shown? Simple algorithm that learns to do nearly as well as best fixed choice. Hedge can exploit any pattern that the best choice does. Works for Adaptive Adversaries. Suitable for playing repeated games. Related ideas appearing in Algorithmic Game Theory literature. 17
18 Related Questions Optimize and get no-regret against richer classes of strategies/experts: All distributions over experts All sequences of experts that have K transitions [Auer et al '02] Various classes of functions of input features [Blum & Mansour '05] E.g., consider time of day when choosing driving route. Arbitrary convex set of experts, metric space of experts, etc, with linear, convex, or Lipschitz costs. [Zinkevich '03, Kleinberg et al '08] All policies of a K-state initially unknown Markov Decision Process that models the world. [Auer et al '08] Arbitrary sets of strategies in R n with linear costs that we can optimize offline. [Hannan'57, Kalai & Vempala '02] 18
19 Related Questions Other notions of regret (see e.g., [Blum & Mansour '05]) Time selection functions: get low regret on mondays, rainy days, etc. Sleeping experts: if rule if(p) then predict Q is right 90% of the time it applies, be right 89% of the time P applies. Internal regret & swap regret: If you played x1,..., xt then have no regret against g(x 1 ),..., g(x T ) for every g:x X 19
20 Sleeping Experts if rule if(p) then predict Q is right 90% of the time it applies, be right 89% of the time P applies. Get this for every rule simultaneously. Idea: Generate lots of hypotheses that specialize on certain inputs, some good, some lousy, and combine them into a great classifier. Many applications: [Freund et al '97, Blum '97, Blum & Mansour '05] Document classification, Spam filtering, Adaptive Uis,... if ( physics in D) then classify D as science. Predicates can overlap. 20
21 Sleeping Experts Predicates can overlap E.g., predict college major given the classes C you're enrolled in? if(ml-101, CS-201 in C) then CS if(ml-101, Stats-201 in C) then Stats What do we predict for students enrolled in ML-101, CS-201, and Stats-201? 21
22 Sleeping Experts [Algorithm from Blum & Mansour '05] SleepingExperts(, E, F) Input: 2 (0; 1), experts E, time selection functions F Initialize we;f 0 = 1 for all e 2 E; f 2 F. In each round t: Let we t = P f f(t)wt e;f. Let W t = P e wt e. Let p t e = we=w t t. Choose expert e t from categorical distribution p t Select x t = x(e t ; t), the advice/prediction of e t. For each e 2 E; f 2 F w t+1 e;f = wt e;f f(t)(c t(e) E[c t (e t )]) 22
23 Sleeping Experts [Algorithm from Blum & Mansour '05] w t+1 e;f = wt e;f f(t)(c t(e) E[c t (e t )]) Ensures total sum of weights can never increase. X e;f w t e;f nm for all t w T e;f = Y t 0 f(t)(c t(e) E[c t (e t )]) = P t 0 [f(t)(c t(e) E[c t (e t )])] nm 23
24 Sleeping Experts Performance Let n = jej; m = jfj. Fix T 2 N. Let C(e; f) := P T t=1 f(t) c t(e) Let C alg (f) := P T t=1 f(t) c t(e t ) Then for all e 2 E; f 2 F E [C alg (f)] 1 If = 1 ² is close to 1, ³ C(e; f) + log 1= (nm) E [C alg (f)] = (1 + (²)) C(e; f) + Optimizing yields a regret bound of O( p C(e; f) log(nm) + log(nm)). µ log2 (nm) ² 24
Notes from Week 1: Algorithms for sequential prediction
CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking
17.3.1 Follow the Perturbed Leader
CS787: Advanced Algorithms Topic: Online Learning Presenters: David He, Chris Hopman 17.3.1 Follow the Perturbed Leader 17.3.1.1 Prediction Problem Recall the prediction problem that we discussed in class.
Foundations of Machine Learning On-Line Learning. Mehryar Mohri Courant Institute and Google Research [email protected]
Foundations of Machine Learning On-Line Learning Mehryar Mohri Courant Institute and Google Research [email protected] Motivation PAC learning: distribution fixed over time (training and test). IID assumption.
The Multiplicative Weights Update method
Chapter 2 The Multiplicative Weights Update method The Multiplicative Weights method is a simple idea which has been repeatedly discovered in fields as diverse as Machine Learning, Optimization, and Game
4 Learning, Regret minimization, and Equilibria
4 Learning, Regret minimization, and Equilibria A. Blum and Y. Mansour Abstract Many situations involve repeatedly making decisions in an uncertain environment: for instance, deciding what route to drive
1 Introduction. 2 Prediction with Expert Advice. Online Learning 9.520 Lecture 09
1 Introduction Most of the course is concerned with the batch learning problem. In this lecture, however, we look at a different model, called online. Let us first compare and contrast the two. In batch
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Online Bandit Learning against an Adaptive Adversary: from Regret to Policy Regret
Online Bandit Learning against an Adaptive Adversary: from Regret to Policy Regret Raman Arora Toyota Technological Institute at Chicago, Chicago, IL 60637, USA Ofer Dekel Microsoft Research, 1 Microsoft
Online Convex Programming and Generalized Infinitesimal Gradient Ascent
Online Convex Programming and Generalized Infinitesimal Gradient Ascent Martin Zinkevich February 003 CMU-CS-03-110 School of Computer Science Carnegie Mellon University Pittsburgh, PA 1513 Abstract Convex
Simple and efficient online algorithms for real world applications
Simple and efficient online algorithms for real world applications Università degli Studi di Milano Milano, Italy Talk @ Centro de Visión por Computador Something about me PhD in Robotics at LIRA-Lab,
Boosting. [email protected]
. Machine Learning Boosting Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg [email protected]
LARGE CLASSES OF EXPERTS
LARGE CLASSES OF EXPERTS Csaba Szepesvári University of Alberta CMPUT 654 E-mail: [email protected] UofA, October 31, 2006 OUTLINE 1 TRACKING THE BEST EXPERT 2 FIXED SHARE FORECASTER 3 VARIABLE-SHARE
Active Learning with Boosting for Spam Detection
Active Learning with Boosting for Spam Detection Nikhila Arkalgud Last update: March 22, 2008 Active Learning with Boosting for Spam Detection Last update: March 22, 2008 1 / 38 Outline 1 Spam Filters
Online Learning. 1 Online Learning and Mistake Bounds for Finite Hypothesis Classes
Advanced Course in Machine Learning Spring 2011 Online Learning Lecturer: Shai Shalev-Shwartz Scribe: Shai Shalev-Shwartz In this lecture we describe a different model of learning which is called online
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
Week 1: Introduction to Online Learning
Week 1: Introduction to Online Learning 1 Introduction This is written based on Prediction, Learning, and Games (ISBN: 2184189 / -21-8418-9 Cesa-Bianchi, Nicolo; Lugosi, Gabor 1.1 A Gentle Start Consider
Strongly Adaptive Online Learning
Amit Daniely Alon Gonen Shai Shalev-Shwartz The Hebrew University [email protected] [email protected] [email protected] Abstract Strongly adaptive algorithms are algorithms whose performance
Online Convex Optimization
E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting
Achieving All with No Parameters: AdaNormalHedge
JMLR: Workshop and Conference Proceedings vol 40:1 19, 2015 Achieving All with No Parameters: AdaNormalHedge Haipeng Luo Department of Computer Science, Princeton University Robert E. Schapire Microsoft
Ensemble Methods. Knowledge Discovery and Data Mining 2 (VU) (707.004) Roman Kern. KTI, TU Graz 2015-03-05
Ensemble Methods Knowledge Discovery and Data Mining 2 (VU) (707004) Roman Kern KTI, TU Graz 2015-03-05 Roman Kern (KTI, TU Graz) Ensemble Methods 2015-03-05 1 / 38 Outline 1 Introduction 2 Classification
Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling
Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one
Introduction to Online Learning Theory
Introduction to Online Learning Theory Wojciech Kot lowski Institute of Computing Science, Poznań University of Technology IDSS, 04.06.2013 1 / 53 Outline 1 Example: Online (Stochastic) Gradient Descent
Robbing the bandit: Less regret in online geometric optimization against an adaptive adversary.
Robbing the bandit: Less regret in online geometric optimization against an adaptive adversary. Varsha Dani Thomas P. Hayes November 12, 2005 Abstract We consider online bandit geometric optimization,
The Advantages and Disadvantages of Online Linear Optimization
LINEAR PROGRAMMING WITH ONLINE LEARNING TATSIANA LEVINA, YURI LEVIN, JEFF MCGILL, AND MIKHAIL NEDIAK SCHOOL OF BUSINESS, QUEEN S UNIVERSITY, 143 UNION ST., KINGSTON, ON, K7L 3N6, CANADA E-MAIL:{TLEVIN,YLEVIN,JMCGILL,MNEDIAK}@BUSINESS.QUEENSU.CA
CS570 Data Mining Classification: Ensemble Methods
CS570 Data Mining Classification: Ensemble Methods Cengiz Günay Dept. Math & CS, Emory University Fall 2013 Some slides courtesy of Han-Kamber-Pei, Tan et al., and Li Xiong Günay (Emory) Classification:
Monotone multi-armed bandit allocations
JMLR: Workshop and Conference Proceedings 19 (2011) 829 833 24th Annual Conference on Learning Theory Monotone multi-armed bandit allocations Aleksandrs Slivkins Microsoft Research Silicon Valley, Mountain
Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski [email protected]
Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trajkovski [email protected] Ensembles 2 Learning Ensembles Learn multiple alternative definitions of a concept using different training
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three
Follow the Leader with Dropout Perturbations
JMLR: Worshop and Conference Proceedings vol 35: 26, 204 Follow the Leader with Dropout Perturbations Tim van Erven Département de Mathématiques, Université Paris-Sud, France Wojciech Kotłowsi Institute
1 Portfolio Selection
COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture # Scribe: Nadia Heninger April 8, 008 Portfolio Selection Last time we discussed our model of the stock market N stocks start on day with
Computational Game Theory and Clustering
Computational Game Theory and Clustering Martin Hoefer [email protected] 1 Computational Game Theory? 2 Complexity and Computation of Equilibrium 3 Bounding Inefficiencies 4 Conclusion Computational
Model Combination. 24 Novembre 2009
Model Combination 24 Novembre 2009 Datamining 1 2009-2010 Plan 1 Principles of model combination 2 Resampling methods Bagging Random Forests Boosting 3 Hybrid methods Stacking Generic algorithm for mulistrategy
The Online Set Cover Problem
The Online Set Cover Problem Noga Alon Baruch Awerbuch Yossi Azar Niv Buchbinder Joseph Seffi Naor ABSTRACT Let X = {, 2,..., n} be a ground set of n elements, and let S be a family of subsets of X, S
Data Mining Practical Machine Learning Tools and Techniques
Ensemble learning Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 8 of Data Mining by I. H. Witten, E. Frank and M. A. Hall Combining multiple models Bagging The basic idea
4.1 Introduction - Online Learning Model
Computational Learning Foundations Fall semester, 2010 Lecture 4: November 7, 2010 Lecturer: Yishay Mansour Scribes: Elad Liebman, Yuval Rochman & Allon Wagner 1 4.1 Introduction - Online Learning Model
Online Learning: Theory, Algorithms, and Applications. Thesis submitted for the degree of Doctor of Philosophy by Shai Shalev-Shwartz
Online Learning: Theory, Algorithms, and Applications Thesis submitted for the degree of Doctor of Philosophy by Shai Shalev-Shwartz Submitted to the Senate of the Hebrew University July 2007 This work
2.1 Complexity Classes
15-859(M): Randomized Algorithms Lecturer: Shuchi Chawla Topic: Complexity classes, Identity checking Date: September 15, 2004 Scribe: Andrew Gilpin 2.1 Complexity Classes In this lecture we will look
CSC2420 Fall 2012: Algorithm Design, Analysis and Theory
CSC2420 Fall 2012: Algorithm Design, Analysis and Theory Allan Borodin November 15, 2012; Lecture 10 1 / 27 Randomized online bipartite matching and the adwords problem. We briefly return to online algorithms
9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1
9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1 Seffi Naor Computer Science Dept. Technion Haifa, Israel Introduction
GI01/M055 Supervised Learning Proximal Methods
GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators
Interactive Machine Learning. Maria-Florina Balcan
Interactive Machine Learning Maria-Florina Balcan Machine Learning Image Classification Document Categorization Speech Recognition Protein Classification Branch Prediction Fraud Detection Spam Detection
Lecture 2: Universality
CS 710: Complexity Theory 1/21/2010 Lecture 2: Universality Instructor: Dieter van Melkebeek Scribe: Tyson Williams In this lecture, we introduce the notion of a universal machine, develop efficient universal
Table 1: Summary of the settings and parameters employed by the additive PA algorithm for classification, regression, and uniclass.
Online Passive-Aggressive Algorithms Koby Crammer Ofer Dekel Shai Shalev-Shwartz Yoram Singer School of Computer Science & Engineering The Hebrew University, Jerusalem 91904, Israel {kobics,oferd,shais,singer}@cs.huji.ac.il
Efficient algorithms for online decision problems
Journal of Computer and System Sciences 71 (2005) 291 307 www.elsevier.com/locate/jcss Efficient algorithms for online decision problems Adam Kalai a,, Santosh Vempala b a Department of Computer Science,
Supervised Learning (Big Data Analytics)
Supervised Learning (Big Data Analytics) Vibhav Gogate Department of Computer Science The University of Texas at Dallas Practical advice Goal of Big Data Analytics Uncover patterns in Data. Can be used
Online Semi-Supervised Learning
Online Semi-Supervised Learning Andrew B. Goldberg, Ming Li, Xiaojin Zhu [email protected] Computer Sciences University of Wisconsin Madison Xiaojin Zhu (Univ. Wisconsin-Madison) Online Semi-Supervised
FilterBoost: Regression and Classification on Large Datasets
FilterBoost: Regression and Classification on Large Datasets Joseph K. Bradley Machine Learning Department Carnegie Mellon University Pittsburgh, PA 523 [email protected] Robert E. Schapire Department
Online Selection of Mediated and Domain-Specific Predictions for Improved Recommender Systems
Online Selection of Mediated and Domain-Specific Predictions for Improved Recommender Systems Stephanie Rosenthal, Manuela Veloso, Anind Dey School of Computer Science Carnegie Mellon University {srosenth,veloso,anind}@cs.cmu.edu
Load Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: [email protected] Abstract Load
Making Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University [email protected] [email protected] I. Introduction III. Model The goal of our research
Algorithms for sustainable data centers
Algorithms for sustainable data centers Adam Wierman (Caltech) Minghong Lin (Caltech) Zhenhua Liu (Caltech) Lachlan Andrew (Swinburne) and many others IT is an energy hog The electricity use of data centers
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan
Offline sorting buffers on Line
Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: [email protected] 2 IBM India Research Lab, New Delhi. email: [email protected]
1 What is Machine Learning?
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #1 Scribe: Rob Schapire February 4, 2008 1 What is Machine Learning? Machine learning studies computer algorithms for learning to do
U.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, 2009. Notes on Algebra
U.C. Berkeley CS276: Cryptography Handout 0.1 Luca Trevisan January, 2009 Notes on Algebra These notes contain as little theory as possible, and most results are stated without proof. Any introductory
Source. The Boosting Approach. Example: Spam Filtering. The Boosting Approach to Machine Learning
Source The Boosting Approach to Machine Learning Notes adapted from Rob Schapire www.cs.princeton.edu/~schapire CS 536: Machine Learning Littman (Wu, TA) Example: Spam Filtering problem: filter out spam
Online Learning and Competitive Analysis: a Unified Approach
Online Learning and Competitive Analysis: a Unified Approach Shahar Chen Online Learning and Competitive Analysis: a Unified Approach Research Thesis Submitted in partial fulfillment of the requirements
Discussion on the paper Hypotheses testing by convex optimization by A. Goldenschluger, A. Juditsky and A. Nemirovski.
Discussion on the paper Hypotheses testing by convex optimization by A. Goldenschluger, A. Juditsky and A. Nemirovski. Fabienne Comte, Celine Duval, Valentine Genon-Catalot To cite this version: Fabienne
Dynamics and Equilibria
Dynamics and Equilibria Sergiu Hart Presidential Address, GAMES 2008 (July 2008) Revised and Expanded (November 2009) Revised (2010, 2011, 2012, 2013) SERGIU HART c 2008 p. 1 DYNAMICS AND EQUILIBRIA Sergiu
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
CSE 473: Artificial Intelligence Autumn 2010
CSE 473: Artificial Intelligence Autumn 2010 Machine Learning: Naive Bayes and Perceptron Luke Zettlemoyer Many slides over the course adapted from Dan Klein. 1 Outline Learning: Naive Bayes and Perceptron
The p-norm generalization of the LMS algorithm for adaptive filtering
The p-norm generalization of the LMS algorithm for adaptive filtering Jyrki Kivinen University of Helsinki Manfred Warmuth University of California, Santa Cruz Babak Hassibi California Institute of Technology
CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
Introduction to Online Optimization
Princeton University - Department of Operations Research and Financial Engineering Introduction to Online Optimization Sébastien Bubeck December 14, 2011 1 Contents Chapter 1. Introduction 5 1.1. Statistical
8.1 Min Degree Spanning Tree
CS880: Approximations Algorithms Scribe: Siddharth Barman Lecturer: Shuchi Chawla Topic: Min Degree Spanning Tree Date: 02/15/07 In this lecture we give a local search based algorithm for the Min Degree
Training Methods for Adaptive Boosting of Neural Networks for Character Recognition
Submission to NIPS*97, Category: Algorithms & Architectures, Preferred: Oral Training Methods for Adaptive Boosting of Neural Networks for Character Recognition Holger Schwenk Dept. IRO Université de Montréal
Distributed Machine Learning and Big Data
Distributed Machine Learning and Big Data Sourangshu Bhattacharya Dept. of Computer Science and Engineering, IIT Kharagpur. http://cse.iitkgp.ac.in/~sourangshu/ August 21, 2015 Sourangshu Bhattacharya
Competitive Analysis of On line Randomized Call Control in Cellular Networks
Competitive Analysis of On line Randomized Call Control in Cellular Networks Ioannis Caragiannis Christos Kaklamanis Evi Papaioannou Abstract In this paper we address an important communication issue arising
Approximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
Reinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Martin Lauer AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg [email protected]
Nonparametric tests these test hypotheses that are not statements about population parameters (e.g.,
CHAPTER 13 Nonparametric and Distribution-Free Statistics Nonparametric tests these test hypotheses that are not statements about population parameters (e.g., 2 tests for goodness of fit and independence).
How to Use Expert Advice
NICOLÒ CESA-BIANCHI Università di Milano, Milan, Italy YOAV FREUND AT&T Labs, Florham Park, New Jersey DAVID HAUSSLER AND DAVID P. HELMBOLD University of California, Santa Cruz, Santa Cruz, California
Big Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
Lecture 6 Online and streaming algorithms for clustering
CSE 291: Unsupervised learning Spring 2008 Lecture 6 Online and streaming algorithms for clustering 6.1 On-line k-clustering To the extent that clustering takes place in the brain, it happens in an on-line
Cell Phone based Activity Detection using Markov Logic Network
Cell Phone based Activity Detection using Markov Logic Network Somdeb Sarkhel [email protected] 1 Introduction Mobile devices are becoming increasingly sophisticated and the latest generation of smart
1 Maximum likelihood estimation
COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N
The Goldberg Rao Algorithm for the Maximum Flow Problem
The Goldberg Rao Algorithm for the Maximum Flow Problem COS 528 class notes October 18, 2006 Scribe: Dávid Papp Main idea: use of the blocking flow paradigm to achieve essentially O(min{m 2/3, n 1/2 }
Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010
Simulation Methods Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Simulation Methods 15.450, Fall 2010 1 / 35 Outline 1 Generating Random Numbers 2 Variance Reduction 3 Quasi-Monte
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Reinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Joschka Bödecker AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg [email protected]
Talk announcement please consider attending!
Talk announcement please consider attending! Where: Maurer School of Law, Room 335 When: Thursday, Feb 5, 12PM 1:30PM Speaker: Rafael Pass, Associate Professor, Cornell University, Topic: Reasoning Cryptographically
Lecture 4 Online and streaming algorithms for clustering
CSE 291: Geometric algorithms Spring 2013 Lecture 4 Online and streaming algorithms for clustering 4.1 On-line k-clustering To the extent that clustering takes place in the brain, it happens in an on-line
How To Find Out How Regret Works In An Online Structured Learning Problem
Alon Cohen [email protected] University of Haifa, University of Haifa, Dept. of Computer Science, 31905 Haifa, Israel Tamir Hazan [email protected] University of Haifa, University of Haifa,
Finite-time Analysis of the Multiarmed Bandit Problem*
Machine Learning, 47, 35 56, 00 c 00 Kluwer Academic Publishers. Manufactured in The Netherlands. Finite-time Analysis of the Multiarmed Bandit Problem* PETER AUER University of Technology Graz, A-8010
Chapter 11 Boosting. Xiaogang Su Department of Statistics University of Central Florida - 1 -
Chapter 11 Boosting Xiaogang Su Department of Statistics University of Central Florida - 1 - Perturb and Combine (P&C) Methods have been devised to take advantage of the instability of trees to create
large-scale machine learning revisited Léon Bottou Microsoft Research (NYC)
large-scale machine learning revisited Léon Bottou Microsoft Research (NYC) 1 three frequent ideas in machine learning. independent and identically distributed data This experimental paradigm has driven
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
Part III: Machine Learning. CS 188: Artificial Intelligence. Machine Learning This Set of Slides. Parameter Estimation. Estimation: Smoothing
CS 188: Artificial Intelligence Lecture 20: Dynamic Bayes Nets, Naïve Bayes Pieter Abbeel UC Berkeley Slides adapted from Dan Klein. Part III: Machine Learning Up until now: how to reason in a model and
