1 Introduction. 2 Prediction with Expert Advice. Online Learning Lecture 09
|
|
|
- Nelson Houston
- 9 years ago
- Views:
Transcription
1 1 Introduction Most of the course is concerned with the batch learning problem. In this lecture, however, we look at a different model, called online. Let us first compare and contrast the two. In batch learning, an algorithm takes the data (training samples) and returns a hypothesis. We assume a static nature of the world: a new example that we will encounter will be similar to the training set. More precisely, we suppose that all the training samples, as well as the test point, are independent and identically distributed. Hence, given the bunch z 1,..., z n of training samples drawn from a distribution p, the quality of the learned function f is E z p l(f, z), where l is some loss function. he questions addressed by statistical learning theory are: how many examples are needed to have small expected error with certain confidence? what is the lowest error that can be achieved under certain conditions on p? etc. If the world is not static, it might be difficult to take advantage of large amounts of data. We can no longer rely on statistical assumptions. In fact, we take an even more dramatic view of the world. We suppose that there are no correlations whatsoever between any two days. As there is no stationary distribution responsible for the data, we no longer want to minimize some expected error. All we want is to survive no matter how adversarial the world is. By surviving, we mean that we do not do too badly relative to other agents in the world. In fact, the goal is to do not much worse than the best agent (this difference is called regret). Note that this does not guarantee that we will be doing well in absolute terms, as the best agent might be quite bad. However, this is the price we have to pay for not having any coherence between today and tomorrow. he goal of online learning is, therefore, to do almost as well as the best comparator. oday we will describe the most famous setting, prediction with expert advice. We will then try to understand why the algorithm for this setting works by making some abstractions and moving to the Online Convex Optimization framework. Finally, we will unify the two via Regularization and provide some powerful tools for proving regret guarantees. 2 Prediction with Expert Advice Suppose we have access to predictions of N experts. Denote these predictions at time t by f 1,t,..., f N,t. Fix a convex loss function l. We suppose that l(a, b) [0, 1] for simplicity. At each time step t = 1 to, Player observes f 1,t,..., f N,t and predicts p t Outcome y t is revealed Player suffers loss l(p t, y t ) and experts suffer l(f i,t, y t ) Denote the cumulative loss of expert i by L i,t = t s=1 l(f i,s, y s ) and the cumulative loss of the player by L t = t s=1 l(p s, y s ). he goal of the Player is to minimize the regret Lecturer: Sasha Rakhlin 1
2 l(p t, y t ) min i 1...N l(f i,t, y t ) = L min L i,. How can we make predictions based on the history so that R is small? Solution (called Exponential Weights or Weighted Majority [4]): keep weights w 1,t,..., w N,t over experts and predict p t = w i,t 1f i,t w. i,t 1 Once the outcome is revealed and the losses l(f i,t, y t ) can be calculated, the weights are updated as where η is the learning rate parameter. w i,t = w i,t 1 exp( ηl(f i,t, y t )), heorem 2.1. For the Exponential Weights algorithm with η = R (/2) ln N ln N, Proof, see [3, 2]. Let W t = w i,t. Suppose we initialize w i,0 = 1 for all experts i. hen ln W 0 N. Furthermore, On the other hand, ln W W 0 N w i, ln N N exp( ηl i, ) ln N ln( max,...,n exp( ηl i, )) ln N = η min,...,n L i, ln N. ln W t W t 1 w i,t w i,t 1 exp( ηl(f i,t, y t )) exp( ηl i,t 1 ) exp( ηl i,t 1) exp( ηl(f i,t, y t )) w i,t 1 w i,t 1 η l(f i,t, y t )w i,t 1 w + η2 i,t 1 ηl(p t, y t ) + η2 Lecturer: Sasha Rakhlin 2
3 where the last inequality follows by the definition of p t and convexity of l (via an application of Jensen s inequality). he next to last inequality holds because for a random variable X [a, b], for any s R. See [3, 2] for more details. ln Ee sx sex + s2 (b a) 2 Summing the last inequality over t = 1,..., and observing that logs telescope, ln W W 0 η l(p t, y t ) + η 2 /. Combining the upper and lower bounds for ln W /W 0, Balancing the two terms with η = ln N L min L i, + ln N i, η gives the bound. + η. We ve presented the Prediction with Expert Advice framework and the Exponential Weights algorithm because they are the most widely known. However, the proof above does not give us much insight into why things work. Let us now go to a somewhat simpler setting by removing the extra layer of combining predictions of experts to make our own prediction. he following setting is closely related, but simpler to understand. 3 Online Gradient Descent Consider the following repeated game: At each time step t = 1 to, Player predicts w t, a distribution over N experts Vector of losses l t R N is revealed Player suffers l t w t, the vector product he goal is to minimize the regret, defined as l t w t min w N simplex l t w. Since the minimum over the simplex occurs at the vertices, we could equivalently write min i. Lecturer: Sasha Rakhlin 3
4 One can see that the game is closely related to the one introduced in the beginning of the lecture. Here w t are the normalized versions of the weights kept by the Exponential Weights algorithm. hink of the loss vector l t as the vector of l(f i,t, y t ). he repeated game we just defined is called an Online Linear Optimization game. In fact, we can define the Online Linear Optimization over any convex set K. In this case l t w t min w K l t w. Suppose we just perform gradient steps w t+1 = w t ηl t, followed by a projection onto the set K. It turns out that this is a very good (often optimal) algorithm. Let s prove that this algorithm is good. heorem 3.1 (Online Gradient Descent [5]). Suppose the Online Linear Game is performed by updating w t+1 = w t ηl t with η = 1/2, followed by the projection onto the set K. hen the regret R GD, where G is the maximum norm of the gradient of l t s and D is the diameter of K. Proof. By the definition of the update, w t+1 w 2 w t ηl t w 2 = w t w 2 + η 2 l t 2 2l t (w t w ), where w K is the optimum which can only be computed in hindsight. Solving for the last term, l t (w t w ) w t w 2 w t+1 w 2 2η + η 2 l t 2 Summing over time, l t w t l t w w 1 w 2 2η + η 2 l t 2. Assuming the upper bound on the norms and choosing η = D G, we obtain the bound on the regret. 4 Regularization Let us compare the bounds DG and (/2) log N. he asymptotic dependence on is the same this is typical for online learning if no second-order curvature information is taken into account. However, the bounds are different. For the simplex, the L 2 -diameter D N instead of log N. his difference is noticeable whenever one has a large (exponential in parameters) number of experts. So, have we lost something by considering the Online Gradient Descent instead of Exponential Weights? he answer is Lecturer: Sasha Rakhlin 4
5 no. It turns out that Exponential Weights can be viewed as Online Gradient Descent, but in a different (dual) space. So, the difference in N vs log N arises from the choice of Euclidean potential vs Entropy potential. he dual-space gradient descent is known as Mirror Descent (see Chapter 11 of [3]). Somewhat surprisingly, regularization leads to Mirror Descent, and this is a very general framework subsuming many known online algorithms. By regularization we mean the following Follow he Regularized Leader (FRL) algorithm w t+1 = arg min w t l s w + η 1 R(w) s=1 where R is an appropriate convex differentiable regularizer. When we take R(w) = 1 2 w 2, we obtain the Online Gradient Descent described above. When we take R(w) = w(i) log w(i), we get Exponential Weights for the simplex (or, more generally, Exponentiated Gradient). References [1] Nicolò Cesa-Bianchi, Yoav Freund, David Haussler, David P. Helmbold, Robert E. Schapire, and Manfred K. Warmuth. How to use expert advice. J. ACM, 44(3):427 45, [2] Nicolò Cesa-Bianchi and Gábor Lugosi. On prediction of individual sequences. Ann. Statist, 27(6): , [3] Nicolò Cesa-Bianchi and Gábor Lugosi. Prediction, Learning, and Games. Cambridge University Press, [4] Nick Littlestone and Manfred K. Warmuth. he weighted majority algorithm. Information and Computation, 10(2): , [5] Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In ICML, pages , Lecturer: Sasha Rakhlin 5
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Online Convex Optimization
E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting
Week 1: Introduction to Online Learning
Week 1: Introduction to Online Learning 1 Introduction This is written based on Prediction, Learning, and Games (ISBN: 2184189 / -21-8418-9 Cesa-Bianchi, Nicolo; Lugosi, Gabor 1.1 A Gentle Start Consider
Foundations of Machine Learning On-Line Learning. Mehryar Mohri Courant Institute and Google Research [email protected]
Foundations of Machine Learning On-Line Learning Mehryar Mohri Courant Institute and Google Research [email protected] Motivation PAC learning: distribution fixed over time (training and test). IID assumption.
17.3.1 Follow the Perturbed Leader
CS787: Advanced Algorithms Topic: Online Learning Presenters: David He, Chris Hopman 17.3.1 Follow the Perturbed Leader 17.3.1.1 Prediction Problem Recall the prediction problem that we discussed in class.
Strongly Adaptive Online Learning
Amit Daniely Alon Gonen Shai Shalev-Shwartz The Hebrew University [email protected] [email protected] [email protected] Abstract Strongly adaptive algorithms are algorithms whose performance
Table 1: Summary of the settings and parameters employed by the additive PA algorithm for classification, regression, and uniclass.
Online Passive-Aggressive Algorithms Koby Crammer Ofer Dekel Shai Shalev-Shwartz Yoram Singer School of Computer Science & Engineering The Hebrew University, Jerusalem 91904, Israel {kobics,oferd,shais,singer}@cs.huji.ac.il
The Advantages and Disadvantages of Online Linear Optimization
LINEAR PROGRAMMING WITH ONLINE LEARNING TATSIANA LEVINA, YURI LEVIN, JEFF MCGILL, AND MIKHAIL NEDIAK SCHOOL OF BUSINESS, QUEEN S UNIVERSITY, 143 UNION ST., KINGSTON, ON, K7L 3N6, CANADA E-MAIL:{TLEVIN,YLEVIN,JMCGILL,MNEDIAK}@BUSINESS.QUEENSU.CA
Introduction to Online Learning Theory
Introduction to Online Learning Theory Wojciech Kot lowski Institute of Computing Science, Poznań University of Technology IDSS, 04.06.2013 1 / 53 Outline 1 Example: Online (Stochastic) Gradient Descent
Introduction to Online Optimization
Princeton University - Department of Operations Research and Financial Engineering Introduction to Online Optimization Sébastien Bubeck December 14, 2011 1 Contents Chapter 1. Introduction 5 1.1. Statistical
Online Learning. 1 Online Learning and Mistake Bounds for Finite Hypothesis Classes
Advanced Course in Machine Learning Spring 2011 Online Learning Lecturer: Shai Shalev-Shwartz Scribe: Shai Shalev-Shwartz In this lecture we describe a different model of learning which is called online
Achieving All with No Parameters: AdaNormalHedge
JMLR: Workshop and Conference Proceedings vol 40:1 19, 2015 Achieving All with No Parameters: AdaNormalHedge Haipeng Luo Department of Computer Science, Princeton University Robert E. Schapire Microsoft
How to Use Expert Advice
NICOLÒ CESA-BIANCHI Università di Milano, Milan, Italy YOAV FREUND AT&T Labs, Florham Park, New Jersey DAVID HAUSSLER AND DAVID P. HELMBOLD University of California, Santa Cruz, Santa Cruz, California
Online Convex Programming and Generalized Infinitesimal Gradient Ascent
Online Convex Programming and Generalized Infinitesimal Gradient Ascent Martin Zinkevich February 003 CMU-CS-03-110 School of Computer Science Carnegie Mellon University Pittsburgh, PA 1513 Abstract Convex
The p-norm generalization of the LMS algorithm for adaptive filtering
The p-norm generalization of the LMS algorithm for adaptive filtering Jyrki Kivinen University of Helsinki Manfred Warmuth University of California, Santa Cruz Babak Hassibi California Institute of Technology
Notes from Week 1: Algorithms for sequential prediction
CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking
The Multiplicative Weights Update method
Chapter 2 The Multiplicative Weights Update method The Multiplicative Weights method is a simple idea which has been repeatedly discovered in fields as diverse as Machine Learning, Optimization, and Game
Online Algorithms: Learning & Optimization with No Regret.
Online Algorithms: Learning & Optimization with No Regret. Daniel Golovin 1 The Setup Optimization: Model the problem (objective, constraints) Pick best decision from a feasible set. Learning: Model the
Online Semi-Supervised Learning
Online Semi-Supervised Learning Andrew B. Goldberg, Ming Li, Xiaojin Zhu [email protected] Computer Sciences University of Wisconsin Madison Xiaojin Zhu (Univ. Wisconsin-Madison) Online Semi-Supervised
Follow the Leader with Dropout Perturbations
JMLR: Worshop and Conference Proceedings vol 35: 26, 204 Follow the Leader with Dropout Perturbations Tim van Erven Département de Mathématiques, Université Paris-Sud, France Wojciech Kotłowsi Institute
4 Learning, Regret minimization, and Equilibria
4 Learning, Regret minimization, and Equilibria A. Blum and Y. Mansour Abstract Many situations involve repeatedly making decisions in an uncertain environment: for instance, deciding what route to drive
Simple and efficient online algorithms for real world applications
Simple and efficient online algorithms for real world applications Università degli Studi di Milano Milano, Italy Talk @ Centro de Visión por Computador Something about me PhD in Robotics at LIRA-Lab,
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
1 Portfolio Selection
COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture # Scribe: Nadia Heninger April 8, 008 Portfolio Selection Last time we discussed our model of the stock market N stocks start on day with
LARGE CLASSES OF EXPERTS
LARGE CLASSES OF EXPERTS Csaba Szepesvári University of Alberta CMPUT 654 E-mail: [email protected] UofA, October 31, 2006 OUTLINE 1 TRACKING THE BEST EXPERT 2 FIXED SHARE FORECASTER 3 VARIABLE-SHARE
How To Find Out How Regret Works In An Online Structured Learning Problem
Alon Cohen [email protected] University of Haifa, University of Haifa, Dept. of Computer Science, 31905 Haifa, Israel Tamir Hazan [email protected] University of Haifa, University of Haifa,
Logistic Regression for Spam Filtering
Logistic Regression for Spam Filtering Nikhila Arkalgud February 14, 28 Abstract The goal of the spam filtering problem is to identify an email as a spam or not spam. One of the classic techniques used
Online Learning and Online Convex Optimization. Contents
Foundations and Trends R in Machine Learning Vol. 4, No. 2 (2011) 107 194 c 2012 S. Shalev-Shwartz DOI: 10.1561/2200000018 Online Learning and Online Convex Optimization By Shai Shalev-Shwartz Contents
6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation
6.207/14.15: Networks Lecture 15: Repeated Games and Cooperation Daron Acemoglu and Asu Ozdaglar MIT November 2, 2009 1 Introduction Outline The problem of cooperation Finitely-repeated prisoner s dilemma
Online Bandit Learning against an Adaptive Adversary: from Regret to Policy Regret
Online Bandit Learning against an Adaptive Adversary: from Regret to Policy Regret Raman Arora Toyota Technological Institute at Chicago, Chicago, IL 60637, USA Ofer Dekel Microsoft Research, 1 Microsoft
Competing in the Dark: An Efficient Algorithm for Bandit Linear Optimization
Competing in the Dark: An Efficient Algorithm for Bandit Linear Optimization Jacob Abernethy Computer Science Division UC Berkeley [email protected] Elad Hazan IBM Almaden [email protected] Alexander
Robbing the bandit: Less regret in online geometric optimization against an adaptive adversary.
Robbing the bandit: Less regret in online geometric optimization against an adaptive adversary. Varsha Dani Thomas P. Hayes November 12, 2005 Abstract We consider online bandit geometric optimization,
arxiv:1112.0829v1 [math.pr] 5 Dec 2011
How Not to Win a Million Dollars: A Counterexample to a Conjecture of L. Breiman Thomas P. Hayes arxiv:1112.0829v1 [math.pr] 5 Dec 2011 Abstract Consider a gambling game in which we are allowed to repeatedly
arxiv:0710.0485v2 [cs.lg] 27 Jun 2008
Prediction with expert advice for the Brier game arxiv:07100485v [cslg] 7 Jun 008 Vladimir Vovk and Fedor Zhdanov Computer Learning Research Centre Department of Computer Science Royal Holloway, University
Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.
1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.
9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1
9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1 Seffi Naor Computer Science Dept. Technion Haifa, Israel Introduction
Gambling and Data Compression
Gambling and Data Compression Gambling. Horse Race Definition The wealth relative S(X) = b(x)o(x) is the factor by which the gambler s wealth grows if horse X wins the race, where b(x) is the fraction
Load Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: [email protected] Abstract Load
Online Learning, Stability, and Stochastic Gradient Descent
Online Learning, Stability, and Stochastic Gradient Descent arxiv:1105.4701v3 [cs.lg] 8 Sep 2011 September 9, 2011 Tomaso Poggio, Stephen Voinea, Lorenzo Rosasco CBCL, McGovern Institute, CSAIL, Brain
Prediction With Expert Advice For The Brier Game
Journal of Machine Learning Research 0 (2009) 2445-247 Submitted 6/09; Published /09 Prediction With Expert Advice For The Brier Game Vladimir Vovk Fedor Zhdanov Computer Learning Research Centre Department
GI01/M055 Supervised Learning Proximal Methods
GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators
(Quasi-)Newton methods
(Quasi-)Newton methods 1 Introduction 1.1 Newton method Newton method is a method to find the zeros of a differentiable non-linear function g, x such that g(x) = 0, where g : R n R n. Given a starting
CS 688 Pattern Recognition Lecture 4. Linear Models for Classification
CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(
Online Lazy Updates for Portfolio Selection with Transaction Costs
Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence Online Lazy Updates for Portfolio Selection with Transaction Costs Puja Das, Nicholas Johnson, and Arindam Banerjee Department
Interior-Point Methods for Full-Information and Bandit Online Learning
4164 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 58, NO. 7, JULY 2012 Interior-Point Methods for Full-Information and Bandit Online Learning Jacob D. Abernethy, Elad Hazan, and Alexander Rakhlin Abstract
Game Theory: Supermodular Games 1
Game Theory: Supermodular Games 1 Christoph Schottmüller 1 License: CC Attribution ShareAlike 4.0 1 / 22 Outline 1 Introduction 2 Model 3 Revision questions and exercises 2 / 22 Motivation I several solution
Economics 1011a: Intermediate Microeconomics
Lecture 12: More Uncertainty Economics 1011a: Intermediate Microeconomics Lecture 12: More on Uncertainty Thursday, October 23, 2008 Last class we introduced choice under uncertainty. Today we will explore
Tail inequalities for order statistics of log-concave vectors and applications
Tail inequalities for order statistics of log-concave vectors and applications Rafał Latała Based in part on a joint work with R.Adamczak, A.E.Litvak, A.Pajor and N.Tomczak-Jaegermann Banff, May 2011 Basic
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem
14.1 Rent-or-buy problem
CS787: Advanced Algorithms Lecture 14: Online algorithms We now shift focus to a different kind of algorithmic problem where we need to perform some optimization without knowing the input in advance. Algorithms
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Approximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
ALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand
Notes V General Equilibrium: Positive Theory In this lecture we go on considering a general equilibrium model of a private ownership economy. In contrast to the Notes IV, we focus on positive issues such
Monotone multi-armed bandit allocations
JMLR: Workshop and Conference Proceedings 19 (2011) 829 833 24th Annual Conference on Learning Theory Monotone multi-armed bandit allocations Aleksandrs Slivkins Microsoft Research Silicon Valley, Mountain
Temporal Difference Learning in the Tetris Game
Temporal Difference Learning in the Tetris Game Hans Pirnay, Slava Arabagi February 6, 2009 1 Introduction Learning to play the game Tetris has been a common challenge on a few past machine learning competitions.
The Online Set Cover Problem
The Online Set Cover Problem Noga Alon Baruch Awerbuch Yossi Azar Niv Buchbinder Joseph Seffi Naor ABSTRACT Let X = {, 2,..., n} be a ground set of n elements, and let S be a family of subsets of X, S
Online Learning: Theory, Algorithms, and Applications. Thesis submitted for the degree of Doctor of Philosophy by Shai Shalev-Shwartz
Online Learning: Theory, Algorithms, and Applications Thesis submitted for the degree of Doctor of Philosophy by Shai Shalev-Shwartz Submitted to the Senate of the Hebrew University July 2007 This work
CSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
Master s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
Degeneracy in Linear Programming
Degeneracy in Linear Programming I heard that today s tutorial is all about Ellen DeGeneres Sorry, Stan. But the topic is just as interesting. It s about degeneracy in Linear Programming. Degeneracy? Students
Adaptive Bound Optimization for Online Convex Optimization
Adaptive Bound Optimization for Online Convex Optimization H. Brendan McMahan Google, Inc. [email protected] Matthew Streeter Google, Inc. [email protected] Abstract We introduce a new online convex
How I won the Chess Ratings: Elo vs the rest of the world Competition
How I won the Chess Ratings: Elo vs the rest of the world Competition Yannis Sismanis November 2010 Abstract This article discusses in detail the rating system that won the kaggle competition Chess Ratings:
Universal Algorithm for Trading in Stock Market Based on the Method of Calibration
Universal Algorithm for Trading in Stock Market Based on the Method of Calibration Vladimir V yugin Institute for Information Transmission Problems, Russian Academy of Sciences, Bol shoi Karetnyi per.
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE YUAN TIAN This synopsis is designed merely for keep a record of the materials covered in lectures. Please refer to your own lecture notes for all proofs.
Big Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
Mathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
A simple and fast algorithm for computing exponentials of power series
A simple and fast algorithm for computing exponentials of power series Alin Bostan Algorithms Project, INRIA Paris-Rocquencourt 7815 Le Chesnay Cedex France and Éric Schost ORCCA and Computer Science Department,
Universal Algorithms for Probability Forecasting
Universal Algorithms for Probability Forecasting Fedor Zhdanov Computer Learning Research Centre, Department of Computer Science, Royal Holloway University of London, Egham, Surrey, TW0 0EX, UK. [email protected]
On Adaboost and Optimal Betting Strategies
On Adaboost and Optimal Betting Strategies Pasquale Malacaria School of Electronic Engineering and Computer Science Queen Mary, University of London Email: [email protected] Fabrizio Smeraldi School of
GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1
(R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter [email protected] February 20, 2009 Notice: This work was supported by the U.S.
Nonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan
Example 4.1 (nonlinear pendulum dynamics with friction) Figure 4.1: Pendulum. asin. k, a, and b. We study stability of the origin x
Lecture 4. LaSalle s Invariance Principle We begin with a motivating eample. Eample 4.1 (nonlinear pendulum dynamics with friction) Figure 4.1: Pendulum Dynamics of a pendulum with friction can be written
A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property
A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property Venkat Chandar March 1, 2008 Abstract In this note, we prove that matrices whose entries are all 0 or 1 cannot achieve
Find-The-Number. 1 Find-The-Number With Comps
Find-The-Number 1 Find-The-Number With Comps Consider the following two-person game, which we call Find-The-Number with Comps. Player A (for answerer) has a number x between 1 and 1000. Player Q (for questioner)
Online Ad Auctions. By Hal R. Varian. Draft: February 16, 2009
Online Ad Auctions By Hal R. Varian Draft: February 16, 2009 I describe how search engines sell ad space using an auction. I analyze advertiser behavior in this context using elementary price theory and
Can We Learn to Beat the Best Stock
Can We Learn to Beat the Allan Borodin Ran El-Yaniv 2 Vincent Gogan Department of Computer Science University of Toronto Technion - Israel Institute of Technology 2 {bor,vincent}@cs.toronto.edu [email protected]
CSC2420 Fall 2012: Algorithm Design, Analysis and Theory
CSC2420 Fall 2012: Algorithm Design, Analysis and Theory Allan Borodin November 15, 2012; Lecture 10 1 / 27 Randomized online bipartite matching and the adwords problem. We briefly return to online algorithms
The Relative Worst Order Ratio for On-Line Algorithms
The Relative Worst Order Ratio for On-Line Algorithms Joan Boyar 1 and Lene M. Favrholdt 2 1 Department of Mathematics and Computer Science, University of Southern Denmark, Odense, Denmark, [email protected]
Big Data Analytics: Optimization and Randomization
Big Data Analytics: Optimization and Randomization Tianbao Yang, Qihang Lin, Rong Jin Tutorial@SIGKDD 2015 Sydney, Australia Department of Computer Science, The University of Iowa, IA, USA Department of
