Summer School Machine Learning Trimester
|
|
- Theodora Hancock
- 8 years ago
- Views:
Transcription
1 Summer School Machine Learning Trimester Université Paul Sabatier, Toulouse September 1418, 2015
2 About the CIMI Machine Learning thematic trimester From September to December 2015, the International Centre for Mathematics and Computer Science (CIMI) organizes a thematic trimester on Machine Learning. The goal of this trimester is to propose a series of scientic and pedagogical events reecting common interests of the two laboratories that founded CIMI: Institut de Mathématiques de Toulouse (IMT) and Institut de Recherche en Informatique de Toulouse (IRIT). The trimester will start with a summer school and will continue with three thematic workshops. Summer school - 14th to 18th September 2015 Workshop 1 - Optimization in Machine Learning, Vision and Image Processing - 6th to 7th October 2015 Workshop 2 - Sequential Learning and Applications - 9th to 10th November 2015 Workshop 3 - Learning with Structured Data and Natural Language - 9th to 11th December 2015 The Summer School We propose four courses: Optimization in Machine Learning, Information Retrieval and Machine Learning, Reinforcement Learning and Dictionary Learning. Each course consists of three 2h lectures, and is illustrated by a hands-on computer session. In addition, an invited talk by Noah Smith will be given on Wednesday morning. For the hands-on sessions: Be aware that no computer will be provided for the hands-on sessions : participants are encouraged to bring their own laptops. The hands-on sessions on Optimization in Machine Learning, Reinforcement Learning, and Dictionary Learning will use the language julia. This free software is available for download at Alternatively, participants will be able to run julia as a web service at (this requires to sign in with a Google account, no download or installation required). Code and data for the sessions are available at For the hands-on session on Information Retrieval and Machine Learning, several languages can be used, but it is recommended to install Python and IPython on your laptop: cf. Local Information All courses and hands-on sessions will take place in Amphitheater Einstein, building 3TP2, Université Paul Sabatier (campus de Rangueil). The map of the campus is available at: Contact aurelien.garivier@math.univ-toulouse.fr sebastien.gerchinovitz@math.univ-toulouse.fr josiane.mothe@irit.fr mathieu.serrurier@irit.fr Trimester Scientic Committee Francis Bach (Ecole Normale Supérieure), Sébastien Bubeck (Microsoft Research), Nicolo Cesa-Bianchi (Universitá degli Studi di Milano), Rémi Gribonval (lrisa Rennes), Marc Sebban (University Jean Monnet), Noah Smith (Carnegie Mellon University), Johan Suykens (KU Leuven), Marc Teboulle (Tel-Aviv University).
3 Schedule Monday, September 14th 9:30 Welcome Reception 10:30 A. Garivier Reinforcement Learning I 14:00 M. Melucci Information Retrieval and Machine Learning I 15:50 Coee Break 16:10 P. Richtárik Optimization in Machine Learning I Tuesday, September 15th 10:00 Coee 10:30 B. Scherrer Reinforcement Learning II 14:00 P. Richtárik Optimization in Machine Learning II 15:50 Coee Break 16:10 A. Lazaric & A. Garivier Hands-on session: Reinforcement Learning Wednesday, September 16th 8:30 N. Smith Structured Sparsity in Natural Language Processing (invited talk) 10:20 Coee Break 10:40 P. Richtárik Optimization in Machine Learning III 14:00 M. Melucci Information Retrieval and Machine Learning II 15:50 Coee Break 16:10 P. Richtárik Hands-on session: Optimization in Machine Learning Thursday, September 17th 8:30 J. Mairal Dictionary Learning I 10:20 Coee Break 10:40 B. Scherrer Reinforcement Learning III 14:00 J. Mairal Dictionary Learning II 15:50 Coee Break 16:10 J. Mairal & S. Gerchinovitz Hands-on session: Dictionary Learning Friday, September 18th 8:30 M. Melucci Information Retrieval and Machine Learning III 10:20 Coee Break 10:40 J. Mairal Dictionary Learning III 14:00 M. Melucci Hands-on session: Information Retrieval and Machine Learning 16:00 Closing Reception
4 Course 1: Optimization in Machine Learning, by Peter Richtárik Peter Richtárik is an Assistant Professor of Optimization at the University of Edinburgh. His research interests are in all areas of data science that intersect with optimization, including algorithms, machine learning, statistics, operations research, mathematics and high performance computing. Abstract: This course covers recent advances in scalable algorithms for convex optimization, with a particular emphasis on training (linear) predictors via the empirical risk minimization (ERM) paradigm. The material will be presented in a unied way wherever possible. Randomized, deterministic, primal, dual, accelerated, serial, parallel and distributed methods will be mentioned. The course will start in an unusual place: a concise yet powerful theory of randomized iterative methods for linear systems. While of an independent interest, this will highlight many of the algorithmic schemes and tools we shall encounter later in the course. Outline of the lectures: Lecture 1: Randomized Iterative Methods for Linear Systems (and more) [14] Lecture 2: Randomized Dual Methods [2, 3, 4, 6, 7, 10, 11, 12, 14] Lecture 3: Randomized Primal Methods [1, 5, 8, 9, 14, 15] Hands-on session: Minimizing Finite Sums via Dual-Free SDCA [15] Some references: 1. S. Shalev-Shwartz, Y. Singer, N. Srebro and A. Cotter. Pegasos: primal estimated sub-gradient solver for SVM, Mathematical Programming, 127(1), pp 330, P. Richtárik and M. Taká. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function, Mathematical Programming 144(2), pp 138, 2014 (arxiv: ) 3. S. Shalev-Shwartz and T. Zhang. Stochastic dual coordinate ascent methods for regularized loss minimization, Journal of Machine Learning Research 14, , 2013 (arxiv: ) 4. P. Richtárik and M. Taká. Parallel coordinate descent methods for big data optimization, Mathematical Programming, 2015 (arxiv: ) 5. N. Le Roux, M. Schmidt, and F. Bach. A stochastic gradient method with an exponential convergence rate for nite training sets, NIPS P. Richtárik and M. Taká. Distributed coordinate descent method for learning with big data, arxiv: , P. Richtárik and M. Taká. On optimal probabilities in stochastic coordinate descent methods, Optimization Letters, 2015 (arxiv: ) 8. R. Johnson and T. Zhang. Accelerating stochastic gradient descent using predictive variance reduction, NIPS, J. Kone ný and P. Richtárik. Semi-stochastic gradient descent methods, arxiv: , O. Fercoq and P. Richtárik Accelerated, Parallel and PROXimal coordinate descent, SIAM Journal on Optimization, 2015 (arxiv: ) 11. Z. Qu, P. Richtárik and T. Zhang. Randomized dual coordinate ascent with arbitrary sampling, NIPS, 2015 (arxiv: ) 12. Z. Qu, P. Richtárik, M. Taká and O. Fercoq. SDNA: Stochastic dual Newton ascent for empirical risk minimization, arxiv: , S. Shalev-Shwartz. SDCA without duality, arxiv: , R. Gower and P. Richtárik. Randomized iterative methods for linear systems, arxiv: , D. Csiba and P. Richtárik. Primal method for ERM with exible mini-batching schemes and non-convex losses, arxiv: , 2015.
5 Course 2: Information Retrieval and Machine Learning, by Massimo Melucci Massimo Melucci has been Associate Professor in Computer Science at the Department of Information Engineering of the University of Padua, Italy, since He is on the Editorial Board of the Journal of IR and Associate Editor of Computer Science Review. His research interests are mainly in IR modeling. He is also currently investigating the intersection between IR and machine learning and the use of quantum mechanics in IR. He has been involved in EU and national research projects. Abstract: This course is an introduction to the intersection between Information Retrieval (IR) and Machine Learning (ML) models. ML has been at the basis of some IR tasks such as document ranking and relevance feedback. On the other hand IR poses new challenges to ML because of the peculiar nature of the context in which data are observed. In this course, I will introduce rst the tasks of IR and then the utilisation of some ML techniques to address these tasks. Outline of the lectures: Lecture 1: Introduction to Information Retrieval (key concepts, relevance feedback, evaluation methodology). Lecture 2: Information Retrieval Modeling (key concepts, boolean modeling, vector space modeling, relevance modeling, language modeling, evaluation results). Lecture 3: Machine Learning and Information Retrieval (key concepts, correspondence between IR and ML, features, approaches, applications, evaluation methodology). Hands-on session: Document Ranking: an Introduction. Some references: R. Baeza-Yates and B. Ribeiro-Neto. Modern Information Retrieval. Addison Wesley, New York, USA, II edition, W. Croft, D. Metzler, and T. Strohman. Search Engines: Information Retrieval in Practice. Addison Wesley, H. Li. Learning to Rank for Information Retrieval and Natural Language Processing. Morgan and Claypool, T.-Y. Liu. Learning to Rank for Information Retrieval. Springer, C. Manning, P. Raghavan, and H. Schütze. An Introduction to Information Retrieval. Cambridge University Press, G. Salton and M. McGill. Introduction to Modern Information Retrieval. McGraw-Hill, New York, NY, C. J. Van Rijsbergen. Information Retrieval. Butterworths, London, second edition, 1979.
6 Course 3: Reinforcement Learning, by Bruno Scherrer and Alessandro Lazaric Bruno Scherrer is a research scientist at INRIA in the project BIGS. He is a member of the Probability and Statistics Team at Institut Elie Cartan of Lorraine (IECL). His main research interests are stochastic optimal control, reinforcement learning, Markov decision processes, approximate dynamic programming, analysis of algorithms and stochastic processes. Alessandro Lazaric is a Junior Researcher (CR1) at INRIA Lille - Nord Europe in the SequeL team. Abstract: The course will cover the basic models and techniques of reinforcement learning (RL). We will begin by reviewing the Markov decision process (MDP) model used to formalize the interaction between a learning agent and an (unknown) dynamic environment. After introducing the dynamic programming techniques used to compute the exact optimal solution of an MDP known in advance, we will move to the actual learning problem where the MDP is unknown. We will introduce popular algorithms such as Q-learning and SARSA. This will lead to the analysis of two of the most important aspects of RL algorithms: how to trade o exploration and exploitation, and how to accurately approximate solutions. The core of the exploration-exploitation problem will be studied in the celebrated multi-armed bandit framework and its application to modern recommendation systems. Finally, a few examples of approximate dynamic programming will be presented together with some guarantees on their performance. The hands-on session will focus on implementing multi-armed bandit algorithms applied to the problem of policy optimization and online RL for a simple management problem. Outline of the lectures: Lecture 1: Introduction to RL: the Multi-Armed Bandit Problem. Lecture 2: Markov Decision Processes, Planning, and Dynamic Programming. Lecture 3: Approximate Solutions for Continuous MDPs. Hands-on session: Bandit Algorithms for Policy Optimization. Some references: R. Bellman, Dynamic Programming, Princeton University Press (1957). D. Bertsekas, Dynamic Programming and Optimal Control, Athena Scientic (2005). R. Sutton, Reinforcement Learning - An Introduction, MIT Press (1998). C. Szepesvari, Algorithms for Reinforcement Learning, Morgan & Claypool Publishers (2010).
7 Course 4: Dictionary Learning, by Julien Mairal Julien Mairal is a research scientist at INRIA in the project LEAR. He was previously a postdoctoral researcher in the statistics department of the university of California, Berkeley. Before that, he did his PhD at INRIA in the project WILLOW under the supervision of Jean Ponce and Francis Bach. He is interested in machine learning, optimization, computer vision, statistical signal and image processing, and also have some interest in bio-informatics and neurosciences. Abstract: In recent years, a large amount of multi-disciplinary research has been conducted on sparse models and their applications. In statistics and machine learning, the sparsity principle is used to perform model selectionthat is, automatically selecting a simple model among a large collection of them. In signal processing, sparse coding consists of representing data with linear combinations of a few dictionary elements. Subsequently, the corresponding tools have been widely adopted by several scientic communities such as neuroscience, bioinformatics, or computer vision. The goal of this course is to oer a self-contained view of sparse modeling for visual recognition and image processing. More specically, we focus on applications where the dictionary is learned and adapted to data, yielding a compact representation that has been successful in various contexts. Outline of the lectures: Lecture 1: A Short Introduction to Parsimony. Lecture 2: Sparse Models for Image Processing. Lecture 3: Optimization for Sparse Estimation. Hands-on session: Sparse Estimation for Image and Vision Processing. Some references: J. Mairal, F. Bach, and J. Ponce, Sparse Estimation for Image and Vision Processing, Foundations and Trends in Computer Graphics and Vision F. Bach, R. Jenatton, J. Mairal, and G. Obozinski, Optimization with sparsity-inducing penalties, Foundations and Trends in Machine Learning, 2012.
8 About CIMI CIMI stands for Centre International de Mathématiques et Informatique de Toulouse and it is one of the Excellence projects selected by the ANR for the period CIMI brings together the teams of the Institut de Mathématiques de Toulouse (IMT) and the Institut de Recherche en Informatique de Toulouse (IRIT). It aims at becoming an international reference in mathematics, computer science and their interactions. The program will attract high-level scientists and students from around the world. It includes actions towards attractiveness, such as Excellence Chairs for long-term visitors, grants for Doctoral and Post-Doctoral students, as well as fellowships for Master students. Attractiveness is further enhanced with thematic trimesters organized within CIMI on specic topics including courses, seminars and workshops. The innovative tools developed at CIMI will also have a strong economic impact on the region and will prot its industrial partners in major technological areas, making CIMI a strategic partner of the social and economic world.
Machine learning challenges for big data
Machine learning challenges for big data Francis Bach SIERRA Project-team, INRIA - Ecole Normale Supérieure Joint work with R. Jenatton, J. Mairal, G. Obozinski, N. Le Roux, M. Schmidt - December 2012
More informationWorkshop 1: Optimization in machine learning, vision and image processing
Workshop 1: Optimization in machine learning, vision and image processing Université Paul Sabatier, Toulouse, October 6th-7th 2015 About the CIMI Machine Learning thematic trimester From September to December
More informationBig learning: challenges and opportunities
Big learning: challenges and opportunities Francis Bach SIERRA Project-team, INRIA - Ecole Normale Supérieure December 2013 Omnipresent digital media Scientific context Big data Multimedia, sensors, indicators,
More informationStochastic gradient methods for machine learning
Stochastic gradient methods for machine learning Francis Bach INRIA - Ecole Normale Supérieure, Paris, France Joint work with Eric Moulines, Nicolas Le Roux and Mark Schmidt - April 2013 Context Machine
More informationStochastic gradient methods for machine learning
Stochastic gradient methods for machine learning Francis Bach INRIA - Ecole Normale Supérieure, Paris, France Joint work with Eric Moulines, Nicolas Le Roux and Mark Schmidt - April 2013 Context Machine
More informationBeyond stochastic gradient descent for large-scale machine learning
Beyond stochastic gradient descent for large-scale machine learning Francis Bach INRIA - Ecole Normale Supérieure, Paris, France Joint work with Eric Moulines, Nicolas Le Roux and Mark Schmidt - ECML-PKDD,
More informationParallel & Distributed Optimization. Based on Mark Schmidt s slides
Parallel & Distributed Optimization Based on Mark Schmidt s slides Motivation behind using parallel & Distributed optimization Performance Computational throughput have increased exponentially in linear
More informationLarge-Scale Similarity and Distance Metric Learning
Large-Scale Similarity and Distance Metric Learning Aurélien Bellet Télécom ParisTech Joint work with K. Liu, Y. Shi and F. Sha (USC), S. Clémençon and I. Colin (Télécom ParisTech) Séminaire Criteo March
More informationFederated Optimization: Distributed Optimization Beyond the Datacenter
Federated Optimization: Distributed Optimization Beyond the Datacenter Jakub Konečný School of Mathematics University of Edinburgh J.Konecny@sms.ed.ac.uk H. Brendan McMahan Google, Inc. Seattle, WA 98103
More informationBig Data Analytics: Optimization and Randomization
Big Data Analytics: Optimization and Randomization Tianbao Yang, Qihang Lin, Rong Jin Tutorial@SIGKDD 2015 Sydney, Australia Department of Computer Science, The University of Iowa, IA, USA Department of
More informationOperations Research and Financial Engineering. Courses
Operations Research and Financial Engineering Courses ORF 504/FIN 504 Financial Econometrics Professor Jianqing Fan This course covers econometric and statistical methods as applied to finance. Topics
More informationLecture 1: Introduction to Reinforcement Learning
Lecture 1: Introduction to Reinforcement Learning David Silver Outline 1 Admin 2 About Reinforcement Learning 3 The Reinforcement Learning Problem 4 Inside An RL Agent 5 Problems within Reinforcement Learning
More informationTD(0) Leads to Better Policies than Approximate Value Iteration
TD(0) Leads to Better Policies than Approximate Value Iteration Benjamin Van Roy Management Science and Engineering and Electrical Engineering Stanford University Stanford, CA 94305 bvr@stanford.edu Abstract
More informationBig Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
More informationAccelerated Parallel Optimization Methods for Large Scale Machine Learning
Accelerated Parallel Optimization Methods for Large Scale Machine Learning Haipeng Luo Princeton University haipengl@cs.princeton.edu Patrick Haffner and Jean-François Paiement AT&T Labs - Research {haffner,jpaiement}@research.att.com
More informationMachine Learning. 01 - Introduction
Machine Learning 01 - Introduction Machine learning course One lecture (Wednesday, 9:30, 346) and one exercise (Monday, 17:15, 203). Oral exam, 20 minutes, 5 credit points. Some basic mathematical knowledge
More informationDistributed Coordinate Descent Method for Learning with Big Data
Peter Richtárik Martin Takáč University of Edinburgh, King s Buildings, EH9 3JZ Edinburgh, United Kingdom PETER.RICHTARIK@ED.AC.UK MARTIN.TAKI@GMAIL.COM Abstract In this paper we develop and analyze Hydra:
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More information350 Serra Mall, Stanford, CA 94305-9515
Meisam Razaviyayn Contact Information Room 260, Packard Building 350 Serra Mall, Stanford, CA 94305-9515 E-mail: meisamr@stanford.edu Research Interests Education Appointments Large scale data driven optimization
More informationStochastic Optimization for Big Data Analytics: Algorithms and Libraries
Stochastic Optimization for Big Data Analytics: Algorithms and Libraries Tianbao Yang SDM 2014, Philadelphia, Pennsylvania collaborators: Rong Jin, Shenghuo Zhu NEC Laboratories America, Michigan State
More informationWinter 2016 Course Timetable. Legend: TIME: M = Monday T = Tuesday W = Wednesday R = Thursday F = Friday BREATH: M = Methodology: RA = Research Area
Winter 2016 Course Timetable Legend: TIME: M = Monday T = Tuesday W = Wednesday R = Thursday F = Friday BREATH: M = Methodology: RA = Research Area Please note: Times listed in parentheses refer to the
More informationemail: marco.frittelli@unimi.it Professor of Mathematical Finance Office phone: Italy+ 02 50316143
MARCO FRITTELLI email: marco.frittelli@unimi.it Professor of Mathematical Finance Office phone: Italy+ 02 50316143 CV Personal data and studies: Italian and USA citizenship. Degree (Laurea) in Mathematics
More informationMachine learning and optimization for massive data
Machine learning and optimization for massive data Francis Bach INRIA - Ecole Normale Supérieure, Paris, France ÉCOLE NORMALE SUPÉRIEURE Joint work with Eric Moulines - IHES, May 2015 Big data revolution?
More informationGroup Sparse Coding. Fernando Pereira Google Mountain View, CA pereira@google.com. Dennis Strelow Google Mountain View, CA strelow@google.
Group Sparse Coding Samy Bengio Google Mountain View, CA bengio@google.com Fernando Pereira Google Mountain View, CA pereira@google.com Yoram Singer Google Mountain View, CA singer@google.com Dennis Strelow
More informationLarge Scale Learning to Rank
Large Scale Learning to Rank D. Sculley Google, Inc. dsculley@google.com Abstract Pairwise learning to rank methods such as RankSVM give good performance, but suffer from the computational burden of optimizing
More informationBig Data Optimization: Randomized lock-free methods for minimizing partially separable convex functions
Big Data Optimization: Randomized lock-free methods for minimizing partially separable convex functions Peter Richtárik School of Mathematics The University of Edinburgh Joint work with Martin Takáč (Edinburgh)
More informationParallel Data Mining. Team 2 Flash Coders Team Research Investigation Presentation 2. Foundations of Parallel Computing Oct 2014
Parallel Data Mining Team 2 Flash Coders Team Research Investigation Presentation 2 Foundations of Parallel Computing Oct 2014 Agenda Overview of topic Analysis of research papers Software design Overview
More informationA primal-dual algorithm for group sparse regularization with overlapping groups
A primal-dual algorithm for group sparse regularization with overlapping groups Sofia Mosci DISI- Università di Genova mosci@disi.unige.it Alessandro Verri DISI- Università di Genova verri@disi.unige.it
More informationAntonino Freno. Curriculum Vitae. Phone (office): Office: +33 (0)3 59 35 87 27. antonino.freno@inria.fr; http://researchers.lille.inria.fr/~freno/.
Antonino Freno Curriculum Vitae Personal Information First name: Antonino Family name: Freno Date of birth: July 1, 1980 Place of birth: Reggio Calabria (RC) Italy Citizenship: Italian Phone (office):
More informationSimple and efficient online algorithms for real world applications
Simple and efficient online algorithms for real world applications Università degli Studi di Milano Milano, Italy Talk @ Centro de Visión por Computador Something about me PhD in Robotics at LIRA-Lab,
More informationOnline Convex Optimization
E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting
More informationKarthik Sridharan. 424 Gates Hall Ithaca, E-mail: sridharan@cs.cornell.edu http://www.cs.cornell.edu/ sridharan/ Contact Information
Karthik Sridharan Contact Information 424 Gates Hall Ithaca, NY 14853-7501 USA E-mail: sridharan@cs.cornell.edu http://www.cs.cornell.edu/ sridharan/ Research Interests Machine Learning, Statistical Learning
More informationMotivation. Motivation. Can a software agent learn to play Backgammon by itself? Machine Learning. Reinforcement Learning
Motivation Machine Learning Can a software agent learn to play Backgammon by itself? Reinforcement Learning Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut
More informationProximal mapping via network optimization
L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
More informationAdaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
More informationChannel Allocation in Cellular Telephone. Systems. Lab. for Info. and Decision Sciences. Cambridge, MA 02139. bertsekas@lids.mit.edu.
Reinforcement Learning for Dynamic Channel Allocation in Cellular Telephone Systems Satinder Singh Department of Computer Science University of Colorado Boulder, CO 80309-0430 baveja@cs.colorado.edu Dimitri
More informationCURRICULUM VITÆ. Education. Latest update : August 2015. Sanjay Ramassamy Graduate student at Brown University.
CURRICULUM VITÆ Latest update : August 2015 Sanjay Ramassamy Graduate student at Brown University. Address : Sanjay Ramassamy Box 1917 151 Thayer Street Providence, RI 02912 USA Webpage : http://www.normalesup.org/~ramassamy/
More informationOnline Learning for Matrix Factorization and Sparse Coding
Journal of Machine Learning Research 11 (2010) 19-60 Submitted 7/09; Revised 11/09; Published 1/10 Online Learning for Matrix Factorization and Sparse Coding Julien Mairal Francis Bach INRIA - WILLOW Project-Team
More informationMachine Learning Big Data using Map Reduce
Machine Learning Big Data using Map Reduce By Michael Bowles, PhD Where Does Big Data Come From? -Web data (web logs, click histories) -e-commerce applications (purchase histories) -Retail purchase histories
More informationPreliminary Syllabus for the course of Data Science for Business Analytics
Preliminary Syllabus for the course of Data Science for Business Analytics Miguel Godinho de Matos 1,2 and Pedro Ferreira 2,3 1 Cato_lica-Lisbon, School of Business and Economics 2 Heinz College, Carnegie
More informationLarge-scale machine learning and convex optimization
Large-scale machine learning and convex optimization Francis Bach INRIA - Ecole Normale Supérieure, Paris, France Eurandom - March 2014 Slides available at www.di.ens.fr/~fbach/gradsto_eurandom.pdf Context
More informationScalable Machine Learning - or what to do with all that Big Data infrastructure
- or what to do with all that Big Data infrastructure TU Berlin blog.mikiobraun.de Strata+Hadoop World London, 2015 1 Complex Data Analysis at Scale Click-through prediction Personalized Spam Detection
More informationMonotone multi-armed bandit allocations
JMLR: Workshop and Conference Proceedings 19 (2011) 829 833 24th Annual Conference on Learning Theory Monotone multi-armed bandit allocations Aleksandrs Slivkins Microsoft Research Silicon Valley, Mountain
More informationA Sarsa based Autonomous Stock Trading Agent
A Sarsa based Autonomous Stock Trading Agent Achal Augustine The University of Texas at Austin Department of Computer Science Austin, TX 78712 USA achal@cs.utexas.edu Abstract This paper describes an autonomous
More informationBIOINF 585 Fall 2015 Machine Learning for Systems Biology & Clinical Informatics http://www.ccmb.med.umich.edu/node/1376
Course Director: Dr. Kayvan Najarian (DCM&B, kayvan@umich.edu) Lectures: Labs: Mondays and Wednesdays 9:00 AM -10:30 AM Rm. 2065 Palmer Commons Bldg. Wednesdays 10:30 AM 11:30 AM (alternate weeks) Rm.
More informationParallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data Jun Wang Department of Mechanical and Automation Engineering The Chinese University of Hong Kong Shatin, New Territories,
More informationClustering Connectionist and Statistical Language Processing
Clustering Connectionist and Statistical Language Processing Frank Keller keller@coli.uni-sb.de Computerlinguistik Universität des Saarlandes Clustering p.1/21 Overview clustering vs. classification supervised
More informationAn Environment Model for N onstationary Reinforcement Learning
An Environment Model for N onstationary Reinforcement Learning Samuel P. M. Choi Dit-Yan Yeung Nevin L. Zhang pmchoi~cs.ust.hk dyyeung~cs.ust.hk lzhang~cs.ust.hk Department of Computer Science, Hong Kong
More informationMaster of Science in Artificial Intelligence
Master of Science in Artificial Intelligence Options: Engineering and Computer Science (ECS) Speech and Language Technology (SLT) Big Data Analytics (BDA) Faculty of Engineering Science Faculty of Science
More informationBig Data Optimization at SAS
Big Data Optimization at SAS Imre Pólik et al. SAS Institute Cary, NC, USA Edinburgh, 2013 Outline 1 Optimization at SAS 2 Big Data Optimization at SAS The SAS HPA architecture Support vector machines
More informationResearch units. Laboratory of Microbial Molecular Genetics LMGM
Research units HCERES report on research unit: Laboratory of Microbial Molecular Genetics LMGM Under the supervision of the following institutions and research bodies: Université Toulouse 3 - Paul Sabatier
More informationNeuro-Dynamic Programming An Overview
1 Neuro-Dynamic Programming An Overview Dimitri Bertsekas Dept. of Electrical Engineering and Computer Science M.I.T. September 2006 2 BELLMAN AND THE DUAL CURSES Dynamic Programming (DP) is very broadly
More information9700 South Cass Avenue, Lemont, IL 60439 URL: www.mcs.anl.gov/ fulin
Fu Lin Contact information Education Work experience Research interests Mathematics and Computer Science Division Phone: (630) 252-0973 Argonne National Laboratory E-mail: fulin@mcs.anl.gov 9700 South
More informationFoundations of Machine Learning On-Line Learning. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu
Foundations of Machine Learning On-Line Learning Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Motivation PAC learning: distribution fixed over time (training and test). IID assumption.
More informationNEURAL NETWORKS AND REINFORCEMENT LEARNING. Abhijit Gosavi
NEURAL NETWORKS AND REINFORCEMENT LEARNING Abhijit Gosavi Department of Engineering Management and Systems Engineering Missouri University of Science and Technology Rolla, MO 65409 1 Outline A Quick Introduction
More informationEfficient Projections onto the l 1 -Ball for Learning in High Dimensions
Efficient Projections onto the l 1 -Ball for Learning in High Dimensions John Duchi Google, Mountain View, CA 94043 Shai Shalev-Shwartz Toyota Technological Institute, Chicago, IL, 60637 Yoram Singer Tushar
More informationSURVEY REPORT DATA SCIENCE SOCIETY 2014
SURVEY REPORT DATA SCIENCE SOCIETY 2014 TABLE OF CONTENTS Contents About the Initiative 1 Report Summary 2 Participants Info 3 Participants Expertise 6 Suggested Discussion Topics 7 Selected Responses
More informationLaboratoire d Informatique de Paris Nord, Institut Galilée, Université. 99 avenue Jean-Baptiste Clément, 93430 Villetaneuse, France.
Domenico Ruoppolo CV Personal Information First Name Domenico. Last Name Ruoppolo. Date of Birth December 16th, 1985. Place of Birth Naples, Italy. Nationality Italian. Location Address Office B311. Contacts
More informationSparse Online Learning via Truncated Gradient
Sparse Online Learning via Truncated Gradient John Langford Yahoo! Research jl@yahoo-inc.com Lihong Li Department of Computer Science Rutgers University lihong@cs.rutgers.edu Tong Zhang Department of Statistics
More informationHow To Get A Computer Science Degree At Appalachian State
118 Master of Science in Computer Science Department of Computer Science College of Arts and Sciences James T. Wilkes, Chair and Professor Ph.D., Duke University WilkesJT@appstate.edu http://www.cs.appstate.edu/
More informationDistributed Machine Learning and Big Data
Distributed Machine Learning and Big Data Sourangshu Bhattacharya Dept. of Computer Science and Engineering, IIT Kharagpur. http://cse.iitkgp.ac.in/~sourangshu/ August 21, 2015 Sourangshu Bhattacharya
More informationCurriculum vitae. July 2007 present Professor of Mathematics (W3), Technische
Peter Bank Institut für Mathematik, Sekr. MA 7-1 Straße des 17. Juni 136 10623 Berlin Germany Tel.: +49 (30) 314-22816 Fax.: +49 (30) 314-24413 e-mail: bank@math.tu-berlin.edu URL: www.math.tu-berlin.de/
More informationMachine Learning. CUNY Graduate Center, Spring 2013. Professor Liang Huang. huang@cs.qc.cuny.edu
Machine Learning CUNY Graduate Center, Spring 2013 Professor Liang Huang huang@cs.qc.cuny.edu http://acl.cs.qc.edu/~lhuang/teaching/machine-learning Logistics Lectures M 9:30-11:30 am Room 4419 Personnel
More informationIntroduction to Online Learning Theory
Introduction to Online Learning Theory Wojciech Kot lowski Institute of Computing Science, Poznań University of Technology IDSS, 04.06.2013 1 / 53 Outline 1 Example: Online (Stochastic) Gradient Descent
More informationEIT ICT Labs MASTER SCHOOL DSS Programme Specialisations
EIT ICT Labs MASTER SCHOOL DSS Programme Specialisations DSS EIT ICT Labs Master Programme Distributed System and Services (Cloud Computing) The programme in Distributed Systems and Services focuses on
More informationlarge-scale machine learning revisited Léon Bottou Microsoft Research (NYC)
large-scale machine learning revisited Léon Bottou Microsoft Research (NYC) 1 three frequent ideas in machine learning. independent and identically distributed data This experimental paradigm has driven
More informationLecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
More informationPerformance evaluation of Web Information Retrieval Systems and its application to e-business
Performance evaluation of Web Information Retrieval Systems and its application to e-business Fidel Cacheda, Angel Viña Departament of Information and Comunications Technologies Facultad de Informática,
More informationSteven C.H. Hoi. School of Computer Engineering Nanyang Technological University Singapore
Steven C.H. Hoi School of Computer Engineering Nanyang Technological University Singapore Acknowledgments: Peilin Zhao, Jialei Wang, Hao Xia, Jing Lu, Rong Jin, Pengcheng Wu, Dayong Wang, etc. 2 Agenda
More informationPh.D., Particle Physics Theory Thesis title: FCNC Processes of B and K Mesons from Lattice QCD University of Edinburgh October 1995 July1999
Curriculum Vitae Date of Birth: 8 th April 1971 Nationality: Place of Birth: Work Address: Taiwanese Taipei City Institute for Physics National Chiao-Tung University Hsinchu 300 Taiwan Work Telephone:
More informationLarge Linear Classification When Data Cannot Fit In Memory
Large Linear Classification When Data Cannot Fit In Memory ABSTRACT Hsiang-Fu Yu Dept. of Computer Science National Taiwan University Taipei 106, Taiwan b93107@csie.ntu.edu.tw Kai-Wei Chang Dept. of Computer
More informationBeyond Reward: The Problem of Knowledge and Data
Beyond Reward: The Problem of Knowledge and Data Richard S. Sutton University of Alberta Edmonton, Alberta, Canada Intelligence can be defined, informally, as knowing a lot and being able to use that knowledge
More informationLearning the Structure of Factored Markov Decision Processes in Reinforcement Learning Problems
Learning the Structure of Factored Markov Decision Processes in Reinforcement Learning Problems Thomas Degris Thomas.Degris@lip6.fr Olivier Sigaud Olivier.Sigaud@lip6.fr Pierre-Henri Wuillemin Pierre-Henri.Wuillemin@lip6.fr
More informationDr V. J. Brown. Neuroscience (see Biomedical Sciences) History, Philosophy, Social Anthropology, Theological Studies.
Psychology - pathways & 1000 Level modules School of Psychology Head of School Degree Programmes Single Honours Degree: Joint Honours Degrees: Dr V. J. Brown Psychology Neuroscience (see Biomedical Sciences)
More informationOnline Lazy Updates for Portfolio Selection with Transaction Costs
Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence Online Lazy Updates for Portfolio Selection with Transaction Costs Puja Das, Nicholas Johnson, and Arindam Banerjee Department
More informationThe Science Education System in Oxford and the UK. David Andrews
The Science Education System in Oxford and the UK David Andrews 1 Outline of this talk High school education (brief) Undergraduate education Oxford Physics Other sciences at Oxford Other universities Postgraduate
More informationSupport Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano
More informationhttp://www.duke.edu/~aa249/ The Fuqua School of Business
Alessandro Arlotto Duke University http://www.duke.edu/~aa249/ The Fuqua School of Business alessandro.arlotto@duke.edu 100 Fuqua Drive, Office W311 Phone: (919) 660-7780 PO Box 90120 Fax: (919) 681-6246
More informationStatistical Inference, Learning and Models for Big Data
Statistical Inference, Learning and Models for Big Data Nancy Reid University of Toronto P.R. Krishnaiah Memorial Lecture 2015 Rao Prize Conference Penn State University May 15, 2015 P. R. Krishnaiah 1932
More informationBayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University caizhua@gmail.com
Bayesian Machine Learning (ML): Modeling And Inference in Big Data Zhuhua Cai Google Rice University caizhua@gmail.com 1 Syllabus Bayesian ML Concepts (Today) Bayesian ML on MapReduce (Next morning) Bayesian
More informationFeature Selection with Monte-Carlo Tree Search
Feature Selection with Monte-Carlo Tree Search Robert Pinsler 20.01.2015 20.01.2015 Fachbereich Informatik DKE: Seminar zu maschinellem Lernen Robert Pinsler 1 Agenda 1 Feature Selection 2 Feature Selection
More informationNumerical Analysis. Professor Donna Calhoun. Fall 2013 Math 465/565. Office : MG241A Office Hours : Wednesday 10:00-12:00 and 1:00-3:00
Numerical Analysis Professor Donna Calhoun Office : MG241A Office Hours : Wednesday 10:00-12:00 and 1:00-3:00 Fall 2013 Math 465/565 http://math.boisestate.edu/~calhoun/teaching/math565_fall2013 What is
More information10. Proximal point method
L. Vandenberghe EE236C Spring 2013-14) 10. Proximal point method proximal point method augmented Lagrangian method Moreau-Yosida smoothing 10-1 Proximal point method a conceptual algorithm for minimizing
More informationTeaching in School of Electronic, Information and Electrical Engineering
Introduction to Teaching in School of Electronic, Information and Electrical Engineering Shanghai Jiao Tong University Outline Organization of SEIEE Faculty Enrollments Undergraduate Programs Sample Curricula
More informationSimon Lacoste-Julien (December 2012 version) Curriculum Vitae
Simon Lacoste-Julien (December 2012 version) Curriculum Vitae Contact Information SIERRA project-team INRIA 23, avenue d Italie Paris, 75013, France Email: see website Web: http://www.di.ens.fr/ slacoste
More informationScheduling Software Projects to Minimize the Development Time and Cost with a Given Staff
Scheduling Software Projects to Minimize the Development Time and Cost with a Given Staff Frank Padberg Fakultät für Informatik Universität Karlsruhe, Germany padberg @ira.uka.de Abstract A probabilistic
More informationLearning outcomes. Knowledge and understanding. Competence and skills
Syllabus Master s Programme in Statistics and Data Mining 120 ECTS Credits Aim The rapid growth of databases provides scientists and business people with vast new resources. This programme meets the challenges
More informationFRANCESCO BELLOCCHIO S CURRICULUM VITAE ET STUDIORUM
FRANCESCO BELLOCCHIO S CURRICULUM VITAE ET STUDIORUM April 2011 Index Personal details and education 1 Research activities 2 Teaching and tutorial activities 3 Conference organization and review activities
More informationCSci 538 Articial Intelligence (Machine Learning and Data Analysis)
CSci 538 Articial Intelligence (Machine Learning and Data Analysis) Course Syllabus Fall 2015 Instructor Derek Harter, Ph.D., Associate Professor Department of Computer Science Texas A&M University - Commerce
More informationEDEN ADVANCED STRATEGIC MANAGEMENT DOCTORAL SEMINAR. When: October 19-23, 2015 Where: INSEAD Europe campus, Fontainebleau, France
EDEN ADVANCED STRATEGIC MANAGEMENT DOCTORAL SEMINAR When: October 19-23, 2015 Where: INSEAD Europe campus, Fontainebleau, France The EDEN Advanced Strategic Management Doctoral Seminar is co-organized
More informationAcademic Calendar for Faculty
Summer 2013 Term June 3, 2013 (Monday) June 3-4, 2013 (Monday Tuesday) June 5, 2013 (Wednesday) June 5-6, 2013 (Wednesday Thursday) June 6, 2013 (Thursday) July 3, 2013 (Wednesday) July 4, 2013 (Thursday)
More informationReinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Martin Lauer AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg martin.lauer@kit.edu
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationCurriculum Vitae. Rohini Kumar
Curriculum Vitae Rohini Kumar Contact Information University Address: Mathematics, 1150 Faculty/Administration Building 656 W. Kirby Detroit, MI 48202 Phone: 313.577.2479 email: rkumar@math.wayne.edu Personal
More informationMaster of Artificial Intelligence
Faculty of Engineering Faculty of Science Master of Artificial Intelligence Options: Engineering and Computer Science (ECS) Speech and Language Technology (SLT) Cognitive Science (CS) K.U.Leuven Masters.
More informationPlease consult the Department of Engineering about the Computer Engineering Emphasis.
COMPUTER SCIENCE Computer science is a dynamically growing discipline. ABOUT THE PROGRAM The Department of Computer Science is committed to providing students with a program that includes the basic fundamentals
More informationB. Delyon. Stochastic approximation with decreasing gain: convergence and asymptotic theory. Unpublished lecture notes, Université de Rennes, 2000.
Some References Laetitia Andrieu, Guy Cohen and Felisa J. Vázquez-Abad. Gradient-based simulation optimization under probability constraints. European Journal of Operational Research, 212, 345-351, 2011.
More informationAn Information Retrieval using weighted Index Terms in Natural Language document collections
Internet and Information Technology in Modern Organizations: Challenges & Answers 635 An Information Retrieval using weighted Index Terms in Natural Language document collections Ahmed A. A. Radwan, Minia
More informationContemporary Techniques for Data Mining Social Media
Contemporary Techniques for Data Mining Social Media Stephen Cutting (100063482) 1 Introduction Social media websites such as Facebook, Twitter and Google+ allow millions of users to communicate with one
More informationOnline Optimization and Personalization of Teaching Sequences
Online Optimization and Personalization of Teaching Sequences Benjamin Clément 1, Didier Roy 1, Manuel Lopes 1, Pierre-Yves Oudeyer 1 1 Flowers Lab Inria Bordeaux Sud-Ouest, Bordeaux 33400, France, didier.roy@inria.fr
More information