Sparse Coding: An Overview
|
|
- Domenic Haynes
- 7 years ago
- Views:
Transcription
1 Sparse Coding: An Overview Brian Booth SFU Machine Learning Reading Group November 12, 2013
2 The aim of sparse coding
3 The aim of sparse coding Every column of D is a prototype
4 The aim of sparse coding Every column of D is a prototype Similar to, but more general than, PCA
5 Example: Sparse Coding of Images
6 Sparse Coding in V1
7 Introduction The Basics Example: Image Denoising Adding Prior Knowledge Conclusions
8 Example: Image Restoration
9 Sparse Coding and Acoustics
10 Sparse Coding and Acoustics Inner ear (cochlea) also does sparse coding of frequencies
11 Sparse Coding and Natural Language Processing
12 Outline 1 Introduction: Why Sparse Coding?
13 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge
14 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
15 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
16 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
17 The aim of sparse coding, revisited We assume our data x satisfies x n α i d i = αd i=1
18 The aim of sparse coding, revisited We assume our data x satisfies x n α i d i = αd i=1 Learning: Given training data x j, j {1,, m} Learn dictionary D and sparse code α
19 The aim of sparse coding, revisited We assume our data x satisfies x n α i d i = αd i=1 Learning: Given training data x j, j {1,, m} Learn dictionary D and sparse code α Encoding: Given test data x, dictionary D Learn sparse code α
20 Learning: The Objective Function Dictionary learning involves optimizing: arg min {d i },{α j } j=1 m x j n α j i d i 2 i=1
21 Learning: The Objective Function Dictionary learning involves optimizing: arg min {d i },{α j } j=1 m x j n m n α j i d i 2 + β α j i i=1 j=1 i=1
22 Learning: The Objective Function Dictionary learning involves optimizing: arg min {d i },{α j } j=1 m x j n m n α j i d i 2 + β α j i i=1 j=1 i=1 subject to d i 2 c, i = 1,, n.
23 Learning: The Objective Function Dictionary learning involves optimizing: arg min In matrix notation: {d i },{α j } j=1 m x j n α j i d i 2 + β i=1 subject to d i 2 c, i = 1,, n. arg min D,A X AD 2 F + β i,j m j=1 i=1 α i,j n α j i subject to i D 2 i,j c, i = 1,, n.
24 Learning: The Objective Function Dictionary learning involves optimizing: arg min In matrix notation: {d i },{α j } j=1 m x j n α j i d i 2 + β i=1 subject to d i 2 c, i = 1,, n. arg min D,A X AD 2 F + β i,j m j=1 i=1 α i,j n α j i subject to i D 2 i,j c, i = 1,, n. Split the optimization over D and A in two.
25 Step 1: Learning the Dictionary Reduced optimization problem: arg min D X AD 2 F subject to i D 2 i,j c, i = 1,, n.
26 Step 1: Learning the Dictionary Reduced optimization problem: arg min D X AD 2 F subject to i D 2 i,j c, i = 1,, n. Introduce Lagrange multipliers: ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D i,j c )
27 Step 1: Learning the Dictionary Reduced optimization problem: arg min D X AD 2 F subject to i D 2 i,j c, i = 1,, n. Introduce Lagrange multipliers: ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D i,j c ) where each λ j 0 is a dual variable...
28 Step 1: Moving to the dual From the Lagrangian ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D 2 i,j c )
29 Step 1: Moving to the dual From the Lagrangian ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D 2 i,j c ) minimize over D to obtain Lagrange dual D (λ) = min D L (D, λ) =
30 Step 1: Moving to the dual From the Lagrangian ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D 2 i,j c ) minimize over D to obtain Lagrange dual ( ( ) 1 D (λ) = min L (D, λ) = tr X T X XA T AA T + Λ (XA ) ) T T cλ D
31 Step 1: Moving to the dual From the Lagrangian ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D 2 i,j c ) minimize over D to obtain Lagrange dual ( ( ) 1 D (λ) = min L (D, λ) = tr X T X XA T AA T + Λ (XA ) ) T T cλ D The dual can be optimized using congugate gradient
32 Step 1: Moving to the dual From the Lagrangian ( ) L (D, λ) = tr (X AD) T (X AD) + n j=1 λ j ( i D 2 i,j c ) minimize over D to obtain Lagrange dual ( ( ) 1 D (λ) = min L (D, λ) = tr X T X XA T AA T + Λ (XA ) ) T T cλ D The dual can be optimized using congugate gradient Only n, λ values compared to D being n k
33 Step 1: Dual to the Dictionary With the optimal Λ, our dictionary is D T = ( ) 1 AA T + Λ (XA ) T T
34 Step 1: Dual to the Dictionary With the optimal Λ, our dictionary is D T = ( ) 1 AA T + Λ (XA ) T T Key point: Moving to the dual reduces the number of optimization variables, speeding up the optimization.
35 Step 2: Learning the Sparse Code With D now fixed, optimize for A arg min A X AD 2 F + β i,j α i,j
36 Step 2: Learning the Sparse Code With D now fixed, optimize for A arg min A X AD 2 F + β i,j α i,j Unconstrained, convex quadratic optimization
37 Step 2: Learning the Sparse Code With D now fixed, optimize for A arg min A X AD 2 F + β i,j α i,j Unconstrained, convex quadratic optimization Many solvers for this (e.g. interior point methods, in-crowd algorithm, fixed-point continuation)
38 Step 2: Learning the Sparse Code With D now fixed, optimize for A arg min A X AD 2 F + β i,j α i,j Note: Unconstrained, convex quadratic optimization Many solvers for this (e.g. interior point methods, in-crowd algorithm, fixed-point continuation) Same problem as the encoding problem. Runtime of optimization in the encoding stage?
39 Speeding up the testing phase Fair amount of work on speeding up the encoding stage: H. Lee et al., Efficient sparse coding algorithms nips06-sparsecoding.pdf K. Gregor and Y. LeCun, Learning Fast Approximations of Sparse Coding gregor-icml-10.pdf S. Hawe et al., Separable Dictionary Learning
40 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
41 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
42 Relationships between Dictionary atoms Dictionaries are over-complete bases
43 Relationships between Dictionary atoms Dictionaries are over-complete bases Dictate relationships between atoms
44 Relationships between Dictionary atoms Dictionaries are over-complete bases Dictate relationships between atoms Example: Hierarchical dictionaries
45 Example: Image Patches
46 Example: Document Topics
47 Problem Statement Goal: Have sub-groups of sparse code α all be non-zero (or zero).
48 Problem Statement Goal: Have sub-groups of sparse code α all be non-zero (or zero). Hierarchical: If a node is non-zero, it s parent must be non-zero If a node s parent is zero, the node must be zero
49 Problem Statement Goal: Have sub-groups of sparse code α all be non-zero (or zero). Hierarchical: If a node is non-zero, it s parent must be non-zero If a node s parent is zero, the node must be zero Implementation: Change the regularization Enforce sparsity differently...
50 Grouping Code Entries Level k included in k + 1 groups
51 Grouping Code Entries Level k included in k + 1 groups Add α i to objective function once for each group
52 Group Regularization Updated objective function: arg min D,{α j } m ] [ x j Dα j 2 j=1
53 Group Regularization Updated objective function: arg min D,{α j } m j=1 [ x j Dα j 2 + βω (α j) ]
54 Group Regularization Updated objective function: arg min D,{α j } m j=1 [ x j Dα j 2 + βω (α j) ] where Ω (α) = g P w g α g
55 Group Regularization Updated objective function: arg min D,{α j } m j=1 [ x j Dα j 2 + βω (α j) ] where Ω (α) = g P w g α g α g are the code values for group g.
56 Group Regularization Updated objective function: arg min D,{α j } m j=1 [ x j Dα j 2 + βω (α j) ] where Ω (α) = g P w g α g α g are the code values for group g. w g weights the enforcement of the hierarchy
57 Group Regularization Updated objective function: arg min D,{α j } m j=1 [ x j Dα j 2 + βω (α j) ] where Ω (α) = g P w g α g α g are the code values for group g. w g weights the enforcement of the hierarchy Solve using proximal methods.
58 Other Examples Other examples of structured sparsity: M. Stojnic et al., On the Reconstruction of Block-Sparse Signals With an Optimal Number of Measurements, J. Mairal et al., Convex and Network Flow Optimization for Structured Sparsity, volume12/mairal11a/mairal11a.pdf
59 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
60 Outline 1 Introduction: Why Sparse Coding? 2 Sparse Coding: The Basics 3 Adding Prior Knowledge 4 Conclusions
61 Summary
62 Summary Two interesting directions:
63 Summary Two interesting directions: Increasing speed of the testing phase
64 Summary Two interesting directions: Increasing speed of the testing phase Optimizing dictionary structure
Big learning: challenges and opportunities
Big learning: challenges and opportunities Francis Bach SIERRA Project-team, INRIA - Ecole Normale Supérieure December 2013 Omnipresent digital media Scientific context Big data Multimedia, sensors, indicators,
More informationMachine learning challenges for big data
Machine learning challenges for big data Francis Bach SIERRA Project-team, INRIA - Ecole Normale Supérieure Joint work with R. Jenatton, J. Mairal, G. Obozinski, N. Le Roux, M. Schmidt - December 2012
More information6. Cholesky factorization
6. Cholesky factorization EE103 (Fall 2011-12) triangular matrices forward and backward substitution the Cholesky factorization solving Ax = b with A positive definite inverse of a positive definite matrix
More informationClass-specific Sparse Coding for Learning of Object Representations
Class-specific Sparse Coding for Learning of Object Representations Stephan Hasler, Heiko Wersing, and Edgar Körner Honda Research Institute Europe GmbH Carl-Legien-Str. 30, 63073 Offenbach am Main, Germany
More informationNon-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning SAMSI 10 May 2013 Outline Introduction to NMF Applications Motivations NMF as a middle step
More informationGroup Sparse Coding. Fernando Pereira Google Mountain View, CA pereira@google.com. Dennis Strelow Google Mountain View, CA strelow@google.
Group Sparse Coding Samy Bengio Google Mountain View, CA bengio@google.com Fernando Pereira Google Mountain View, CA pereira@google.com Yoram Singer Google Mountain View, CA singer@google.com Dennis Strelow
More informationProximal mapping via network optimization
L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
More informationReprésentations parcimonieuses pour le traitement d image et la vision artificielle
Représentations parcimonieuses pour le traitement d image et la vision artificielle Julien Mairal Department of Statistics, University of California, Berkeley INRIA - projet Willow, Paris Journées ORASIS,
More informationBig Data Techniques Applied to Very Short-term Wind Power Forecasting
Big Data Techniques Applied to Very Short-term Wind Power Forecasting Ricardo Bessa Senior Researcher (ricardo.j.bessa@inesctec.pt) Center for Power and Energy Systems, INESC TEC, Portugal Joint work with
More informationA Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Large-margin linear
More informationC variance Matrices and Computer Network Analysis
On Covariance Structure in Noisy, Big Data Randy C. Paffenroth a, Ryan Nong a and Philip C. Du Toit a a Numerica Corporation, Loveland, CO, USA; ABSTRACT Herein we describe theory and algorithms for detecting
More informationSeveral Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
More informationPart II Redundant Dictionaries and Pursuit Algorithms
Aisenstadt Chair Course CRM September 2009 Part II Redundant Dictionaries and Pursuit Algorithms Stéphane Mallat Centre de Mathématiques Appliquées Ecole Polytechnique Sparsity in Redundant Dictionaries
More informationAn Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.
An Overview Of Software For Convex Optimization Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.edu In fact, the great watershed in optimization isn t between linearity
More informationOnline Learning for Matrix Factorization and Sparse Coding
Journal of Machine Learning Research 11 (2010) 19-60 Submitted 7/09; Revised 11/09; Published 1/10 Online Learning for Matrix Factorization and Sparse Coding Julien Mairal Francis Bach INRIA - WILLOW Project-Team
More informationLarge-Scale Similarity and Distance Metric Learning
Large-Scale Similarity and Distance Metric Learning Aurélien Bellet Télécom ParisTech Joint work with K. Liu, Y. Shi and F. Sha (USC), S. Clémençon and I. Colin (Télécom ParisTech) Séminaire Criteo March
More informationLecture Topic: Low-Rank Approximations
Lecture Topic: Low-Rank Approximations Low-Rank Approximations We have seen principal component analysis. The extraction of the first principle eigenvalue could be seen as an approximation of the original
More informationBig Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
More informationSupport Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationOptimization Modeling for Mining Engineers
Optimization Modeling for Mining Engineers Alexandra M. Newman Division of Economics and Business Slide 1 Colorado School of Mines Seminar Outline Linear Programming Integer Linear Programming Slide 2
More informationSupport Vector Machines
Support Vector Machines Charlie Frogner 1 MIT 2011 1 Slides mostly stolen from Ryan Rifkin (Google). Plan Regularization derivation of SVMs. Analyzing the SVM problem: optimization, duality. Geometric
More informationSupport Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
More informationMatrix Multiplication
Matrix Multiplication CPS343 Parallel and High Performance Computing Spring 2016 CPS343 (Parallel and HPC) Matrix Multiplication Spring 2016 1 / 32 Outline 1 Matrix operations Importance Dense and sparse
More informationEfficient online learning of a non-negative sparse autoencoder
and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil
More informationReduced echelon form: Add the following conditions to conditions 1, 2, and 3 above:
Section 1.2: Row Reduction and Echelon Forms Echelon form (or row echelon form): 1. All nonzero rows are above any rows of all zeros. 2. Each leading entry (i.e. left most nonzero entry) of a row is in
More informationA REVIEW OF FAST l 1 -MINIMIZATION ALGORITHMS FOR ROBUST FACE RECOGNITION
A REVIEW OF FAST l 1 -MINIMIZATION ALGORITHMS FOR ROBUST FACE RECOGNITION ALLEN Y. YANG, ARVIND GANESH, ZIHAN ZHOU, S. SHANKAR SASTRY, AND YI MA Abstract. l 1 -minimization refers to finding the minimum
More informationNumerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen
(für Informatiker) M. Grepl J. Berger & J.T. Frings Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2010/11 Problem Statement Unconstrained Optimality Conditions Constrained
More informationImage Super-Resolution via Sparse Representation
1 Image Super-Resolution via Sparse Representation Jianchao Yang, Student Member, IEEE, John Wright, Student Member, IEEE Thomas Huang, Life Fellow, IEEE and Yi Ma, Senior Member, IEEE Abstract This paper
More informationBag of Pursuits and Neural Gas for Improved Sparse Coding
Bag of Pursuits and Neural Gas for Improved Sparse Coding Kai Labusch, Erhardt Barth, and Thomas Martinetz University of Lübec Institute for Neuro- and Bioinformatics Ratzeburger Allee 6 23562 Lübec, Germany
More informationNumerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems
Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course G63.2010.001 / G22.2420-001,
More informationHow To Teach A Dictionary To Learn For Big Data Via Atom Update
IEEE COMPUTING IN SCIENCE AND ENGINEERING, VOL. XX, NO. XX, XXX 2014 1 IK-SVD: Dictionary Learning for Spatial Big Data via Incremental Atom Update Lizhe Wang 1,KeLu 2, Peng Liu 1, Rajiv Ranjan 3, Lajiao
More informationCSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
More informationAbstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication).
MAT 2 (Badger, Spring 202) LU Factorization Selected Notes September 2, 202 Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix
More information2) Based on the information in the table which choice BEST shows the answer to 1 906? 906 899 904 909
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ) Multiplying a number by results in what type of. even. 0. even.,0. odd..,0. even ) Based on the information in the table which choice BEST shows the answer to 0? 0 0 0 )
More informationVariational approach to restore point-like and curve-like singularities in imaging
Variational approach to restore point-like and curve-like singularities in imaging Daniele Graziani joint work with Gilles Aubert and Laure Blanc-Féraud Roma 12/06/2012 Daniele Graziani (Roma) 12/06/2012
More informationMatrix-vector multiplication in terms of dot-products
Matrix-vector multiplication in terms of dot-products Let M be an R C matrix. Dot-Product Definition of matrix-vector multiplication: M u is the R-vector v such that v[r] is the dot-product of row r of
More informationBig Data Analytics. Lucas Rego Drumond
Big Data Analytics Lucas Rego Drumond Information Systems and Machine Learning Lab (ISMLL) Institute of Computer Science University of Hildesheim, Germany Going For Large Scale Going For Large Scale 1
More informationStatistical machine learning, high dimension and big data
Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,
More informationIEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, 2013. ACCEPTED FOR PUBLICATION 1
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, 2013. ACCEPTED FOR PUBLICATION 1 Active-Set Newton Algorithm for Overcomplete Non-Negative Representations of Audio Tuomas Virtanen, Member,
More informationA numerically adaptive implementation of the simplex method
A numerically adaptive implementation of the simplex method József Smidla, Péter Tar, István Maros Department of Computer Science and Systems Technology University of Pannonia 17th of December 2014. 1
More informationThe Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem
More informationGraph Based Processing of Big Images
Graph Based Processing of Big Images HUI HAN CHIN DSO NATIONAL LABORATORIES, SINGAPORE 31 OCT 2014 Introduction Member of Technical Staff at Signal Processing Laboratory, Sensors Division, DSO National
More informationDantzig-Wolfe bound and Dantzig-Wolfe cookbook
Dantzig-Wolfe bound and Dantzig-Wolfe cookbook thst@man.dtu.dk DTU-Management Technical University of Denmark 1 Outline LP strength of the Dantzig-Wolfe The exercise from last week... The Dantzig-Wolfe
More information10. Proximal point method
L. Vandenberghe EE236C Spring 2013-14) 10. Proximal point method proximal point method augmented Lagrangian method Moreau-Yosida smoothing 10-1 Proximal point method a conceptual algorithm for minimizing
More informationDistributed Machine Learning and Big Data
Distributed Machine Learning and Big Data Sourangshu Bhattacharya Dept. of Computer Science and Engineering, IIT Kharagpur. http://cse.iitkgp.ac.in/~sourangshu/ August 21, 2015 Sourangshu Bhattacharya
More informationDuality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725
Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
More informationMorphological Diversity and Sparsity for Multichannel Data Restoration
Morphological Diversity and Sparsity for Multichannel Data Restoration J.Bobin 1, Y.Moudden 1, J.Fadili and J-L.Starck 1 1 jerome.bobin@cea.fr, ymoudden@cea.fr, jstarck@cea.fr - CEA-DAPNIA/SEDI, Service
More informationCouple Dictionary Training for Image Super-resolution
IEEE TRANSACTIONS ON IMAGE PROCESSING 1 Couple Dictionary Training for Image Super-resolution Jianchao Yang, Student Member, IEEE, Zhaowen Wang, Student Member, IEEE, Zhe Lin, Member, IEEE, Scott Cohen,
More information1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
More informationMathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
More informationUsing Microsoft Excel to build Efficient Frontiers via the Mean Variance Optimization Method
Using Microsoft Excel to build Efficient Frontiers via the Mean Variance Optimization Method Submitted by John Alexander McNair ID #: 0061216 Date: April 14, 2003 The Optimal Portfolio Problem Consider
More information3. Interpolation. Closing the Gaps of Discretization... Beyond Polynomials
3. Interpolation Closing the Gaps of Discretization... Beyond Polynomials Closing the Gaps of Discretization... Beyond Polynomials, December 19, 2012 1 3.3. Polynomial Splines Idea of Polynomial Splines
More information7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix
7. LU factorization EE103 (Fall 2011-12) factor-solve method LU factorization solving Ax = b with A nonsingular the inverse of a nonsingular matrix LU factorization algorithm effect of rounding error sparse
More informationApplications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
More informationSparse Signal Recovery: Theory, Applications and Algorithms
Sparse Signal Recovery: Theory, Applications and Algorithms Bhaskar Rao Department of Electrical and Computer Engineering University of California, San Diego Collaborators: I. Gorodonitsky, S. Cotter,
More informationNonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
More informationDual Methods for Total Variation-Based Image Restoration
Dual Methods for Total Variation-Based Image Restoration Jamylle Carter Institute for Mathematics and its Applications University of Minnesota, Twin Cities Ph.D. (Mathematics), UCLA, 2001 Advisor: Tony
More informationA primal-dual algorithm for group sparse regularization with overlapping groups
A primal-dual algorithm for group sparse regularization with overlapping groups Sofia Mosci DISI- Università di Genova mosci@disi.unige.it Alessandro Verri DISI- Università di Genova verri@disi.unige.it
More informationSection for Cognitive Systems DTU Informatics, Technical University of Denmark
Transformation Invariant Sparse Coding Morten Mørup & Mikkel N Schmidt Morten Mørup & Mikkel N. Schmidt Section for Cognitive Systems DTU Informatics, Technical University of Denmark Redundancy Reduction
More informationSublinear Algorithms for Big Data. Part 4: Random Topics
Sublinear Algorithms for Big Data Part 4: Random Topics Qin Zhang 1-1 2-1 Topic 1: Compressive sensing Compressive sensing The model (Candes-Romberg-Tao 04; Donoho 04) Applicaitons Medical imaging reconstruction
More informationSECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the
More informationParallel & Distributed Optimization. Based on Mark Schmidt s slides
Parallel & Distributed Optimization Based on Mark Schmidt s slides Motivation behind using parallel & Distributed optimization Performance Computational throughput have increased exponentially in linear
More informationA Lagrangian-DNN Relaxation: a Fast Method for Computing Tight Lower Bounds for a Class of Quadratic Optimization Problems
A Lagrangian-DNN Relaxation: a Fast Method for Computing Tight Lower Bounds for a Class of Quadratic Optimization Problems Sunyoung Kim, Masakazu Kojima and Kim-Chuan Toh October 2013 Abstract. We propose
More informationBig Data Analytics: Optimization and Randomization
Big Data Analytics: Optimization and Randomization Tianbao Yang, Qihang Lin, Rong Jin Tutorial@SIGKDD 2015 Sydney, Australia Department of Computer Science, The University of Iowa, IA, USA Department of
More informationEfficient variable selection in support vector machines via the alternating direction method of multipliers
Efficient variable selection in support vector machines via the alternating direction method of multipliers Gui-Bo Ye Yifei Chen Xiaohui Xie University of California, Irvine University of California, Irvine
More informationHPC Deployment of OpenFOAM in an Industrial Setting
HPC Deployment of OpenFOAM in an Industrial Setting Hrvoje Jasak h.jasak@wikki.co.uk Wikki Ltd, United Kingdom PRACE Seminar: Industrial Usage of HPC Stockholm, Sweden, 28-29 March 2011 HPC Deployment
More informationThe Many Facets of Big Data
Department of Computer Science and Engineering Hong Kong University of Science and Technology Hong Kong ACPR 2013 Big Data 1 volume sample size is big feature dimensionality is big 2 variety multiple formats:
More informationAn Introduction to Physically Based Modeling: Constrained Dynamics
An Introduction to Physically Based Modeling: Constrained Dynamics Andrew Witkin Robotics Institute Carnegie Mellon University Please note: This document is 1997 by Andrew Witkin. This chapter may be freely
More informationFactoring nonnegative matrices with linear programs
Factoring nonnegative matrices with linear programs Victor Bittorf bittorf@cs.wisc.edu Benjamin Recht brecht@cs.wisc.edu Computer Sciences University of Wisconsin Christopher Ré chrisre@cs.wisc.edu Joel
More informationMatrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.
Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that
More informationSolutions Of Some Non-Linear Programming Problems BIJAN KUMAR PATEL. Master of Science in Mathematics. Prof. ANIL KUMAR
Solutions Of Some Non-Linear Programming Problems A PROJECT REPORT submitted by BIJAN KUMAR PATEL for the partial fulfilment for the award of the degree of Master of Science in Mathematics under the supervision
More informationNumerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course G63.2010.001 / G22.2420-001, Fall 2010 September 30th, 2010 A. Donev (Courant Institute)
More informationIntroduction to Sparsity in Signal Processing 1
Introduction to Sparsity in Signal Processing Ivan Selesnick November, NYU-Poly Introduction These notes describe how sparsity can be used in several signal processing problems. A common theme throughout
More informationRECENTLY, there has been a great deal of interest in
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 47, NO. 1, JANUARY 1999 187 An Affine Scaling Methodology for Best Basis Selection Bhaskar D. Rao, Senior Member, IEEE, Kenneth Kreutz-Delgado, Senior Member,
More informationSparsity-promoting recovery from simultaneous data: a compressive sensing approach
SEG 2011 San Antonio Sparsity-promoting recovery from simultaneous data: a compressive sensing approach Haneet Wason*, Tim T. Y. Lin, and Felix J. Herrmann September 19, 2011 SLIM University of British
More informationImage Super-Resolution as Sparse Representation of Raw Image Patches
Image Super-Resolution as Sparse Representation of Raw Image Patches Jianchao Yang, John Wright, Yi Ma, Thomas Huang University of Illinois at Urbana-Champagin Beckman Institute and Coordinated Science
More informationEFFICIENT QUANTIZATION PARAMETER ESTIMATION IN HEVC BASED ON ρ-domain
EFFICIENT QUANTIZATION PARAMETER ESTIMATION IN HEVC BASED ON ρ-domain T. Biatek, M. Raulet, J.-F. Travers, O. Deforges bcom, Cesson-Sevigne, France IETR / INSA de Rennes, Rennes, France TDF, Cesson-Sevigne,
More informationInterior-Point Algorithms for Quadratic Programming
Interior-Point Algorithms for Quadratic Programming Thomas Reslow Krüth Kongens Lyngby 2008 IMM-M.Sc-2008-19 Technical University of Denmark Informatics and Mathematical Modelling Building 321, DK-2800
More informationControlling the Internet in the era of Software Defined and Virtualized Networks. Fernando Paganini Universidad ORT Uruguay
Controlling the Internet in the era of Software Defined and Virtualized Networks Fernando Paganini Universidad ORT Uruguay CDS@20, Caltech 2014 Motivation 1. The Internet grew in its first 30 years with
More informationLinear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
More informationFurther Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 J. Zhang Institute of Applied Mathematics, Chongqing University of Posts and Telecommunications, Chongqing
More informationSSketch: An Automated Framework for Streaming Sketch-based Analysis of Big Data on FPGA *
SSketch: An Automated Framework for Streaming Sketch-based Analysis of Big Data on FPGA * Bita Darvish Rouhani, Ebrahim Songhori, Azalia Mirhoseini, and Farinaz Koushanfar Department of ECE, Rice University
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationIncreasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.
1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.
More informationMassive Data Classification via Unconstrained Support Vector Machines
Massive Data Classification via Unconstrained Support Vector Machines Olvi L. Mangasarian and Michael E. Thompson Computer Sciences Department University of Wisconsin 1210 West Dayton Street Madison, WI
More informationSee All by Looking at A Few: Sparse Modeling for Finding Representative Objects
See All by Looking at A Few: Sparse Modeling for Finding Representative Objects Ehsan Elhamifar Johns Hopkins University Guillermo Sapiro University of Minnesota René Vidal Johns Hopkins University Abstract
More informationLearning A Discriminative Dictionary for Sparse Coding via Label Consistent K-SVD
Learning A Discriminative Dictionary for Sparse Coding via Label Consistent K-SVD Zhuolin Jiang, Zhe Lin, Larry S. Davis University of Maryland, College Park, MD, 7 Adobe Systems Incorporated, San Jose,
More informationSupport Vector Machines
CS229 Lecture notes Andrew Ng Part V Support Vector Machines This set of notes presents the Support Vector Machine (SVM) learning algorithm. SVMs are among the best (and many believe are indeed the best)
More informationTHE problem of computing sparse solutions (i.e., solutions
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 7, JULY 2005 2477 Sparse Solutions to Linear Inverse Problems With Multiple Measurement Vectors Shane F. Cotter, Member, IEEE, Bhaskar D. Rao, Fellow,
More informationA Distributed Line Search for Network Optimization
01 American Control Conference Fairmont Queen Elizabeth, Montréal, Canada June 7-June 9, 01 A Distributed Line Search for Networ Optimization Michael Zargham, Alejandro Ribeiro, Ali Jadbabaie Abstract
More informationFleet Assignment Using Collective Intelligence
Fleet Assignment Using Collective Intelligence Nicolas E Antoine, Stefan R Bieniawski, and Ilan M Kroo Stanford University, Stanford, CA 94305 David H Wolpert NASA Ames Research Center, Moffett Field,
More informationA HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER. Figure 1. Basic structure of an encoder.
A HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER Manoj Kumar 1 Mohammad Zubair 1 1 IBM T.J. Watson Research Center, Yorktown Hgts, NY, USA ABSTRACT The MPEG/Audio is a standard for both
More informationOnline Semi-Supervised Discriminative Dictionary Learning for Sparse Representation
Online Semi-Supervised Discriminative Dictionary Learning for Sparse Representation Guangxiao Zhang, Zhuolin Jiang, Larry S. Davis University of Maryland, College Park, MD, 2742 {gxzhang,zhuolin,lsd}@umiacs.umd.edu
More informationUNIVERSAL SPEECH MODELS FOR SPEAKER INDEPENDENT SINGLE CHANNEL SOURCE SEPARATION
UNIVERSAL SPEECH MODELS FOR SPEAKER INDEPENDENT SINGLE CHANNEL SOURCE SEPARATION Dennis L. Sun Department of Statistics Stanford University Gautham J. Mysore Adobe Research ABSTRACT Supervised and semi-supervised
More informationAnalysis of data separation and recovery problems using clustered sparsity
Analysis of data separation and recovery problems using clustered sparsity Emily J. King a, Gitta Kutyniok a, and Xiaosheng Zhuang a a Institute of Mathematics, Universität Osnabrück, Osnabrück, Germany
More informationGroup Testing a tool of protecting Network Security
Group Testing a tool of protecting Network Security Hung-Lin Fu 傅 恆 霖 Department of Applied Mathematics, National Chiao Tung University, Hsin Chu, Taiwan Group testing (General Model) Consider a set N
More informationBig Data Analytics in Future Internet of Things
Big Data Analytics in Future Internet of Things Guoru Ding, Long Wang, Qihui Wu College of Communications Engineering, PLA University of Science and Technology, Nanjing 210007, China email: dingguoru@gmail.com;
More informationDATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
More informationSolving Linear Programs in Excel
Notes for AGEC 622 Bruce McCarl Regents Professor of Agricultural Economics Texas A&M University Thanks to Michael Lau for his efforts to prepare the earlier copies of this. 1 http://ageco.tamu.edu/faculty/mccarl/622class/
More information