MS&E 310 Course Project III: First-Order Potential Reduction for Linear Programming

Similar documents
Nonlinear Programming Methods.S2 Quadratic Programming

Interior Point Methods and Linear Programming

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

Nonlinear Optimization: Algorithms 3: Interior-point methods

Summer course on Convex Optimization. Fifth Lecture Interior-Point Methods (1) Michel Baes, K.U.Leuven Bharath Rangarajan, U.

Duality in General Programs. Ryan Tibshirani Convex Optimization /36-725

2.3 Convex Constrained Optimization Problems

Mathematical finance and linear programming (optimization)

Duality of linear conic problems

LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method

Interior-Point Methods for Linear Optimization

SOLUTIONS. f x = 6x 2 6xy 24x, f y = 3x 2 6y. To find the critical points, we solve

Big Data - Lecture 1 Optimization reminders

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

4.6 Linear Programming duality

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

SMOOTHING APPROXIMATIONS FOR TWO CLASSES OF CONVEX EIGENVALUE OPTIMIZATION PROBLEMS YU QI. (B.Sc.(Hons.), BUAA)

Linear Programming I

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

Date: April 12, Contents

Linear Programming Notes V Problem Transformations

A Lagrangian-DNN Relaxation: a Fast Method for Computing Tight Lower Bounds for a Class of Quadratic Optimization Problems

International Doctoral School Algorithmic Decision Theory: MCDA and MOO

10. Proximal point method

3. Linear Programming and Polyhedral Combinatorics

Optimisation et simulation numérique.

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

An Introduction on SemiDefinite Program

constraint. Let us penalize ourselves for making the constraint too big. We end up with a

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

24. The Branch and Bound Method

Practical Guide to the Simplex Method of Linear Programming

Can linear programs solve NP-hard problems?

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

Advanced Lecture on Mathematical Science and Information Science I. Optimization in Finance

Optimization Methods in Finance

. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

An Accelerated First-Order Method for Solving SOS Relaxations of Unconstrained Polynomial Optimization Problems

CONSTRAINED NONLINEAR PROGRAMMING

1 Lecture: Integration of rational functions by decomposition

26 Linear Programming

Proximal mapping via network optimization

1 Linear Programming. 1.1 Introduction. Problem description: motivate by min-cost flow. bit of history. everything is LP. NP and conp. P breakthrough.

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Optimization Modeling for Mining Engineers

4.1 Learning algorithms for neural networks

Convex Programming Tools for Disjunctive Programs

Study Guide 2 Solutions MATH 111

CS 295 (UVM) The LMS Algorithm Fall / 23

An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM

An Introduction to Linear Programming

5.1 Bipartite Matching

Stationarity Results for Generating Set Search for Linearly Constrained Optimization

Applied Algorithm Design Lecture 5

Inner Product Spaces

Pattern Analysis. Logistic Regression. 12. Mai Joachim Hornegger. Chair of Pattern Recognition Erlangen University

Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach

CHAPTER 9. Integer Programming

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

GI01/M055 Supervised Learning Proximal Methods

Notes from Week 1: Algorithms for sequential prediction

Computing a Nearest Correlation Matrix with Factor Structure

Lecture 2: August 29. Linear Programming (part I)

Discrete Optimization

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

1 Solving LPs: The Simplex Algorithm of George Dantzig

Optimization in R n Introduction

Massive Data Classification via Unconstrained Support Vector Machines

Minimally Infeasible Set Partitioning Problems with Balanced Constraints

GENERALIZED INTEGER PROGRAMMING

Approximation Algorithms

Support Vector Machine (SVM)

Interior-Point Algorithms for Quadratic Programming

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method

Advances in Convex Optimization: Interior-point Methods, Cone Programming, and Applications

Lecture 11: 0-1 Quadratic Program and Lower Bounds

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

(Quasi-)Newton methods

The p-norm generalization of the LMS algorithm for adaptive filtering

( ) = ( ) = {,,, } β ( ), < 1 ( ) + ( ) = ( ) + ( )

The Expectation Maximization Algorithm A short tutorial

CS 688 Pattern Recognition Lecture 4. Linear Models for Classification

Online Learning and Competitive Analysis: a Unified Approach

1 Introduction. 2 Prediction with Expert Advice. Online Learning Lecture 09

Operation Research. Module 1. Module 2. Unit 1. Unit 2. Unit 3. Unit 1


The Multiplicative Weights Update method

A Distributed Line Search for Network Optimization

Chapter 4. Duality. 4.1 A Graphical Example

Homework # 3 Solutions

Similarity and Diagonalization. Similar Matrices

Two-Stage Stochastic Linear Programs

IEOR 4404 Homework #2 Intro OR: Deterministic Models February 14, 2011 Prof. Jay Sethuraman Page 1 of 5. Homework #2

1 if 1 x 0 1 if 0 x 1

Statistical Machine Learning

Cost Minimization and the Cost Function

Actually Doing It! 6. Prove that the regular unit cube (say 1cm=unit) of sufficiently high dimension can fit inside it the whole city of New York.

Transcription:

MS&E 310 Course Project III: First-Order Potential Reduction for Linear Programming Yinyu Ye November 15, 2016 1 Convex optimization over the simplex constraint We consider the following optimization problem over the simplex: Minimize Subject To e T x = n; x 0, 1 where e is the vector of all ones. Such a problem in considered in [7], where function does not need to be convex and a FPTAS algorithm was developed for computing an approximate KKT point of general quadratic programming. The following algorithm and analysis resemble those in [7]. We assume that is a convex function in x R n and fx = 0 where x is a minimizer of the problem. Furthermore, we make a standard Lipschitz assumption such that fx + d T d + γ 2 d 2, where positive γ is the Lipschitz parameter. Note that any homogeneous linear feasibility problem, e.g., the canonical Karmarkar form in [2]: Ax = 0; e T x = n; x 0. can be formulated as the model with = 1 2 Ax 2 and γ as the half of the largest eigenvalue of matrix A T A. Furthermore, any linear programming problem in the standard form and its dual Minimize c T x Subject to Ax = b; x 0; Maximize b T y Subject to A T y + s = c; s 0 1

can be represented as a homogeneous linear feasibility problem Ye et al. [5]: Ax bτ = 0; A T y s + cτ = 0; b T y c T x κ = 0; e T x + e T s + τ + κ = 2n + 2; x, s, τ, κ 0. We consider the potential function e.g., see [2, 4, 1, 6] φx = ln j lnx j, where n over the simplex. Clearly, if we start from x 0 = e, the analytic center of the simplex, and generate a sequence of points x k, k = 1,...,, whose potential value is strictly decreased, then when φx k φx 0 ln1/ɛ, we must have lnfx k lnfx 0 ln1/ɛ or This is because on the simplex lnx k j j j fx k fx 0 ɛ. lnx 0 j, k = 1,... We now describe a first order steepest descent potential reduction algorithm in the next section. 2 Steepest-Descent Potential Reduction and Complexity Analysis Note that the gradient vector of the potential function of x > 0 is φx = X 1 e. where in this note X denotes the diagonal matrix whose diagonal entries are elements of vector x. The following lemma is well known in the literature of interior-point algorithms [2, 1, 6]: Lemma 1. Let x > 0 and X 1 d β < 1. Then j lnx j + d j + j lnx j e T X 1 d + β 2 21 β. 2

Lemma 2. For any x > 0 and x x, a matrix A R m n with Ax = Ax, and a vector λ R m, consider vector px = X φx A T λ. Then, px 1. Proof. First, px = X X 1 e A T λ = X AT λ e. If any entry of AT λ is equal or less than 0, then px px 1. On the other hand, if AT λ λ T > 0, we have AT x 0. Then, from convexity and Ax = Ax, Thus, from fx = 0 Furthermore, where fx T x x = px 2 = 2 X 2 2 n X 2 2 n T AT λ x. T AT λ x x. AT λ λ T 2 2 AT x + n AT λ λ T 2 1 2 AT x + n λ AT T 2 x λ 2 AT T x + n = z2 n 2z + n = 1 n z n2, z = AT λ T x 1. The above quadratic function of z has the minimizer at z = 1 if n, so that for n + n. 1 n z n2 1 n n2 1 For any given x > 0 in the simplex and any d with e T d = 0, fx + d T d + γ 2 d 2 T d + γ 2 XX 1 d 2 T d + γ 2 X 1 d 2, 3

where the last inequality is due to X 1. Let X 1 d = β < 1 and x + = x + d = Xe + X 1 d > 0. Then, from Lemma 1 φx + φx ln 1 + T d+ γ 2 X 1 d 2 e T X 1 d + β2 21 β T d+ γ 2 X 1 d 2 e T X 1 d + β2 21 β = φx T d + γ 2 β2 + β2 21 β. The first order steepest descent potential reduction algorithm would update x by solving Minimize φx T d Subject to e T d = 0, X 1 d β; or Minimize φx T Xd Subject to e T Xd = 0, d β; where parameter β < 1 is yet to be determined. 2 where Let the scaled gradient projection vector px = I 1 x 2 XeeT X Then the minimizer of problem 2 would be and since px 1 based on Lemma 2. Thus, X φx = X λx = et X 2 φx x 2. d = β px Xpx, e λx e, φx T d = β px px 2 = β px β, φx + φx β + For β 1/2, the above quantity is less than β + Thus, one can choose β to minimize the quantity at γ β 2 2 β2 + 21 β 2 + γ β 2 /2. so that β = 1 2 + γ φx + φx 1/2 2 + 2γ. 4

One can see that the larger value of, the greater reduction of the potential function. Starting from x 0 = 1 n e, we iteratively generate xk, k = 1,..., such that φx k+1 φx k fx k 2fx k + 2γ fx k 2fx 0 + 2γ fx k 4 max{fx 0, 2γ}. The second inequality is due to fx k < fx 0 from φx k < φx 0 for all k 1 and x 0 is the analytic center of the simplex. Thus, if fxk fx 0 ɛ for 1 k K, we must have φx 0 φx K ln 1 ɛ, so that or K k=1 fx k 4 max{fx 0, 2γ} ln1 ɛ Kɛfx 0 4 max{fx 0, 2γ} ln 1 ɛ. Note that = n + n 2n. We conclude Theorem 3. The steepest descent potential reduction algorithm generates a x k with fx k /fx 0 ɛ in no more than steps. 4n + n max{1, 2n + nγ/fx 0 } ɛ ln 1 ɛ 3 Extension, Implementation and Further Analysis? Question 1: Develop a similar analysis for solving Minimize Subject To 0 x 2, 3 where we start x 0 = e, the analytic center of the BOX constraint. Question 2: Implement the algorithm and perform numerical tests to solve for = 1 2 Ax 2. Question 3: Implement the algorithm and perform numerical tests to solve for = 1 2 AAT 1/2 Ax 2, 5

and compare the performance with that in Question 2. Question 4: Test your implementation on homogeneous and self LP models for various linear programs feasible or infeasible, where you may eliminate free variables y from the formulation. References [1] C. C. Gonzaga, Polynomial affine algorithms for linear programming, Math. Programming 49 1990 7 21. [2] N. Karmarkar, A new polynomial-time algorithm for linear programming, Combinatorica 4 1984 373-395. [3] S. Mehrotra. On the implementation of a primal dual interior point method. SIAM J. Optimization, 24:575 601, 1992. [4] M. J. Todd and Y. Ye, A centered projective algorithm for linear programming, Math. Oper. Res. 15 1990 508-529. [5] Y. Ye, M. J. Todd, and S. Mizuno, An O nl - iteration homogeneous and self-dual linear programming algorithm, Math. Oper. Res. 19 1994 53 67. [6] Y. Ye, An On 3 L potential reduction algorithm for linear programming, Math. Programming 50 1991 239 258. [7] Y. Ye, On the complexity of approximating a KKT point of quadratic programming, Math. Programming 80 1998 195-211. 6