Primal-dual interior-point methods part I. Javier Peña (guest lecturer) Convex Optimization /36-725

Similar documents
Interior Point Methods and Linear Programming

Duality in General Programs. Ryan Tibshirani Convex Optimization /36-725

Nonlinear Optimization: Algorithms 3: Interior-point methods

Mathematical finance and linear programming (optimization)

4.6 Linear Programming duality

Summer course on Convex Optimization. Fifth Lecture Interior-Point Methods (1) Michel Baes, K.U.Leuven Bharath Rangarajan, U.

Nonlinear Programming Methods.S2 Quadratic Programming

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method

Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method

Linear Programming Notes V Problem Transformations

1 Solving LPs: The Simplex Algorithm of George Dantzig

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

2.3 Convex Constrained Optimization Problems

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

Lecture 2: August 29. Linear Programming (part I)

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

Duality of linear conic problems

An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Interior-Point Methods for Linear Optimization

Advanced Lecture on Mathematical Science and Information Science I. Optimization in Finance

International Doctoral School Algorithmic Decision Theory: MCDA and MOO

. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2

10. Proximal point method

Practical Guide to the Simplex Method of Linear Programming

Can linear programs solve NP-hard problems?

Optimization Methods in Finance

LECTURE: INTRO TO LINEAR PROGRAMMING AND THE SIMPLEX METHOD, KEVIN ROSS MARCH 31, 2005

Dual Methods for Total Variation-Based Image Restoration

MA4001 Engineering Mathematics 1 Lecture 10 Limits and Continuity

Some representability and duality results for convex mixed-integer programs.

Optimization Modeling for Mining Engineers

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

Lecture Notes 10

Support Vector Machine (SVM)

1 Review of Least Squares Solutions to Overdetermined Systems

Big Data - Lecture 1 Optimization reminders

Degeneracy in Linear Programming

Optimization in R n Introduction

5.1 Bipartite Matching

Chapter 4. Duality. 4.1 A Graphical Example

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: FILTER METHODS AND MERIT FUNCTIONS

This exposition of linear programming

What is Linear Programming?

Duality in Linear Programming

A Lagrangian-DNN Relaxation: a Fast Method for Computing Tight Lower Bounds for a Class of Quadratic Optimization Problems

9th Max-Planck Advanced Course on the Foundations of Computer Science (ADFOCS) Primal-Dual Algorithms for Online Optimization: Lecture 1

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

Linear Programming in Matrix Form

3. Linear Programming and Polyhedral Combinatorics

Applied Algorithm Design Lecture 5

Interior-Point Algorithms for Quadratic Programming

CHAPTER 9. Integer Programming

INTERIOR POINT POLYNOMIAL TIME METHODS IN CONVEX PROGRAMMING

An Introduction on SemiDefinite Program

Max-Min Representation of Piecewise Linear Functions

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Adaptive Online Gradient Descent

ALMOST COMMON PRIORS 1. INTRODUCTION

Tetris: A Study of Randomized Constraint Sampling

24. The Branch and Bound Method

Principles of demand management Airline yield management Determining the booking limits. » A simple problem» Stochastic gradients for general problems

Module1. x y 800.

1 Linear Programming. 1.1 Introduction. Problem description: motivate by min-cost flow. bit of history. everything is LP. NP and conp. P breakthrough.

Optimisation et simulation numérique.

Linear Programming. March 14, 2014

Roots of Polynomials

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.

Proximal mapping via network optimization

Two-Stage Stochastic Linear Programs

Integrating Benders decomposition within Constraint Programming

4.1 Learning algorithms for neural networks

Chapter Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling

The Equivalence of Linear Programs and Zero-Sum Games

Factoring & Primality

IEOR 4404 Homework #2 Intro OR: Deterministic Models February 14, 2011 Prof. Jay Sethuraman Page 1 of 5. Homework #2

26 Linear Programming

Linear Programming Notes VII Sensitivity Analysis

Discrete Optimization

Equilibrium computation: Part 1

LU Factorization Method to Solve Linear Programming Problem

How To Understand Max Min Fairness

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

1 Lecture: Integration of rational functions by decomposition

Optimization Theory for Large Systems

The Multiplicative Weights Update method

Linear Programming: Chapter 11 Game Theory

Math 120 Final Exam Practice Problems, Form: A

Distributed Machine Learning and Big Data

7 Gaussian Elimination and LU Factorization

Linear Programming I

Follow the Perturbed Leader

Chapter 6. Cuboids. and. vol(conv(p ))

Rolle s Theorem. q( x) = 1

Conic optimization: examples and software

Support Vector Machines

An Introduction to Linear Programming

Transcription:

Primal-dual interior-point methods part I Javier Peña (guest lecturer) Convex Optimization 10-725/36-725

Last time: barrier method Approximate min subject to f(x) h i (x) 0, i = 1,... m Ax = b with min subject to tf(x) + φ(x) Ax = b where φ is the log-barrier function m φ(x) = log( h i (x)). i=1 2

Solve a sequence of problems Barrier method min subject to tf(x) + φ(x) Ax = b for increasing values of t > 0, until m/t ɛ. Start with t = t (0) > 0, and solve the above problem using Newton s method to produce x (0) = x (t). For k = 1, 2, 3,... Solve the barrier problem at t = t (k), using Newton s method initialized at x (k 1), to produce x (k) = x (t) Stop if m/t ɛ Else update t (k+1) = µt, where µ > 1 3

Outline Today: Recap of linear programming and duality The central path Feasible path-following interior-point methods Infeasible interior-point methods 4

Linear program in standard form Problem of the form min x subject to c T x Ax = b x 0, where A R m n, b R m, c R n. Recall: Any linear program can be rewritten in standard form. Standard assumption: A is full row-rank. 5

Linear programming duality The dual of the above problem is Or equivalently max y b T y subject to A T y c. max y,s subject to b T y A T y + s = c s 0. Throughout the sequel refer to the LP from the previous slide as the primal problem and to the above LP as the dual problem. 6

Linear programming duality Theorem (Weak duality) Assume x is primal feasible and y is dual feasible. Then b T y c T x. Theorem (Strong duality) Assume primal LP is feasible. Then it is bounded if and only if the dual is feasible. In that case their optimal values are the same and they are attained. 7

Optimality conditions The points x and (y, s ) are respectively primal and dual optimal solutions if and only if (x, y, s ) solves Ax = b A T y + s = c x j s j = 0, j = 1,..., n x, s 0. Two main classes of algorithms for linear programming Simplex method: Maintain first three conditions and aim for the fourth one. Interior-point methods: Maintain first two and the fourth conditions and aim for the third one. 8

Some history Dantzig (1940s): the simplex method. Still one of the most popular algorithms for linear programming. Klee and Minty (1960s): LP with n variables and 2n constraints that the simplex method needs to perform 2 n iterations to solve. Khachiyan (1979): first polynomial-time algorithm for LP based on the ellipsoid method of Nemirovski and Yudin (1976). Theoretically strong but computationally weak. Karmarkar (1984): first interior-point polynomial-time algorithm for LP. Renegar (1988): Newton-based interior-point algorithm for LP. Best known theoretical complexity to date. Modern state-of-the-art LP solvers typically use both simplex and interior-point methods. 9

10 Barrier method for primal and dual problems Pick τ > 0. Approximate the LP primal with min x subject to c T x τ Ax = b n log x j j=1 and the LP dual with max y,s b T y + τ n log s j j=1 subject to A T y + s = c. Neat fact: The above two problems are, modulo a constant, Lagrangian duals of each other.

11 Primal-dual central path Assume the primal and dual problems are strictly feasible. The primal-dual central path is the set {(x(τ), y(τ), s(τ)) : τ > 0} where x(τ), and (y(τ), s(τ)) solve the above pair of barrier problems. Equivalently, (x(τ), y(τ), s(τ)) is the solution to Ax = b A T y + s = c x j s j = τ, j = 1,..., n x, s > 0.

12 Path following interior-point methods Main idea: Generate (x k, y k, s k ) (x(τ k ), y(τ k ), s(τ k )) for τ k 0. Key details: Proximity to the central path Decrease τ k Update (x k, y k, s k )

13 Neighborhoods of the central path Notation: F 0 := {(x, y, s) : Ax = b, A T y + s = c, x, s > 0}. For x, s R n, X := diag(x), S := diag(s). Given x, s R n +, µ(x, s) := xt s n For θ (0, 1), two-norm neighborhood: N 2 (θ) := {(x, y, s) F 0 : XS1 µ(x, s)1 2 θµ(x, s)} For γ (0, 1), one-sided infinity-norm neighborhood: N (γ) := {(x, y, s) F 0 : x i s i γµ(x, s), i = 1,..., n}

14 Newton step Recall: (x(τ), y(τ), s(τ)) solution to A T y + s c 0 Ax b XS1 = 0, x, s > 0. τ1 Newton step equations: 0 A T I x 0 A 0 0 y = 0. S 0 X s τ1 XS1

15 Short-step path following algorithm Algorithm SPF 1. Let θ, δ (0, 1) be such that θ 2 +δ 2 2 3/2 (1 θ) ( 1 δ n ) θ. 2. Let (x 0, y 0, s 0 ) N 2 (θ). 3. For k = 0, 1,... Compute Newton step for ( ) (x, y, s) = (x k, y k, s k ), τ = 1 δ n µ(x, s). Set (x k+1, y k+1, s k+1 ) := (x k, y k, s k ) + ( x, y, s).

16 Theorem The sequence generated by Algorithm SPF satisfies (x k, y k, s k ) N 2 (θ), and µ(x k+1, s k+1 ) = ( 1 δ n ) µ(x k, s k ) Corollary ( n ( )) In O log nµ(x 0,s 0 ) ɛ the algorithm yields (x k, y k, s k ) F 0 such that c T x k b T y k ɛ.

17 Long-step path following algorithm Algorithm LPF 1. Choose γ (0, 1) and 0 < σ min < σ max < 1 2. Let (x 0, y 0, s 0 ) N (γ) 3. For k = 0, 1,... Choose σ [σ min, σ max ] Compute Newton step for (x, y, s) = (x k, y k, s k ), τ = σµ(x k, s k ) Choose α k as the largest α [0, 1] such that (x k, y k, s k ) + α( x, y, s) N (γ) Set (x k+1, y k+1, s k+1 ) := (x k, y k, s k ) + α k ( x, y, s)

18 Theorem The sequence generated by Algorithm LPF satisfies (x k, y k, s k ) N (γ), and µ(x k+1, s k+1 ) ( 1 δ ) µ(x k, s k ) n for some constant δ that depends on γ, σ min, σ max but not on n. Corollary In O such that ( n log ( )) nµ(x 0,s 0 ) ɛ the algorithm yields (x k, y k, s k ) F 0 c T x k b T y k ɛ.

19 Infeasible interior-point algorithms Algorithms SPF and LPF require an initial point in F 0. Can we eliminate this requirement? Given (x, y, s), let r b := Ax b, r c := A T y + s c. Assume (x 0, y 0, s 0 ) with x 0, s 0 > 0 is given. Extend N (γ) to N (γ, β) := {(x, y, s) : (r b, r c ) [ (r 0 b, r0 c ) /µ 0 ]βµ, for γ (0, 1), β 1. x, s > 0, x i s i γµ, i = 1,..., n} Newton step equations: 0 A T I x r c A 0 0 y = r b. S 0 X s XS1 τ1

20 Algorithm IPF 1. Choose γ (0, 1), 0 < σ min < σ max < 0.5, and β 1. 2. Choose (x 0, y 0, s 0 ) with x 0, s 0 > 0 3. For k = 0, 1,... Choose σ [σ min, σ max ] Compute Newton step for (x, y, s) = (x k, y k, s k ), τ = σµ(x k, s k ) Choose α k as the largest α [0, 1] such that (x k, y k, s k ) + α( x, y, s) N (γ, β) and µ(x k + α x, s k + α s) (1 0.01α)µ(x k, s k ) Set (x k+1, y k+1, s k+1 ) := (x k, y k, s k ) + α k ( x, y, s)

21 Theorem Assume the primal and dual problems have optimal solutions. Then the sequence (x k, y k, s k ), k = 0, 1,... generated by Algorithm IPF satisfies µ k := µ(x k, y k, s k ) 0 linearly. In particular (r k b, rk c ) 0 R-linearly. Remark: a k 0 R-linearly a k b k and b k 0 linearly.

22 IPM for more general convex optimization Consider a convex minimization problem min f(x) subject to h(x) 0 Ax = b. Assume f, h smooth and strong duality holds. Then x and (u, v ) are respectively primal and dual optimal solutions if and only if (x, u, v ) solves the KKT conditions f(x) + A T v + h(x)u = 0 Uh(x) = 0 Ax = b u, h(x) 0.

23 Central path and Newton step Central path: {(x(τ), u(τ), v(τ)) : τ > 0} where (x(τ), u(τ), v(τ)) solves f(x) + A T v + h(x)u = 0 Uh(x) = τ1 Ax = b u, h(x) > 0. Newton step: 2 f(x) + i u i 2 h i (x) h(x) A T x r dual U h(x) T H(x) 0 u = r cent, A 0 0 v r pri r dual = f(x)+a T v+ h(x)u, r cent = Uh(x)+τ1, r pri = Ax b.

Given x, u such that h(x) 0, u 0, define µ(x, u) := h(x)t u m. Primal-Dual Algorithm 1. Choose σ (0, 1) 2. Choose (x 0, u 0, v 0 ) such that h(x 0 ) < 0, u 0 > 0 3. For k = 0, 1,... Compute Newton step for (x, u, v) = (x k, u k, v k ), τ := σµ(x k, u k ) Choose steplength αk via line-search and set (x k+1, u k+1, v k+1 ) := (x k, u k, v k ) + α k ( x, u, v) Line-search: Maintain h(x) < 0, u > 0 and reduce r dual, r cent, r pri. 24

25 References and further reading S. Wright (1997), Primal-Dual Interior-Point Methods, Chapters 5 and 6 S. Boyd and L. Vandenberghe (2004), Convex optimization, Chapter 11 J. Renegar (2001), A Mathematical View of Interior-Point Methods