A Brief Introduction to Duality Theory Paul J. Atzberger

Similar documents
2.3 Convex Constrained Optimization Problems

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

Nonlinear Programming Methods.S2 Quadratic Programming

Duality of linear conic problems

Date: April 12, Contents

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Convex analysis and profit/cost/support functions

Notes on Symmetric Matrices

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

Support Vector Machine (SVM)

An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM

Nonlinear Optimization: Algorithms 3: Interior-point methods

Linear Programming Notes V Problem Transformations

Mathematical finance and linear programming (optimization)

Optimization in R n Introduction

Support Vector Machines

Inner Product Spaces and Orthogonality

Duality in General Programs. Ryan Tibshirani Convex Optimization /36-725

Inner Product Spaces

An Introduction on SemiDefinite Program

Several Views of Support Vector Machines

SMOOTHING APPROXIMATIONS FOR TWO CLASSES OF CONVEX EIGENVALUE OPTIMIZATION PROBLEMS YU QI. (B.Sc.(Hons.), BUAA)

Big Data - Lecture 1 Optimization reminders

Proximal mapping via network optimization

Interior Point Methods and Linear Programming

Solutions Of Some Non-Linear Programming Problems BIJAN KUMAR PATEL. Master of Science in Mathematics. Prof. ANIL KUMAR

Lecture 2: The SVM classifier

(Quasi-)Newton methods

Practical Guide to the Simplex Method of Linear Programming

Classification of Cartan matrices

CHAPTER 9. Integer Programming

Continuity of the Perron Root

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

Lecture 7: Finding Lyapunov Functions 1

MINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS. denote the family of probability density functions g on X satisfying

Adaptive Online Gradient Descent

Linear Programming I

10. Proximal point method

Chapter 4. Duality. 4.1 A Graphical Example

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

Convex Programming Tools for Disjunctive Programs

A Lagrangian-DNN Relaxation: a Fast Method for Computing Tight Lower Bounds for a Class of Quadratic Optimization Problems

Factorization Theorems

Support Vector Machines Explained

Summer course on Convex Optimization. Fifth Lecture Interior-Point Methods (1) Michel Baes, K.U.Leuven Bharath Rangarajan, U.

Conic optimization: examples and software

Lecture 5 Principal Minors and the Hessian

constraint. Let us penalize ourselves for making the constraint too big. We end up with a

Support Vector Machines

Derivative Free Optimization

LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method

TOPIC 4: DERIVATIVES

Section Inner Products and Norms

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

1 Solving LPs: The Simplex Algorithm of George Dantzig

Follow the Perturbed Leader

Convex Rationing Solutions (Incomplete Version, Do not distribute)

3. INNER PRODUCT SPACES

Solving polynomial least squares problems via semidefinite programming relaxations

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

4.6 Linear Programming duality

Biggar High School Mathematics Department. National 5 Learning Intentions & Success Criteria: Assessing My Progress

3. Linear Programming and Polyhedral Combinatorics

Advanced Microeconomics

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Cost Minimization and the Cost Function

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

by the matrix A results in a vector which is a reflection of the given

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

This unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

Advanced Lecture on Mathematical Science and Information Science I. Optimization in Finance

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

Numerical Analysis Lecture Notes

Linear Algebra Review. Vectors

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Mathematical Methods of Engineering Analysis

Lecture 2: August 29. Linear Programming (part I)

Chapter 6. Linear Programming: The Simplex Method. Introduction to the Big M Method. Section 4 Maximization and Minimization with Problem Constraints

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

CONSTRAINED NONLINEAR PROGRAMMING

Vector and Matrix Norms

Completely Positive Cone and its Dual

Some Lecture Notes and In-Class Examples for Pre-Calculus:

Math 120 Final Exam Practice Problems, Form: A

Inner products on R n, and more

DRAFT. Algebra 1 EOC Item Specifications

Applications to Data Smoothing and Image Processing I

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

Section 1.1. Introduction to R n

BANACH AND HILBERT SPACE REVIEW

T ( a i x i ) = a i T (x i ).

1 if 1 x 0 1 if 0 x 1

Big Data Analytics: Optimization and Randomization

Stationarity Results for Generating Set Search for Linearly Constrained Optimization

Lecture Topic: Low-Rank Approximations

Network Traffic Modelling

CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA

What is Linear Programming?

Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method

Transcription:

A Brief Introduction to Duality Theory Paul J. Atzberger Comments should be sent to: atzberg@math.ucsb.edu

Abstract These notes give an introduction to duality theory in the context of linear and positive semidefinite programming. These notes are based on material from Convex Analysis and Nonlinear Optimization by Borwein and Lewis and Numerical Optimization by Nocedal and Wright. Two examples are given to show how duality can be used. The first optimization application is to find the matrix in an affine family that has the least 2-norm. This problem can be formulated as the dual to a positive semi-definite program. The second problem is to find for an empirical correlation matrix Q the closest positive semidefinite matrix Q with respect to the Frobenius norm. This problem can be formulated as a positive semidefinite program. We also derive it s dual problem. Lastly, we briefly discuss some numerical methods useful in solving these constrained optimization problems. Cones An important construct that arises in optimization problems is a cone K. Cones are often used to formulate constraints in optimization problems and have some important properties with respect to the dual. We review here some basic definitions and results concerning cones. Definition 1 A cone K is a set invariant under multiplication by non-negative scalars. If x K and λ 0 then λx K. Definition 2 The dual of a cone is Dual(K) = {z R n z T x 0} Three important cones are: Definition 3 Quadratic Cone: K q = {z R m (z 2,...,z m ) 2 z 1 } Definition 4 Positive Orthant Cone: K + = {z R m z 1 0,z 2 0,...,z m 0} Definition 5 Positive Semidefinite Cone: K S+ = {X S n n X 0} Each of these cones is self-dual: Lemma 1 The positive orthant cone K + in R m is equal to its dual cone: Proof: We will show that z T x 0 x K + z i 0 i. If z T x 0 x K + then since e i K we have that z i 0 i. If z i 0 i then since any x K + has x i 0 i we have that z T x 0. Therefore z Dual(K + ) z K +. 2

Lemma 2 The positive semidefinite cone K S+ in S n n is equal to its dual cone: Proof: We will show that Z X 0 X K S+ Z K S+. If Z X 0 X K S+ then since Z S n n we can choose a basis so that Z is diagonal. Let X = e k e T k then X K S+ and Z X = Z k,k 0. This requires that Z be positive semidefinite and Z K S+. Let Z K S+ then by definition of positive semidefiniteness v T Zv 0, v. In particular e T k Ze k = Z k,k 0. If for any X K S+ we represent matrices in the eigenbasis of X we obtain (µ k 0 eigenvalues of X): Z X = µ 1 Z 1,1 + µ 2 Z 2,2 + + µ n Z n,n min{µ k } tr(z) 0 Therefore dual(k S+ ) = K S+. Lemma 3 The quadratic cone K q in R is equal to its dual cone: Proof: We first will show that the quadratic code is contained in the dual cone. Since all vectors in the quadratic cone have first component non-negative we have: z T x 0 z 1 x 1 z 2 x 2 z n+1 x n+1 z 2 x 2 z n+1 x n+1 (z 2,...,z n+1 ) T (x 2,...,x n+1 ) = (z 2,...,z n+1 ) (x 2,...,x n+1 ) cos(θ) z 1 x 1 Thus K Dual(K). Next we show that the dual cone is contained in the quadratic cone. If z Dual(K) then for all x K: z 1 x 1 z 2 x 2 z n+1 x n+1 = (z 2,...,z n+1 ) T (x 2,...,x n+1 )cos(θ) = (z 2,...,z n+1 ) (x 2,...,x n+1 ) cos(θ) Now choose the member x of K so that x 1 = 1, sign(x 1) = sign(z 1 ), (x 2,...,x n+1) = 1 and cos(θ) = 1 then z 1 x 1 = z 1 and we obtain: z 1 (z 2,...,z n+1 ) 3

Thus z K and K Dual(K). Therefore K = Dual(K). Definition 6 The indicator function for a set U denoted δ U is defined as { } 0 if x U δ U (x) = if x U Definition 7 The Fenchel Conjugate of a function f(x) is f (φ) = sup x R n {< φ,x > f(x)} Lemma 4 If K is a cone and δ K (x) is the indicator function the Fenchel Conjugate of the indicator function is δ K (φ) = δ Dual(K)(φ). Proof: δ K(φ) = sup x R n {< φ,x > δ K (x)} = sup{< φ,x >} x K { if x K s.t. < φ,x >> 0 = 0 if < φ,x > 0, x K = δ Dual(K) (φ) Linear and Semidefinite Programs Many optimization problems can be arranged into a standard form. Below we discuss some important standard forms that can arise. Optimization problems involving vectors with linear inequality and equality constraints and a linear objective function can be formulated as: LP Program: inf x Rn{< c,x > x 0,Ax = b} This optimization problem is referred to as a linear program. To formulate inequalities a slack variable z can be introduced and the linear inequality can be formulated as an equality. For example the inequality ax 1 2 can be formulated as z 0 ax 1 z = 2 4

Optimization problems involving matrices with linear equality constraints, linear objective function, and inequality constraints involving the definiteness of the matrix can often be formulated as: SDP Program: inf {C X X 0,A i X = b i } X S n n This optimization problem is referred to as a semidefinite program. The dot product refers to vector dot product regarding the entries of each matrix as a vector A B = tr(ab). Duality The central result we will use in the derivation of the dual problem is the Fenchel-Young Inequality. Theorem 1 (Fenchel-Young Inequality) Let f (φ) = sup x R {< φ,x > f(x)}. This is called the Fenchel Conjugate and the following inequality holds: f(x) + f (φ) < φ,x > Equality holds if φ is a subgradient of f at x, φ f(x). Proof: < φ,x > f(x) f (φ) f(x) + f (φ) < φ,x > If φ is a subgradient of f at x then: f(x)+ < φ,(y x) > f(y) f(x) < φ,x > f(y) < φ,y > f(x) < φ,x > f (φ) f(x) + f (φ) < φ,x > Therefore: f(x) + f (φ) =< φ,x > The Fenchel-Young Inequality has a geometric interpretation if we view the graph of f(x) as a surface over the x hyperplane in the z-direction. We can rearrange the inequality to become: < φ,x > f (φ) f(x) The inequality then states that out of all hyperplanes with slope vector φ the greatest z-intercept attainable while still remaining below the surface is f (φ). 5

Theorem 2 inf {f(x)} = inf x R sup x R φ R {< φ,x > f (φ)} Proof: By the Fenchel-Young Inequality: For the other inequality we use: f(x) sup{< φ,x > f (φ)} φ R inf {f(x)} x R inf {< φ,x > f (φ)} x R φ R f (0) = sup{ f(x)} = inf {f(x)} x R x R Then sup{< φ,x > f (φ)} inf {f(x)} φ R x R This consequence of the Fenchel-Young Inequality offers an alternative formulation an optimization problem. Geometrically this can be thought of as the following procedure to find the minimum: 1. At a given x, find for all the hyperplanes that lie below the surface the one that is closest to attaining f(x) at x, denote the value at x by g(x). (If no one hyperplane exists then use the monotonic sequence of approximations to the value of f(x) at x to define g(x)) 2. Minimize the function g(x). Below are some examples using the alternative approach to find the minimizer. For the first example we demonstrate that the alternative problem amounts to the same thing as the primal problem. Example 1 inf + 1} x R 2{x2 The Fenchel Conjugate is f (φ) = sup{φx x 2 1} x R = sup{ (x φ x R 2 )2 + φ2 4 1} = φ2 4 1 6

This gives for the dual formulation of the problem: inf sup x R φ R {φx φ2 + 1} = inf 4 sup x R φ R = inf x R {x2 + 1} = 1 { 1 4 (φ 2x)2 + x 2 + 1} Example 2 The alternative formulation of the problem can sometimes lead to a problem with poor properties. One example were this occurs is the following: f(x) = { ln(1 + x) if x 0 ln(1 x) if x 0 } 1 The minimizer is at x = 0. Since the tangent lines to ln(1 + x) have slopes 1+x 0 as x ± we see no line other than the one with slope 0 lies below the graph of f(x). This means f (φ) = for φ 0 and f (0) = 0. The new objective function inside the infimum becomes discontinuous and infinity everywhere except at the solution point. In fact even formulating the problem which involves finding the Fenchel Conjugate (z-intercept) is the same as having found the minimizer of the problem. While the alternative formulation in the examples above might not offer much of an advantage. We will see below that there are cases in which the alternative form does offer some advantages after an appropriate change of variable in φ. One important advantage is that it is often possible to turn equality constraints into inequality constraints under an appropriate change of variable. In addition many practical problems can be formulated as the alternative problem associated with one of the standard forms above. We refer to the alternative formulation of the minimization problem as the dual problem. The original minimization problem is referred to as the primal problem. We now derive the dual problem for linear programming. LP Primal: LP Dual: inf x Rn{< c,x > x 0,Ax = b} Derivation: Let sup {< y,b > A T y c 0} y R m V = {x R n Ax = b,a R m n } K + = {x R n x 0} 7

Then f (φ) = sup x R n {< φ c,x > δ K+ (x) δ V (x)} sup {< φ c,x > δ K+ (x)} + sup { δ V (x)} x R n x R n = δ K+ (φ c) inf x R n{δ V (x)} If we make a change of variable: The alternative formulation becomes: φ = A T y inf x Rn{f(x)} = inf x R n inf x R n sup {< φ,x > f (φ)} φ R m sup y R m {< y,b > f (A T y)} sup {< y,b > δ K+ (A T y c) + inf y R m x R n{δ V (x)}} sup {< y,b >} y R m,a T y c 0 Next we find the dual problem for semidefinite programming. SDP Primal: inf X S n n{c X X 0,A i X = b i } SDP Dual: Derivation of dual: Let sup y R,C yiai 0 {y b} K S+ = {X S n n X 0} V = {X S n n A i X = b i } By the theorems above: inf X S n n{c X X 0,A i X = b i } = inf X S n n{c X + δ K(X) + δ V (X)} = inf sup {Φ X f (Φ)} X S n n Φ R n n 8

The Fenchel Conjugate of f(x) = C X + δ K (X) + δ V (X) is by the lemmas above: f (Φ) = sup X S n n {(Φ C) X δ KS+ (X) δ V (X)} δ Dual(K) (Φ C) inf {δ V (X)} X S n n = δ K (Φ C) inf {δ V (X)} X S n n Under the change of variable: We obtain: Φ = y i A i inf X S n n,a i X=b i sup {Φ X f (Φ)} Φ R n n sup { y i b i f ( y i A i )} y R sup {y b δ K ( y i A i C) + y R sup y R,C yiai 0 {y b} inf {δ V (X)}} X S n n By examining the derivations above we obtain the following theorem. Theorem 3 (Weak Duality) For the linear and positive semidefinite program if we let p denote the solution value obtained in the primal problem and d denote the solution value obtained in the dual problem. Then Proof:(see the derivations of the dual problem) d p Under addition conditions on the linear and semi-definite programs it is possible to show that the primal and dual problems obtain the same values p = d. See reference [1] for the Strong Duality Theorem. Two Applications of SDP Programming and Duality Before discussing the applications we first prove a theorem that will be useful. Theorem 4 Let M be any m n matrix. Let C = [ 0 M M T 0 ] The non-negative eigenvalues of C are the singular values of M. 9

Proof: Every matrix has a SVD decomposition. For M we have: Let then WMV = Σ U = 1 [ ] W W 2 V V UU = 1 [ ] [ ] W V W W 2 W V V V [ WW + V V WW + V V ] = 1 2 WW + V V [ ] I 0 = 0 I WW + V V and UCU = 1 [ ] [ ] W V MV MV 2 W V M T W M T W [ WMV + V M T W WMV V M T W ] = 1 2 WMV + V M T W [ ] Σ 0 = 0 Σ WMV V M T W Since C is symmetric and U is unitary the eigenvalues are σ i and σ i where Problem 1 Find the member of the matrix family: Σ = [δ ij σ i ] B(y) = B 0 + m y k B k that has the least 2-norm. This problem can be formulated as the dual problem of a semidefinite program. k=1 SDP Dual: sup y R,C yiai 0 {y b} 10

Let t := y b i := 1 for i = m + 1 b i := 0 for 1 i m [ ] I 0 A i := for i = m + 1 0 I [ ] 0 Bi A i := Bi T for 1 i m 0 [ ] 0 B0 C := B0 T 0 [ ] ti B(y) Q(y,t) := B(y) T ti ([ ] [ ]) = U Σ 0 ti 0 + U 0 Σ 0 ti The dual SDP problem becomes sup { t} (y,t) R m R,Q(y,t) 0 Now we show that this formulation is equivalent to the problem of finding the matrix with least 2-norm in the affine family. The 2-norm of a matrix is the maximum singular value. For Q(y,t) we have: Q(y,t) 2 = t + σ max (y) The positive semidefinite constraint on the matrix Q requires that the minimum eigenvalue is nonnegative: t σ max (y) 0 Since the positive semidefinite program takes sup{ t} we see that it is maximized when t is smallest. For a given y the smallest value of t is obtained when: t = σ max (y) = B(y) 2 Therefore the 2-norm problem is equivalent to the dual positive semi-definite program. Now we derive the primal version of the positive semidefinite program. SDP Primal: Which can be simplified a little. For a general matrix P of the type above: inf X S n n{c X X 0,A i X = b i } P X = [ 0 M M T 0 ] X 11

= = M i,j x i,(j+n) + (1 i n),(1 j n) (1 i n),(1 j n) 2M i,j x i,(j+n) (1 i n),(1 j n) M T i,jx (i+n),j This gives: C X = (1 i n),(1 j n) B 0i,j x i,(j+n) A k X = 0 B ki,j x i,(j+n) = 0 (1 i n),(1 j n) [ I 0 0 I ] X = = 1 2n x i,i Now we show that the primal and dual problems both have strictly feasible points: The dual problem requires: (y,t) R m R [ 0 B(y) B(y) T 0 ] [ ti 0 + 0 ti ] 0 Given any y there is a minimum eigenvalue of the first matrix, so if t is chosen sufficiently large all eigenvalues will be positive. Therefore there is a strictly feasible point for the dual problem. The primal problem requires: X S n n X 0 B ki,j x i,(j+n) = 0 (1 i n),(1 j n) 2n x i,i = 1 If all x i,j = 0 with i j > 0 and x i,i = 1 2n then the above equality constraints are satisfied and the matrix X is positive definite. Therefore there is a strictly feasible point for the primal problem. 12

Problem 2 Given a symmetric matrix Q find the best approximating positive semidefinite matrix Q with respect to the Frobenius norm. This problem arises naturally when computing a correlation matrix from empirical data. The empirical correlation matrix may be symmetric but lack positive definiteness. The problem can be formulated as a mixed semidefinite-quadratic cone program. inf {C X + c x X 0,x K quadratic,a i X + a i x = b i } X S n n,x R m Let K quadratic := {x R m (x 2,x 3,...,x m ) 2 x 1 } D = Q Q then since D is symmetric we can represent the matrix as a vector d of length n(n+1) 2 = m. d k = Q α(k),β(k) Q α(k),β(k) where α( ),β( ) are the maps that give the corresponding i,j entry of the matrix. Since the 2-norm of d is proportional the the Frobenius norm of D finding the best Frobenius approximation is the same as minimizing the 2-norm of d. We can use the quadratic cone constraint to formulate this by introducing equality constraints: By the quadratic cone constraint x i = d i 1 for 2 i m + 1 Let (x 2,x 3,...,x ) 2 x 1 The objective is then to minimizer x 1 subject to the constraints. C = 0 c 1 = e 1 c i = 0 for (2 i m + 1) A 1 = 0 A i = 1 2 e α(i 1)e T β(i 1) + 1 2 e β(i 1)e T α(i 1) for (2 i m + 1) a 1 = 0 a i = e i for (2 i m + 1) b 1 = 0 b i = Q α(i 1),β(i 1) for (2 i m + 1) The primal problem becomes: inf {x 1 X 0,x K quadratic,x i + X α(i 1),β(i 1) = Q α(i 1),β(i 1) } X S n n,x R 13

Now we derive the dual problem. Using the derivation above for the semidefinite problem and the fact that the dual cone of the quadratic cone K is itself we obtain by a similar computation: sup {z b z i a i c K quadratic, z i A i C 0} z R For the given problem this becomes: sup { z R i=2 z i Qα(i 1),β(i 1) e 1 i=2 z i e i K quadratic, i=2 ) z i (e α(i 1) e T β(i 1) + e β(i 1)e T α(i 1) 0} If we let Z := 1 2 i=2 z i (e α(i 1) e T β(i 1) + e β(i 1)e T α(i 1) ) then the problem can be written more simply as: sup z R {Z Q (1, z 2,..., z ) K quadratic,z 0} We now show each method has a strictly feasible point. For the primal problem the constraints require: X S n n x R X 0 x K quadratic x i + X α(i 1),β(i 1) = Q α(i 1),β(i 1) for (2 i m + 1) If we set X α(i 1),β(i 1) = Q α(i 1),β(i 1) + M α(i 1),β(i 1) where λ min is the minimum eigenvalue of Q and M = 2λ min I then X 0. We set corresponding entries of X by symmetry. We then must set x i = M α(i 1),β(i 1) for (2 i m + 1) to satisfy the equality constraint and since x 1 is free we can set it to x 1 = 2 (x 2,...,x 2. This is a strictly feasible point for the primal problem. For the dual problem the constraints require: z R e 1 i=2 i=2 z i e i K quadratic ) z i (e α(i 1) e T β(i 1) + e β(i 1)e T α(i 1) 0 Let z γ(k) = 1 2n and all other z i = 0 where γ(k) is the map which gives the corresponding index of the k th diagonal entry of Z. Then the negative semidefinite constraint will be satisfied strictly. The first constraint will be satisfied since there are only n diagonal entries of Z. This gives a feasible point for the dual problem. 14

Numerical Methods Two important classes of numerical methods used for solving constrained optimization problems are Log Barrier Methods and Primal-Dual Interior Point Methods. We will only briefly sketch the ideas behind the numerical methods here and refer the reader to reference [2] for more details. The log barrier methods approximately solve constrained optimization problems by augmenting the objective function by a penalty term: h(x) = f(x) + µ m log(c i (x)) The constraints are formulated as c i (x) 0. A sequence of optimization problems is then solve for successively smaller values of µ. Typically the initial values of µ are set to be large so that the constraints can be easily satisfied. Next µ is gradually reduced to zero which has the effect of increasing the importance of the objective function and results in a better approximation to the true solution. The advantage of this approach is that if a solution lies on the boundary of the feasible set the approximate solutions will be guided along a path in the interior of the set of feasible points as µ 0. It can be shown that if at each stage the optimization problem corresponding to µ is solved to sufficient accuracy and the parameter µ is taken sufficiently close to zero then the solutions become close to the true constrained minimizer of the objective function. For more details see Chapter 17 of Nocedal and Wright [2]. A second class of numerical methods useful in solving linear and semidefinite programs with constraints are the Primal-Dual Methods. The basic idea behind these methods is to search for a point that satisfies the Karush-Kuhn-Tucker necessary conditions for the constrained minimization problem. The Karush-Kuhn- Tucker conditions are: f(x ) = i A λ i c i (x ) λ i 0 λ i c i (x ) = 0 where A = {i c i (x) = 0} is the active set. In addition it is usually required that the constraints be such that the gradients of { c i (x) i A(x)} are linearly independent. This is referred to as the Linear Independent Constraint Qualification (LICQ). Geometrically the KKT conditions with LICQ can be interpreted as stating that at a local minimizer of the constrained optimization problem the objective function f(x) must have either all derivatives zero or any non-zero components of the steepest descent vector must point outside the feasible set. In other words the steepest descent vector must be in the span of the negative gradients of the constraint functions c i (x). Thus any step in a descent direction leaves the feasible set. The Primal-Dual Interior Methods search for a minimizing feasible point by satisfying approximate KKT conditions and the constraints of the primal and dual problems. In particular they search for a point (x τ,λ τ,s τ ) solving: A T λ + s = c Ax = b x i s i = τ (x,s) 0 where s is the slack variables used to represent the dual inequality constraint. 15

If the RHS is subtracted from each equation and the difference is denoted by the vector valued function F(x,λ,s) then finding a root of F amounts to finding a point satisfying the above conditions. This can be achieved numerically using Newton s method. As τ is reduced to zero the sequence of points (x τ,λ τ,s τ ) approaches a point satisfying the constraints and KKT conditions. It turns out the path followed by the points in primal space x is similar to the path obtained from the log barrier method. For a more thorough discussion of the methods and implementation see Chapter 14 of Nocedal and Wright [2]. Two implementations of Primal-Dual Methods available freely and worth experimenting with are SeDuMi and SDPT3. As websites quickly become outdated we do not give the URL s here. The distributions are easy to locate using any of the popular search engines. 16

References [1] Convex Analysis and Nonlinear Optimization, Borwein and Lewis, Canadian Mathematical Society, 2000. [2] Numerical Optimization, Nocedal and Wright, Springer, 1999. Paul J. Atzberger Any comments should be send to: atzberg@math.ucsb.edu