5. Linear Inequalities and Elimination

Similar documents
3. Linear Programming and Polyhedral Combinatorics

Actually Doing It! 6. Prove that the regular unit cube (say 1cm=unit) of sufficiently high dimension can fit inside it the whole city of New York.

. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2

What is Linear Programming?

1 Solving LPs: The Simplex Algorithm of George Dantzig

Some representability and duality results for convex mixed-integer programs.

Duality of linear conic problems

2.3 Convex Constrained Optimization Problems

Linear Programming. March 14, 2014

CHAPTER 9. Integer Programming

Linear Programming I

Can linear programs solve NP-hard problems?

Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method

This exposition of linear programming

Arrangements And Duality

Minkowski Sum of Polytopes Defined by Their Vertices

Math 215 HW #6 Solutions

3.1 Solving Systems Using Tables and Graphs

26 Linear Programming

Practical Guide to the Simplex Method of Linear Programming

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Applied Algorithm Design Lecture 5

International Doctoral School Algorithmic Decision Theory: MCDA and MOO

Lecture 2: August 29. Linear Programming (part I)

Largest Fixed-Aspect, Axis-Aligned Rectangle

LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method

Polytope Examples (PolyComp Fukuda) Matching Polytope 1

1 Symmetries of regular polyhedra

(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties

Duality in General Programs. Ryan Tibshirani Convex Optimization /36-725

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

4.6 Linear Programming duality

Special Situations in the Simplex Algorithm

Nonlinear Programming Methods.S2 Quadratic Programming

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

PYTHAGOREAN TRIPLES KEITH CONRAD

THE DIMENSION OF A VECTOR SPACE

Inner Product Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Linear Programming. April 12, 2005

Minimally Infeasible Set Partitioning Problems with Balanced Constraints

FACTORIZATION OF TROPICAL POLYNOMIALS IN ONE AND SEVERAL VARIABLES. Nathan B. Grigg. Submitted to Brigham Young University in partial fulfillment

Mathematical finance and linear programming (optimization)

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Linearly Independent Sets and Linearly Dependent Sets

The Graphical Method: An Example

Solving Systems of Linear Equations

Linear Programming Notes V Problem Transformations

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the school year.

Systems of Linear Equations

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

STURMFELS EXPANDED. Niels Lauritzen

5.5. Solving linear systems by the elimination method

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Linear Programming in Matrix Form

arxiv: v1 [math.co] 7 Mar 2012

1 Linear Programming. 1.1 Introduction. Problem description: motivate by min-cost flow. bit of history. everything is LP. NP and conp. P breakthrough.

1 Review of Least Squares Solutions to Overdetermined Systems

Zeros of a Polynomial Function

24. The Branch and Bound Method

NOTES ON LINEAR TRANSFORMATIONS

Orthogonal Projections

Solutions to Math 51 First Exam January 29, 2015

On Minimal Valid Inequalities for Mixed Integer Conic Programs

by the matrix A results in a vector which is a reflection of the given

OPRE 6201 : 2. Simplex Method

Linear Programming Notes VII Sensitivity Analysis

4.1 Learning algorithms for neural networks

1 Norms and Vector Spaces

Lecture 2: Homogeneous Coordinates, Lines and Conics

Fiber Polytopes for the Projections between Cyclic Polytopes

Linear Algebra Notes

5.1 Bipartite Matching

An interval linear programming contractor

Intersection of Convex Objects: The Method of Separating Axes

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Notes on Factoring. MA 206 Kurt Bryan

Definition Given a graph G on n vertices, we define the following quantities:

Transportation Polytopes: a Twenty year Update

Linear Programming: Theory and Applications

University of Lille I PC first year list of exercises n 7. Review

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

T ( a i x i ) = a i T (x i ).

Linear Programming Problems

Gautam Appa and H. Paul Williams A formula for the solution of DEA models

1 if 1 x 0 1 if 0 x 1

SOLVING POLYNOMIAL EQUATIONS

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

18.06 Problem Set 4 Solution Due Wednesday, 11 March 2009 at 4 pm in Total: 175 points.

1 Determinants and the Solvability of Linear Systems

Two-Stage Stochastic Linear Programs

Lecture 7: Finding Lyapunov Functions 1

1 Homework 1. [p 0 q i+j p i 1 q j+1 ] + [p i q j ] + [p i+1 q j p i+j q 0 ]

3. INNER PRODUCT SPACES

Section 1.4. Lines, Planes, and Hyperplanes. The Calculus of Functions of Several Variables

The Equivalence of Linear Programs and Zero-Sum Games

NP-Hardness Results Related to PPAD

Solving Systems of Linear Equations

1.7. Partial Fractions Rational Functions and Partial Fractions. A rational function is a quotient of two polynomials: R(x) = P (x) Q(x).

Transcription:

5-1 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 5. Linear Inequalities and Elimination Searching for certificates Projection of polyhedra Quantifier elimination Constructing valid inequalities Fourier-Motzkin elimination Efficiency Certificates Farkas lemma Representations Polytopes and combinatorial optimization Efficient representations

5-2 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Searching for Certificates Given a feasibility problem does there exist x such that f i (x) 0 for all i = 1,..., m We would like to find certificates of infeasibility. Two important methods include Optimization Automated inference, or constructive methods In this section, we will describe some constructive methods for the special case of linear equations and inequalities

5-3 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Polyhedra A set S R n is called a polyhedron if it is the intersection of a finite set of closed halfspaces { } S = x R n Ax b A bounded polyhedron is called a polytope The dimension of a polyhedron is the dimension of its affine hull { } affine(s) = λx + νy λ + ν = 1, x, y S If b = 0 the polyhedron is a cone Every polyhedron is convex

5-4 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Faces of Polyhedra given a R n, the corresponding face of polyhedron P is { face(a, P ) = x P a T x a T y for all y P } face( [ 1 1 ] T, P ), dimension 1 face( [ 2 1 ] T, P ), dimension 0 Faces of dimension 0 are called vertices 1 edges d 1 facets, where d = dim(p ) Facets are also said to have codimension 1

5-5 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Projection of Polytopes Suppose we have a polytope { S = x R n Ax b } We d like to construct the projection onto the hyperplane { } x R n x 1 = 0 Call this projection P (S) In particular, we would like to find the inequalities that define P (S)

5-6 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Projection of Polytopes We have P (S) = { x 2 there exists x 1 such that [ x1 x 2 ] S } Our objective is to perform quantifier elimination to remove the existential quantifier and find a basic semialgebraic representation of P (S) Alternatively, we can interpret this as finding valid inequalities that do not depend on x 1 ; i.e., the intersection cone{f 1,..., f m } R[x 2,..., x n ] This is called the elimination cone of valid inequalities

5-7 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Projection of Polytopes Intuitively, P (S) is a polytope; what are its vertices? Every face of P (S) is the projection of a face of S Hence every vertex of P (S) is the projection of some vertex of S What about the facets? So one algorithm is Find the vertices of S, and project them Find the convex hull of the projected points But how do we do this?

5-8 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Example The polytope S has dimension 55, 2048 vertices, billions of facets The 3d projection P (S) has 92 vertices and 74 facets

5-9 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Simple Example 5 4x 1 x 2 9 (1) x 1 2x 2 4 (2) 4 3 7 5 2x 1 + x 2 0 (3) 3 x 2 6x 2 6 (4) x 1 + 2x 2 11 (5) 2 1 6 6x 1 + 2x 2 17 (6) x 2 4 (7) 1 2 1 2 3 4 5

5-10 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Constructing Valid Inequalities We can generate new valid inequalities from the given set; e.g., if a T 1 x b 1 and a T 2 x b 2 then λ 1 (b 1 a T 1 x) + λ 2(b 2 a T 2 x) 0 is a valid inequality for all λ 1, λ 2 0 Here we are applying the inference rule, for λ 1, λ 2 0 f 1, f 2 0 = λ 1 f 1 + λ 2 f 2 0

5-11 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Constructing Valid Inequalities For example, use inequalities (2) and (6) above x 1 + 2x 2 4 6x 1 2x 2 17 Pick λ 1 = 6 and λ 2 = 1 to give 6( x 1 2x 2 ) + (6x 1 2x 2 ) 6( 4) + 17 2x 2 1 The corresponding vector is in the cone generated by a 1 and a 2 If a 1 and a 2 have opposite sign coefficients of x 1, then we can pick some element of the cone with x 1 coefficient zero.

5-12 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Fourier-Motzkin Elimination Write the original inequalities as along with x 2 4 x 2 4 + 9 4 2x 2 + 4 x 2 2 6x 2 + 6 x 1 2x 2 11 x 2 3 + 17 6 Hence every expression on the left hand side is less than every expression on the right, for every (x 1, x 2 ) P Together with x 2 4, this set of pairs specifies exactly P (S)

5-13 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 The Projected Set This gives the following system of inequalities for P (S) x 2 5 x 2 1 0 7 x 2 1 2 x 2 4 2 5 x 2 17 x 2 4 5 x 2 1 2 x 2 4 There are many redundant inequalities P (S) is defined by the tightest pair x 2 1 2 x 2 4 When performing repeated projection, it is very important to eliminate redundant inequalities

5-14 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Efficiency of Fourier-Motzkin elimination If A has m rows, then after elimination of x 1 we can have no more than m 2 facets 4 If m/2 inequalities have a positive coefficient of x 1, and m/2 have a negative coefficient, then FM constructs exactly m 2 /4 new inequalities Repeating this, eliminating d dimensions gives m 2 2 d inequalities Key question: how many are redundant? i.e., does projection produce exponentially more facets?

5-15 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Inequality Representation Constructing such inequalities corresponds to multiplication of the original constraint Ax b by a positive matrix C In this case C = 1 0 0 0 4 0 0 6 0 0 0 0 4 0 0 1 0 0 1 0 0 0 6 0 0 0 1 0 0 0 1 0 2 0 0 0 0 6 0 0 2 0 0 0 0 1 1 0 0 0 0 0 6 0 1 0 0 0 0 0 0 0 1 A = 4 1 1 2 2 1 1 6 1 2 6 2 0 1 b = 9 4 0 6 11 17 4

5-16 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Inequality Representation The resulting inequality system is CAx Cb, since x 0 and C 0 = Cx 0 We find CA = 0 7 0 14 0 0 0 14 0 5 0 2 0 4 0 38 0 1 Cb = 35 14 7 7 22 34 5 19 4

5-17 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Feasibility In the example above, we eliminated x 1 to find x 2 1 2 x 2 4 We can now eliminate x 2 to find 0 7 2 which is obviously true; it s valid for every x S, but happens to be independent of x If we had arrived instead at 0 2 then we would have derived a contradiction, and the original system of inequalities would therefore be infeasible

5-18 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Example Consider the infeasible system x 1 0 x 2 0 x 1 + x 2 2 Write this as x 1 0 x 1 + x 2 2 x 2 0 Eliminating x 1 gives x 2 2 x 2 0 Subsequently eliminating x 2 gives the contradiction 0 2

5-19 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Inequality Representation The original system is Ax b with A = 1 0 1 1 and b = 0 1 0 2 0 To eliminate x 1, multiply [ ] 1 1 0 (Ax b) 0 0 0 1 Similarly to eliminate x 2 we form [ ] [ ] 1 1 0 1 1 0 0 1 (Ax b) 0

5-20 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Certificates of Infeasibility The final elimination is [ ] 1 1 1 (Ax b) 0 Hence we have found a vector λ such that λ 0 (since its a product of positive matrices) λ T A = 0 and λ T b < 0 (since it gives a contradiction) Fourier-Motzkin constructs a certificate of infeasibility; the vector λ Exactly decides feasibility of linear inequalities Hence this gives an extremely inefficient way to solve a linear program

5-21 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Farkas Lemma Hence Fourier-Motzkin gives a proof of Farkas lemma The primal problem is x Ax b The dual problem is a strong alternative λ λ T A = 0, λ T b < 0, λ 0 The beauty of this proof is that it is algebraic It does not require any compactness or topology It works over general fields, e.g. Q, It is a syntactic proof, just requiring the axioms of positivity

5-22 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Gaussian Elimination We can also view Gaussian elimination in the same way Constructing linear combination of rows is inference Every such combination is a valid equality If we find 0x = 1 then we have a proof of infeasibility The corresponding strong duality result is Primal: x Ax = b Dual: λ λ T A = 0, λ T b 0 Of course, this is just the usual range-nullspace duality

5-23 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Computation One feature of FM is that it allows exact rational arithmetic Just like Groebner basis methods Consequently very slow; the numerators and denominators in the rational numbers become large Even Gaussian elimination is slow in exact arithmetic (but still polynomial) Optimization Approach Solving the inequalities using interior-point methods is much faster than testing feasibility using FM Allows floating-point arithmetic We will see similar methods for polynomial equations and inequalities

5-24 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Representation of Polytopes We can represent a polytope in the following ways an intersection of halfspaces, called an H-polytope { } S = x R n Ax b the convex hull of its vertices, called a V -polytope { } S = co a 1,..., a m

5-25 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Size of representations In some cases, one representation is smaller than the other The n-cube { C n = x R n 1 x i 1 for all i } has 2n facets, and 2 n vertices The polar of the cube is the n-dimensional crosspolytope { Cn = x R n } x i 1 i = co { e 1, e 1,..., e n, e n } which has 2n vertices and 2 n facets Consequently projection is exponential.

5-26 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Problem Solving using Different Representations If S is a V -polytope Optimization is easy; evaluate c T x at all vertices To check membership of a given y S, we need to solve an LP; duality will give certificate of infeasibility If S is an H-polytope Membership is easy; simply evaluate Ay b The certificate of infeasibility is just the violated inequality Optimization is an LP

5-27 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Polytopes and Combinatorial Optimization Recall the MAXCUT problem maximize trace(qx) subject to diag X = 1 rank(x) = 1 X 0 The cut polytope is the set C = co { X S n X = vv T, v { 1, 1} n } = co { X S n rank(x) = 1, diag(x) = 1, X 0 } Maximizing trace QX over X C gives exactly the MAXCUT value This is equivalent to a linear program

5-28 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 MAXCUT Although we can formulate MAXCUT as an LP, both the V -representation and the H-representation are exponential in the number of vertices e.g., for n = 7, the cut polytope has 116, 764 facets for n = 8, there are approx. 217, 000, 000 facets Exponential description is not necessarily fatal; we may still have a polynomial-time separation oracle For MAXCUT, several families of valid inequalities are known, e.g., the triangle inequalities give LP relaxations of MAXCUT

5-29 Linear Inequalities and Elimination P. Parrilo and S. Lall, CDC 2003 2003.12.07.02 Efficient Representation Projecting a polytope can dramatically change the number of facets Fundamental question: are polyhedral feasible sets the projection of higher dimensional polytope with fewer facets? If so, the problem is solvable by a simpler LP in higher dimensions The projection is performed implicitly Convex Relaxation For any optimization problem, we can always construct an equivalent problem with a linear cost function Then, replacing the feasible set with its convex hull does not change the optimal value Fundamental question: how to efficiently construct convex hulls?