Convex Programming Tools for Disjunctive Programs
|
|
|
- Abigail Smith
- 10 years ago
- Views:
Transcription
1 Convex Programming Tools for Disjunctive Programs João Soares, Departamento de Matemática, Universidade de Coimbra, Portugal Abstract A Disjunctive Program (DP) is a mathematical program whose feasible region is the convex hull of the union of convex sets. The objective function is also convex. Disjunctive Programming models are very frequent as a modeling framework for setup costs or constraints, and models with constraints that are better expressed as disjunctions. Some Process Sinthesis Design models arising from Chemical Engineering are mixed integer convex programming models which are particular instances of a Disjunctive Program. In this talk we will address questions that are raised when conceptualizing a Branch-and-cut algorithm for mixed-integer convex programming. 1 Introduction The process synthesis network problem problem in Chemical Engineering is the problem of simultaneously determining the optimal structure and operating parameters for a chemical synthesis problem. This problem can be modeled as a mixed 0-1 convex program where the continuous variables represent process parameters such as flowrates and the 0-1 variables represent the potential existence of a process unit. The nonlinear elements come from the intrinsic nonlinear input-output performance equations of some process units, see [3] where this model is proposed and [8, 4] for related models. Other models of Network Design in communication and transportation networks are discrete by nature. The 0-1 variables represent the potential existence of multiplexers, concentrators, or interface message processors in computer communication networks, junctions in pipeline networks, interchanges in highway networks, and so on. Discrete variables may also represent the discrete quantity of physical arc units of certain characteristics between two junctions of the network. In a simple model, described in [5], the nonlinear element comes from modelling delay at some link (i, j) as proportional to the fraction of the rate of messages crossing the link (i, j) to the available capacity of the same link. We propose a cutting-plane algorithm for solving the following mathematical program that we will refer to as a mixed zero-one convex program, min f(x) G(x) 0 x i {0, 1}, i = 1,..., p, (1) 1
2 where f: IR n IR is a closed convex function and G: IR n IR m is a vector function of closed convex functions. The variables x i, for i = 1,..., p, are zero-one constrained and the variables x i, for i = p + 1,..., n, are simply nonnegative. We will further assume that both f and G are continuous in an open set containing the continuous relaxation, i.e., when {0, 1} is replaced by [0, 1]. Our work extends to the nonlinear setting the lift-and-project approach of Balas, Ceria and Cornuéjols [1, 2], which is seen as one of the most important practical contributions to the solution of mixed zero-one linear programs by general-purpose cutting-plane based algorithms since the work of Gomory in the sixties. As proposed by Stubbs and Mehrotra [7], we solve the cut generation problem in its dual form. Some of the distinctive features of our algorithm are the following: our algorithm guarantees the existence of a cut whenever an optimal solution was not yet found; we solve the cut generation problem using standard nonlinear programming algorithms; and, we fully extend the lifting procedure to the nonlinear setting. The article is structured in the following way. In Section 2 we describe the basic cuttingplane algorithm specialized to solve Program (1). In Section 3 we explain how the cut generation problem can be solved in a smaller-dimension space, taking advantage of the fact that some variables are already integral. In the talk, the algorithm will be illustrated on a small example. A practical implementation of this method is part of an ongoing research project. 2 The basic cutting plane algorithm Our approach requires that we use the following equivalent formulation of program (1), f(x) x n+1 G(x) 0 x i {0, 1}, i = 1,..., p. (2) Since f is convex then this formulation is still a mixed zero-one convex program. Moreover, the feasible region K can be replaced by P = conv (K) without loss of generality, where we note that P is closed. As a matter of notation, we will still use the same f(x) and G(x) eventhough we are refering to these functions as functions of the first n components of the vector x that now lies in IR n+1. A specialization of the basic cutting-plane algorithm is presented in Figure 1. The algorithm requires performing three basic steps in each iteration. In the first step, the relaxation step, we seek an optimal solution x of the following convex program x P, whose feasible region P is defined by { P x IR n+1 f(x) x : n+1, a i x b i, i = 1,..., m 1, G(x) 0, x i [0, 1], i = 1,..., p, (3) }, (4) where m 1 is the number of cuts generated so far. In the second step, the optimality check step, we try to reduce as much as possible the number of fractional components of x while keeping the same value of the component x n+1. In the third step, the separation step, we 2
3 use the last index j tried in the second step to define the following disjunctive programming relaxation P j of P, P j conv (( P {x: xj = 0} ) ( P {x: xj = 1} )). (5) The following proposition shows that a nonoptimal x guarantee the existence of a separating hyperplane. P j, from where we are able to Proposition 1 In each iteration of the algorithm BCP4MINLP, Step 2 is performed at most p times. Moreover, if j is the last index tried in Step 2 then either x is optimal or x P j. Proof: We recall that in Step 2 of the algorithm BCP4MINLP, the integer-constrained variables are sequentially fixed at one of their bounds, zero or one, until an index j is found such that min min x n+1 x P, > x n+1, (6) i=0,1 x F = x F, x j = i where F identifies the variables that are fixed in the process. Since F can have at most p elements then Step 2 is performed at most p times until either (6) holds or all the integer constrained variables are fixed in which case we would have found an optimal solution. Now, we prove the second part of this proposition. Let j be the last index tried in Step 2 so that (6) holds. Assume, by contradiction, that x P j. Then, the point x can be represented by one of the following three possible ways: a. x = δz + (1 δ)y, where δ (0, 1), z P {x: x j = 0} and y P {x: x j = 1}; b. x = z + dy, where z P {x: x j = 0} and dy is a direction of the set P {x: x j = 1} if this set is nonempty or the zero vector otherwise. c. x = dz + y, where y P {x: x j = 1} and dz is a direction of the set P {x: x j = 0} if this set is nonempty or the zero vector otherwise. If x can be decomposed as in a. then, since x k {0, 1}, for every k F, we must have z k = y k = x k, for every k F. Thus, x conv (( P {x: xf = x F, x j = 0} ) ( P {x: xf = x F, x j = 1} )) which contradicts (6). If x can be decomposed as in b. then, since dy i = 0, for every i {1,..., p}, z k = x k, for every k F {j}. Thus, x P {x: x F = x F, x j = 0} which contradicts (6) once again. If x can be decomposed as in c. an analogous argument as in b. applies. 3 The cut generation problem We explain how the cut generation solution procedure should be implemented to take advantage of the fact that many variables have been fixed during the second step of the algorithm BCP4MINLP. Our cut generation problem uses the following duality result sup α x β (α, β) polar ( P j ), α F 1 = inf x x x P j, x F = x F. (7) 3
4 Data: Functions f and G. The scalars n and p. Initialization: Set k { = 0 and define P 0 as P 0 x IR n+1 f(x) x : n+1, G(x) 0, x i [0, 1], i = 1,..., p, Iteration-k: }, Step 1: Step 2: (Relaxation) Let x be the optimal solution of x P k, (Optimality check) Define F {j {1,..., p}: 0 < x j < 1} and F = {1,..., p} \ F. If F is empty then stop: x is an optimal solution and x n+1 is the optimal value. Otherwise, let j F. Step 2.1: Find an optimal solution ˆx of min x P k, x F = x F x j = 0 x P k x F = x F x j = 1 Step 2.2: If ˆx n+1 = x n+1 then let x = ˆx and restart Step 2; Otherwise set x k = x and continue to Step 3. Step 3: (Separation) Let j be the last index tried in Step 2. Find a separating hyperplane a k+1 x b k+1 between P k j and xk. Define P k+1 = P k { x: a k+1 x b k+1 } and set k := k + 1. Figure 1: The basic cutting-plane algorithm for mixed zero-one convex programming (BCP4MINLP) 4
5 Let us assume without loss of generality that the relaxation P is defined by P {x IR n : G(x) 0, x 0, x i 1, i = 1,... p}. (8) and P j is defined by (5). Let F be an index set that may or may not be related to the Step 2 of the cutting-plane algorithm, and F = {1,..., n}\f be its complement. If ˆx = (ˆx F, x F ) P j is a known optimal primal solution in (7) then the subgradient ˆξ = (ˆξ F, ˆξ F ) of the function f(x) = { xf x F if x F = x F, + otherwise, (9) at the point ˆx that satisfies ˆξ(x ˆx) 0, for every x P j, defines an optimal dual solution (ˆα, ˆβ) polar ( P j ). However, the main purpose of (7) is to define the cut generation problem using a smaller number of variables. This means that after solving the primal problem min x F x F x F { x F : (x F, x F ) P } (10) j we have at hand an optimal primal solution ˆx F and a subgradient ˆξ F of the function x F at ˆx F, such that ˆξ F (x F ˆx F ) 0, for every x F { x F : (x F, x F ) P } j. Thus, a natural question is whether we can extend ˆξ F so that ˆξ = (ˆξ F, ˆξ F ) is a subgradient of the function f defined by (9) at ˆx = (ˆx F, x F ) that satisfies ˆξ(x ˆx) 0, for every x P j. Another natural question is whether we can apply a similar mechanism even when ˆx F is not optimal. Our answers to these questions require that x F = 0. This can be done without loss of generality because when x k, for some k F, is nonzero then as long as it coincides with one of its bounds on Program (2) a variable transformation allows for the requirement to hold. In this setting, x F is feasible for Program (10) if and only if x F belongs to conv ({ x F : (x F, 0) P, x j = 0 } { x F : (x F, 0) P, x j = 1 }). (11) Note that the two individual sets that define this convex hull are the feasible regions in (6) and consequently at the end of Step 2 we already know whether those sets are empty or nonempty. This feature is important because it determines which is the best solution procedure to use on Program (10). If both sets are nonempty then the program can be handled using the solution procedures described on Sections 5.4 and 5.5 of [6]. If one of them is empty then Program (10) is a standard convex program, and may therefore be solved by a standard nonlinear programming algorithm. If the two sets are empty then there is no feasible solution x to Program (2) such that x F = x F. In this case the following inequality x k + (1 x k ) 1, k F : x k =0 k F : x k =1 separates x from the convex hull of the feasible region of Program (2). Now, we explain how the lifting procedure works under two distinct situations, depending on the fact that one or none of the sets in (11) is empty. We start by assuming that none of them is empty. Let ˆx F be a feasible solution for Program (10) and ˆξ F be a subgradient of the function x F at ˆx F such that for a given scalar β satisfying ˆξ F x F < β the following holds min min ˆξ F z F (z F, 0) P, β. (12) i=0,1 z j = i 5
6 We remark that ˆx F need not be optimal for Program (10), though if it were optimal then the existence of a subgradient and a scalar satisfying (12) would be guaranteed. Under a constraint qualification, the optimal solution ẑf i of each one of the problems in (12) also solves a linear program defined by a suitable matrix A i G(ẑF i, 0) so that min ˆξ F z F min G(ẑF i i=0,1, 0) + Ai (z F ẑf i, 0) 0, z k 0, k F, z j = i, β (13) The feasible region of each one of these linear programs defines an outer-approximation of each one of the sets in (11). Our lifting procedure applies to these linear programs, so that by the outer-approximation argument it also applies to our original nonlinear sets. Proposition 2 below describes the lifting mechanism in generic terms. Proposition 2 Let F be an index set and F = {1,..., n} \ F. For a given arbitrary vector α F, let ẑ F be an optimal solution of the following linear program: min α F z F A F z F b, l F z F u F, (14) where l F and u F are the, possibly infinite, lower and upper bounds, A F IR m F and b IR m. Then, for any extended matrix A = [A F, A F ] IR m n there is a closed-form extended vector α = (α F, α F ) such that the vector ẑ = (ẑ F, 0) is an optimal solution of the following linear program: min αz Az b, (15) l F z F u F, z F 0. Proof: Let ˆv 0 be the optimal dual multipliers associated with the matrix constraints in Program (14) and define ( ) m α k max 0, ˆv l a lk, for every k F. Now, consider Program (15) and use the same dual variables to price the new primal variables z k, for every k F. Since the reduced costs are ρ k = α k m ˆv la lk 0, for every k F, we conclude that ẑ = (ẑ F, 0) is optimal for Program (15). This proposition shows by construction how to define extended vectors ˆξ i = (ˆξ F, ˆξ F i ) such that ẑ i = (ẑf i, 0), for i = 0, 1 are still optimal in the following linear programs min ˆξ i z min G(ẑ i ) + A i (z ẑ i ) 0, i=0,1 z 0, z j = i, β (16) whose feasible regions are larger than the set P {x: x j = i}, respectively. Since z F 0, for every z P j, then ˆξ = (ˆξ F, max i=0,1 (ˆξ i F )) 6
7 is a subgradient of f at ˆx such that ˆξx β, for every x P j. Moreover, since ˆξ x < β then we have found a separating hyperplane. Now, we assume that one of the sets in (11) is empty. Thus, Program (10) is solved as a standard convex program because its feasible region is defined by the nonempty set only. However, the fact that one of the sets in (11) is empty does not imply that the same has to occur in (5), when the variables x F are no longer fixed. Proposition 3 below describes in generic terms how to define the extended vector ˆξ i = (ˆξ F, ˆξ F i ) so that ˆξ i z β, for every z P {x: x j = i}, when the set P {x: x F = 0, x j = i} is empty. Proposition 3 Let F be an index set and F = {1,..., n} \ F. For a given arbitrary vector α F, let ẑ F be the optimal value of the following linear program: min α F z F A F z F b + ˆte, l F z F u F, (17) where l F and u F are the, possibly infinite, lower and upper bounds, A F IR m F, b, e IR m where e is a vector of all-ones, and ˆt min{t: A F z F b + te, l F z F u F } > 0. Then, for any extended matrix A = [A F, A F ] IR m n and scalar β there is a closed-form extended vector α = (α F, α F ) such that αz β, for every z such that Az b, l F z F u F, z F 0. Proof: First, consider the linear program that defines ˆt. Let (ˆt, z) be an optimal solution and ŵ be the optimal dual multipliers associated with the matrix constraints. Then, ˆt = ŵb + ˆγ F z F, (18) where ˆγ k = 0 m ŵla lk is the reduced cost associated with the variable z k, for each k F. Now, consider Program (17) and let ˆv be the optimal dual multipliers associated with the matrix constraints. Then, α F ẑ F = ˆv ( b + ˆte ) + ˆρ F ẑ F α F ẑ F ˆtˆve = ˆvb + ˆρ F ẑ F, (19) where ˆρ k = α k m ˆv la lk is the reduced cost associated with the variable z k, for each k F. If β α F ẑ F ˆtˆve then define α k = max(0, m ˆv la lk ), for every k F. For every z such that Az b, l F z F u F, z F 0 we have that αz = α F z F + α F z F ( ) m ˆρ k + ˆv l a lk z k + ( m ) ˆv l a lk z k (20) k F k F = ˆρ F z F + ˆvAz ˆρ F z F + ˆvb (21) ˆρ F ẑ F + ˆvb (22) = α F ẑ F ˆtˆve (23) β, (24) where the inequality (20) follows from the definition of ˆρ F, the definition of α F and the fact that z F 0; the inequality (21) follows from the fact that ˆv 0 and Az b; the inequality 7
8 (22) follows the fact that ˆρ k (z k ẑ k ) 0, for every k F, which is consequence of the values of the reduced costs at optimality; the inequality (23) follows from (19); and finally the inequality (24) holds by hypothesis. If β > α F ẑ F ˆtˆve then a similar formula works but we need to increase ˆv by a suitable positive multiplier of ŵ. Observe that ẑ F is feasible for the linear program that defines ˆt and so, from (18), we have that ˆt ŵb + ˆγ F ẑ F, or equivalently, β ( α F ẑ F ˆtˆve ) δŵb + δˆγ F ẑ F, (25) where δ = ( β ( α F ẑ F ˆtˆve )) /ˆt > 0. Now, define α k = max (0, m (ˆv + δŵ) l a lk), for every k F. For every z such that Az b, l F z F u F, z F 0 we have that αz = α F z F + α F z F ( ) m ˆρ k + ˆv l a lk z k + ( m ) (ˆv + δŵ) l a lk z k (26) k F k F = ˆρ F z F + (ˆv + δŵ) Az + δˆγ F z F ˆρ F z F + ˆvb + δŵb + δˆγ F z F (27) ˆρ F ẑ F + ˆvb + δŵb + δˆγ F ẑ F (28) ˆρ F ẑ F + ˆvb + ( β ( α F ẑ F ˆtˆve )) (29) = β (30) where the inequality (26) follows from the definition of ˆρ F, the definition of α F and the fact that z F 0; the inequality (27) follows from the fact that ˆv + δŵ 0 and Az b; the inequality (28) follows the fact that ˆρ k (z k ẑ k ) 0 and ˆγ k (z k ẑ k ) 0, for every k F, which is consequence of the values of the reduced costs at optimality; the inequality (29) follows from (25); and finally the equality (30) is a consequence of (19). This result can be easily generalized to a situation in which a distinct t variable occurs for each constraint. This is in fact the usual procedure with most phase-one implementations of the Simplex algorithm for linear programs. When solving a nonlinear program whose feasible region P {z: z F = 0, z j = i} is empty, most standard nonlinear programming algorithms are not ready to provide some point (ˆt, z F i ) that solves the following program min t G(z F, 0) te, z k 0, k F z j = i, and in this way proving infeasibility. In fact, it may occur that what seems to be an infeasible problem is just a numerical difficulty of meeting the constraints to a desired accuracy. The solution of the program (31) provides a verification of infeasibility and, as saw in the proof of Proposition 3, the dual variables that may be required for the lifting procedure. Since the Slater condition holds, the optimal solution (ˆt, ẑf i ) of Program (31) also solves the following linear program defined by a suitable matrix A i G(ẑF i, 0), (31) min t G(ẑF i, 0) + Ai (z F ẑf i, 0) te, z k 0, k F, z j = i, (32) 8
9 Then, to complete the lifting procedure we just have to solve min ˆξ F z F G(ẑF i, 0) + Ai (z F ẑf i, 0) ˆte, z k 0, k F, z j = i, (33) and, as explained in the proof of Proposition 3 we are now able to define ˆξ i = (ˆξ F, ˆξ F i ) f(ˆx) such that ˆξ i z β, for every z P {z: z j = i}. Since z F 0, for every z P j, then again ˆξ = (ˆξ F, max (ˆξ F i )) i=0,1 is a subgradient of the function f at ˆx such that ˆξx β, for every x P j. Moreover, since ˆξ x < β then we have found a separating hyperplane. References [1] Egon Balas, Sebastián Ceria, and Gérard Cornuéjols. A lift-and-project cutting plane algorithm for mixed 0-1 programs. Math. Programming, 58(3, Ser. A): , [2] E. Balas, S. Ceria, and G. Cornuejols. Mixed 0-1 programming by lift-and-project in a branch-and-cut framework. Management Science, 42(9): , Sep [3] M. Duran and I. Grossmann. An outer-approximation algorithm for a class of mixedinteger nonlinear programs. Mathematical Programming, 36: , [4] C. Floudas. Nonlinear and Mixed-integer Optimization. Oxford University Press, [5] B. Gavish. Topological design of computer communication networks - the overall design problem. European Journal of Operational Research, 58: , [6] J. Soares. Disjunctive Convex Optimization. PhD thesis, Graduate School of Business, Columbia Univeristy, Jun [7] Robert A. Stubbs and Sanjay Mehrotra. A branch-and-cut method for 0-1 mixed convex programming. Math. Program., 86(3, Ser. A): , [8] M. Turkay and I. Grossmann. Disjunctive programming techniques for the optimization of process systems with discontinuous investment costs - multiple size regions. Industrial & Engineering Chemistry Research, 35: ,
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 J. Zhang Institute of Applied Mathematics, Chongqing University of Posts and Telecommunications, Chongqing
2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
Convex analysis and profit/cost/support functions
CALIFORNIA INSTITUTE OF TECHNOLOGY Division of the Humanities and Social Sciences Convex analysis and profit/cost/support functions KC Border October 2004 Revised January 2009 Let A be a subset of R m
4.6 Linear Programming duality
4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal
Mathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION
1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of
3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the
Chapter 13: Binary and Mixed-Integer Programming
Chapter 3: Binary and Mixed-Integer Programming The general branch and bound approach described in the previous chapter can be customized for special situations. This chapter addresses two special situations:
Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725
Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
Linear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
Date: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
Integrating Benders decomposition within Constraint Programming
Integrating Benders decomposition within Constraint Programming Hadrien Cambazard, Narendra Jussien email: {hcambaza,jussien}@emn.fr École des Mines de Nantes, LINA CNRS FRE 2729 4 rue Alfred Kastler BP
CHAPTER 9. Integer Programming
CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral
10. Proximal point method
L. Vandenberghe EE236C Spring 2013-14) 10. Proximal point method proximal point method augmented Lagrangian method Moreau-Yosida smoothing 10-1 Proximal point method a conceptual algorithm for minimizing
Discrete Optimization
Discrete Optimization [Chen, Batson, Dang: Applied integer Programming] Chapter 3 and 4.1-4.3 by Johan Högdahl and Victoria Svedberg Seminar 2, 2015-03-31 Todays presentation Chapter 3 Transforms using
BANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
Some representability and duality results for convex mixed-integer programs.
Some representability and duality results for convex mixed-integer programs. Santanu S. Dey Joint work with Diego Morán and Juan Pablo Vielma December 17, 2012. Introduction About Motivation Mixed integer
1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.
Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S
1. Prove that the empty set is a subset of every set.
1. Prove that the empty set is a subset of every set. Basic Topology Written by Men-Gen Tsai email: [email protected] Proof: For any element x of the empty set, x is also an element of every set since
Big Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
Proximal mapping via network optimization
L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
Duality in Linear Programming
Duality in Linear Programming 4 In the preceding chapter on sensitivity analysis, we saw that the shadow-price interpretation of the optimal simplex multipliers is a very useful concept. First, these shadow
Noncommercial Software for Mixed-Integer Linear Programming
Noncommercial Software for Mixed-Integer Linear Programming J. T. Linderoth T. K. Ralphs December, 2004. Revised: January, 2005. Abstract We present an overview of noncommercial software tools for the
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
Optimization Theory for Large Systems
Optimization Theory for Large Systems LEON S. LASDON CASE WESTERN RESERVE UNIVERSITY THE MACMILLAN COMPANY COLLIER-MACMILLAN LIMITED, LONDON Contents 1. Linear and Nonlinear Programming 1 1.1 Unconstrained
Continued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
Separation Properties for Locally Convex Cones
Journal of Convex Analysis Volume 9 (2002), No. 1, 301 307 Separation Properties for Locally Convex Cones Walter Roth Department of Mathematics, Universiti Brunei Darussalam, Gadong BE1410, Brunei Darussalam
4: SINGLE-PERIOD MARKET MODELS
4: SINGLE-PERIOD MARKET MODELS Ben Goldys and Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2015 B. Goldys and M. Rutkowski (USydney) Slides 4: Single-Period Market
No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics
No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results
What is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
Applied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach
MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical
1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
constraint. Let us penalize ourselves for making the constraint too big. We end up with a
Chapter 4 Constrained Optimization 4.1 Equality Constraints (Lagrangians) Suppose we have a problem: Maximize 5, (x 1, 2) 2, 2(x 2, 1) 2 subject to x 1 +4x 2 =3 If we ignore the constraint, we get the
THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING
THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING 1. Introduction The Black-Scholes theory, which is the main subject of this course and its sequel, is based on the Efficient Market Hypothesis, that arbitrages
CONSTANT-SIGN SOLUTIONS FOR A NONLINEAR NEUMANN PROBLEM INVOLVING THE DISCRETE p-laplacian. Pasquale Candito and Giuseppina D Aguí
Opuscula Math. 34 no. 4 2014 683 690 http://dx.doi.org/10.7494/opmath.2014.34.4.683 Opuscula Mathematica CONSTANT-SIGN SOLUTIONS FOR A NONLINEAR NEUMANN PROBLEM INVOLVING THE DISCRETE p-laplacian Pasquale
Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.
1. Introduction Linear Programming for Optimization Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1.1 Definition Linear programming is the name of a branch of applied mathematics that
FIRST YEAR CALCULUS. Chapter 7 CONTINUITY. It is a parabola, and we can draw this parabola without lifting our pencil from the paper.
FIRST YEAR CALCULUS WWLCHENW L c WWWL W L Chen, 1982, 2008. 2006. This chapter originates from material used by the author at Imperial College, University of London, between 1981 and 1990. It It is is
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
Optimization Methods in Finance
Optimization Methods in Finance Gerard Cornuejols Reha Tütüncü Carnegie Mellon University, Pittsburgh, PA 15213 USA January 2006 2 Foreword Optimization models play an increasingly important role in financial
GENERALIZED INTEGER PROGRAMMING
Professor S. S. CHADHA, PhD University of Wisconsin, Eau Claire, USA E-mail: [email protected] Professor Veena CHADHA University of Wisconsin, Eau Claire, USA E-mail: [email protected] GENERALIZED INTEGER
An optimal transportation problem with import/export taxes on the boundary
An optimal transportation problem with import/export taxes on the boundary Julián Toledo Workshop International sur les Mathématiques et l Environnement Essaouira, November 2012..................... Joint
Nonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
Introduction: Models, Model Building and Mathematical Optimization The Importance of Modeling Langauges for Solving Real World Problems
Introduction: Models, Model Building and Mathematical Optimization The Importance of Modeling Langauges for Solving Real World Problems Josef Kallrath Structure of the Lecture: the Modeling Process survey
Inner product. Definition of inner product
Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product
LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method
LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method Introduction to dual linear program Given a constraint matrix A, right
Principles of demand management Airline yield management Determining the booking limits. » A simple problem» Stochastic gradients for general problems
Demand Management Principles of demand management Airline yield management Determining the booking limits» A simple problem» Stochastic gradients for general problems Principles of demand management Issues:»
Optimization Modeling for Mining Engineers
Optimization Modeling for Mining Engineers Alexandra M. Newman Division of Economics and Business Slide 1 Colorado School of Mines Seminar Outline Linear Programming Integer Linear Programming Slide 2
Approximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three
Arrangements And Duality
Arrangements And Duality 3.1 Introduction 3 Point configurations are tbe most basic structure we study in computational geometry. But what about configurations of more complicated shapes? For example,
Optimal shift scheduling with a global service level constraint
Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The
Linear Programming in Matrix Form
Linear Programming in Matrix Form Appendix B We first introduce matrix concepts in linear programming by developing a variation of the simplex method called the revised simplex method. This algorithm,
Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011
Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011 A. Miller 1. Introduction. The definitions of metric space and topological space were developed in the early 1900 s, largely
Adaptive Linear Programming Decoding
Adaptive Linear Programming Decoding Mohammad H. Taghavi and Paul H. Siegel ECE Department, University of California, San Diego Email: (mtaghavi, psiegel)@ucsd.edu ISIT 2006, Seattle, USA, July 9 14, 2006
Notes from February 11
Notes from February 11 Math 130 Course web site: www.courses.fas.harvard.edu/5811 Two lemmas Before proving the theorem which was stated at the end of class on February 8, we begin with two lemmas. The
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS KEITH CONRAD 1. Introduction The Fundamental Theorem of Algebra says every nonconstant polynomial with complex coefficients can be factored into linear
GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1
(R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter [email protected] February 20, 2009 Notice: This work was supported by the U.S.
α = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Several Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
Metric Spaces. Chapter 1
Chapter 1 Metric Spaces Many of the arguments you have seen in several variable calculus are almost identical to the corresponding arguments in one variable calculus, especially arguments concerning convergence
An interval linear programming contractor
An interval linear programming contractor Introduction Milan Hladík Abstract. We consider linear programming with interval data. One of the most challenging problems in this topic is to determine or tight
Minimizing costs for transport buyers using integer programming and column generation. Eser Esirgen
MASTER STHESIS Minimizing costs for transport buyers using integer programming and column generation Eser Esirgen DepartmentofMathematicalSciences CHALMERS UNIVERSITY OF TECHNOLOGY UNIVERSITY OF GOTHENBURG
PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5.
PUTNAM TRAINING POLYNOMIALS (Last updated: November 17, 2015) Remark. This is a list of exercises on polynomials. Miguel A. Lerma Exercises 1. Find a polynomial with integral coefficients whose zeros include
The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line
The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line D. Bongiorno, G. Corrao Dipartimento di Ingegneria lettrica, lettronica e delle Telecomunicazioni,
The Ideal Class Group
Chapter 5 The Ideal Class Group We will use Minkowski theory, which belongs to the general area of geometry of numbers, to gain insight into the ideal class group of a number field. We have already mentioned
Stochastic Inventory Control
Chapter 3 Stochastic Inventory Control 1 In this chapter, we consider in much greater details certain dynamic inventory control problems of the type already encountered in section 1.3. In addition to the
Linear Programming I
Linear Programming I November 30, 2003 1 Introduction In the VCR/guns/nuclear bombs/napkins/star wars/professors/butter/mice problem, the benevolent dictator, Bigus Piguinus, of south Antarctica penguins
CHAPTER 11: BASIC LINEAR PROGRAMMING CONCEPTS
Linear programming is a mathematical technique for finding optimal solutions to problems that can be expressed using linear equations and inequalities. If a real-world problem can be represented accurately
(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties
Lecture 1 Convex Sets (Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties 1.1.1 A convex set In the school geometry
Solving Systems of Linear Equations
LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how
Systems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
Online Convex Optimization
E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting
Nonlinear Optimization: Algorithms 3: Interior-point methods
Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris [email protected] Nonlinear optimization c 2006 Jean-Philippe Vert,
Instituto de Engenharia de Sistemas e Computadores de Coimbra Institute of Systems Engineering and Computers INESC Coimbra
Instituto de Engenharia de Sistemas e Computadores de Coimbra Institute of Systems Engineering and Computers INESC Coimbra João Clímaco and Marta Pascoal A new method to detere unsupported non-doated solutions
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Section 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics
Undergraduate Notes in Mathematics Arkansas Tech University Department of Mathematics An Introductory Single Variable Real Analysis: A Learning Approach through Problem Solving Marcel B. Finan c All Rights
The Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
Solutions Of Some Non-Linear Programming Problems BIJAN KUMAR PATEL. Master of Science in Mathematics. Prof. ANIL KUMAR
Solutions Of Some Non-Linear Programming Problems A PROJECT REPORT submitted by BIJAN KUMAR PATEL for the partial fulfilment for the award of the degree of Master of Science in Mathematics under the supervision
ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING. 0. Introduction
ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING Abstract. In [1] there was proved a theorem concerning the continuity of the composition mapping, and there was announced a theorem on sequential continuity
A simple criterion on degree sequences of graphs
Discrete Applied Mathematics 156 (2008) 3513 3517 Contents lists available at ScienceDirect Discrete Applied Mathematics journal homepage: www.elsevier.com/locate/dam Note A simple criterion on degree
Batch Production Scheduling in the Process Industries. By Prashanthi Ravi
Batch Production Scheduling in the Process Industries By Prashanthi Ravi INTRODUCTION Batch production - where a batch means a task together with the quantity produced. The processing of a batch is called
ALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!
Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
