Preprint Ayalew Getachew Mersha, Stephan Dempe Feasible Direction Method for Bilevel Programming Problem ISSN

Size: px
Start display at page:

Download "Preprint 2009-02. Ayalew Getachew Mersha, Stephan Dempe Feasible Direction Method for Bilevel Programming Problem ISSN 1433-9307"

Transcription

1 Fakultät für Mathematik und Informatik Preprint Ayalew Getachew Mersha, Stephan Dempe Feasible Direction Method for Bilevel Programming Problem ISSN

2 Ayalew Getachew Mersha, Stephan Dempe Feasible Direction Method for Bilevel Programming Problem TU Bergakademie Freiberg Fakultät für Mathematik und Informatik Institut für Numerische Mathematik und Optimierung Akademiestr. 6 (Mittelbau) FREIBERG

3 ISSN Herausgeber: Herstellung: Dekan der Fakultät für Mathematik und Informatik Medienzentrum der TU Bergakademie Freiberg

4 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM AYALEW GETACHEW MERSHA AND STEPHAN DEMPE Abstract. In this paper, we investigate the application of feasible direction method for an optimistic nonlinear bilevel programming problem. The convex lower level problem of an optimistic nonlinear bilevel programming problem is replaced by relaxed KKT conditions. The feasible direction method developed by Topkis and Veinott [22] is applied to the auxiliary problem to get a Bouligand stationary point for an optimistic bilevel programming problem. (1.1) 1. Introduction In this paper we consider the following bilevel programming problem. with F (x, y) min x G(x) 0, y Ψ(x) (1.2) y Ψ(x) := arg min{f(x, y) : g(x, y) 0} y where F, f : R n+m R, g : R n+m R p, G : R n R l are at least twice continuously differentiable functions. The following assumption is made throughout this paper. Assumption 1. f(x,.), g i (x,.), i = 1,..., p are convex for every x. If the solution of the lower level problem (1.2) corresponding to any parameter x is not unique, the upper level problem (1.1) is not a well defined optimization problem. In this case the bilevel programming problem may not be solvable. This situation is depicted by examples constructed in [2] and [4, example on page 121]. The following Lemma substantiates this claim. Lemma 1.1. [3] If Ψ(x) is not a singleton for all parameters x, the leader may not achieve his infimum objective function value. To overcome such an unpleasant situation there are three strategies available for the leader. The first strategy is to replace min with inf in the formulation of problem (1.1) and define ɛ optimal solutions. The second strategy is to allow cooperation between the two players: the upper level decision maker, the leader, and the lower level decision maker, the follower. This resulted in the so called optimistic or weak bilevel programming problem. The third is a conservative strategy. In this case the Work of the first author was supported by DAAD(Deutscher Akademischer Austausch Dienst) with a scholarship grant. 1

5 2 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE leader is forced to bound the damage caused by a follower s unfavorable choice. Define the following sets, M := {(x, y) : G(x) 0, g(x, y) 0}, M(X) := {x R n : y s.t. (x, y) M}, M(x) := {y R m : g(x, y) 0, } I(x, y ) := {i : g i (x, y ) = 0}. Definition 1.2. The lower level problem (1.2) is said to satisfy the Mangasarian- Fromowitz Constraint Qualification (MFCQ) at (x, y ), y Ψ(x ) if { r R m r y g i (x, y ) < 0, i I(x, y ) }. The Lagrangian function for the lower level problem is given by Consider the set of Lagrange multipliers Let L(x, y, λ) := f(x, y) + λ g(x, y). Λ(x, y) := {λ : λ 0, λ g(x, y) = 0, y L(x, y, λ) = 0}. J(λ) := {j : λ j > 0}. Definition 1.3. The lower level problem (1.2) is said to satisfy a strong sufficient optimality condition of second order (SSOC) at a point (x, y ) if for each λ Λ(x, y ) and for every nonzero element of the set { r R m r y g i (x, y ) = 0, i J(λ) } we have r T 2 yyl(x, y, λ)r > 0. Definition 1.4. The constant rank constraint qualification (CRCQ) is valid for problem (1.2) at a point (x, y ) if there is an open neighborhood W ε (x, y ), ε > 0 of (x, y ) such that for each subsets I I(x, y ), the family of gradient vectors { y g i (x, y) : i I} has the same rank for all (x, y) W ε (x, y ). Assume the following is satisfied: Assumption 2. The set M is nonempty and compact. 2. Optimistic bilevel programming If the follower agrees to cooperate with the leader in the decision making process, we get the weak bilevel programming problem: (2.1) where ϕ (x) min x s.t. G(x) 0 (2.2) ϕ (x) := min{f (x, y) : y Ψ(x)} y As can be seen from the formulation, this problem is weaker that the original bilevel programming problem.

6 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 3 Definition 2.1. [4] A point (x, y ) R n R m is a local solution of problem (2.1), (2.2) if x {x : G(x) 0}, y Ψ(x ) with F (x, y) F (x, y ), y Ψ(x ) and there exists a sufficiently small positive number ɛ > 0, with ϕ (x ) ϕ (x), x B ɛ (x ). If ɛ = then (x, y ) is a global solution. Now we further require to replace the optimistic bilevel programming problem with another version of hierarchical problem. Consider the following problem (2.3) min{f (x, y) : G(x) 0, y Ψ(x)} x,y with the properties F, g, h explained at the beginning of this paper. The relation between problem (2.3) and (2.1), (2.2) follows. If Assumption 2 and MFCQ satisfied, the global optimal solutions of (2.3) and (2.1), (2.2) coincide [4]. With respect to a local optimal solution, Dutta and Dempe [6] showed that every local optimal solution of the optimistic bilevel problem is a local optimal solution of problem (2.3) as the following result shows. Theorem 2.2. [6] Let the lower level problem (1.2) satisfy Assumption 2 and MFCQ and let x be a local optimal solution of problem (2.1) with corresponding ȳ arg min{f ( x, y) : y Ψ( x)}. y Then ( x, ȳ) is a local optimal solution of problem (2.3). Problem (2.3) is used to derive the necessary optimality condition for the optimistic bilevel programming problem (2.1), (2.2) because it can be reformulated to a single level problem provided that the lower level problem is convex and satisfies the Mangasarian-Fromowitz constraint qualification. 3. Piecewise continuously differentiable functions Definition 3.1. A function y : R n R m is a P C 1 function at x if it is continuous and there exists an open neighborhood V of x and a finite number of continuously differentiable functions y i : V R m, i = 1,..., b such that y(x) {y 1 (x),..., y b (x)} for all x V. The function y is a P C 1 function on some open set O provided it is a P C 1 function at every point x O. Theorem 3.2. [18] Let the lower level problem (1.2) at x = x be convex satisfying (MFCQ), (SSOC) and (CRCQ) at a stationary solution y. Then, the locally uniquely determined function y(x) Ψ(x), and hence F (x, y(x)) is a P C 1 function. It has been shown in [11] that P C 1 functions are locally Lipschitz continuous. The directional differentiability of the solution function y(x) and hence F (x, y(x)) is by product of the result in [16]. The following result shows that P C 1 functions are directional differentiable and locally Lipschitz continuous. Theorem 3.3. [4, Theorem 4.8] Let f : R n R be a P C 1 function. Then f is directionally differentiable and locally Lipschitz continuous.

7 4 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE Define K(i) := {x : y(x) = y i (x)} and the index set of essentially active functions I s y(x) := {i : y(x) = y i (x), x cl int K(i)} {1,..., b}. The usual tangent cone to the set K(i) is defined by { T K(i) (x) := d : x k x, x k K(i) k, t k 0 such that x } k x d. t k Once the set of essentially active selection function is known, it is shown in Scholtes [20, Proposition A.4.1] that the generalized Jacobian of y( ) is given by y(x) = conv{ y i (x) : i I s y(x)}. The generalized directional derivative of solution function y( ) is given by y (x; d) = sup{ d, z : z y(x)}. 4. Optimality It is known that a point (x, y ) is said to be a local optimal solution for a given optimization problem if it is the best point in an small neighborhood of (x, y ). Definition 4.1. [4]: A feasible point (x, y ) is said to be a local optimal solution of problem (2.3) if there exists an open neighborhood U(x, y ) such that F (x, y ) F (x, y), (x, y) U(x, y ), y Ψ(x), G(x) 0. Remember that the validity of conditions MFCQ, SSOC and CRCQ in the lower level problem implies that the solution function y(x) is a P C 1 function: a continuous selection of continuously differentiable functions y 1,..., y b. The so called Clarke and Bouligand stationarity are defined as: Definition 4.2. [20] Let the lower level problem satisfy MFCQ, SSOC and CRCQ at a point (x, y ), y Ψ(x ). Such a point (x, y ) is said to be a Clarke stationary point for the optimistic bilevel programming problem if d, i I s y(x ) such that d T K(i) (x ), x F (x, y )d + y F (x, y ) y i (x )d 0. Definition 4.3. Let conditions MFCQ, SSOC and CRCQ be satisfied for the lower level problem under consideration. A point (x, y ) is said to be a Bouligand stationary point for problem (1.1), (1.2) if (4.1) x F (x, y )d + y F (x, y )y (x ; d) 0 Or in other words j I s y(x ) we have (4.2) d : d G i (x ) 0, i : G i (x ) = 0. x F (x, y )d + y F (x, y ) y j (x )d 0 d : d G i (x ) 0, i : G i (x ), where I s y(x ) is the set of essentially active indices.

8 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 5 If the lower level problem (1.2) is convex satisfying MFCQ, it can be replaced by its sufficient and necessary conditions and hence we get the following problem: (4.3) F (x, y) min x,y,λ G(x) 0 y L(x, y, λ) = 0 g(x, y) 0 λ 0 λ i g i (x, y) = 0, i = 1,..., p. Problems (1.1), (1.2) and problem (4.3) are equivalent if the lower level problem is convex satisfying MFCQ at each feasible point, the solution function is uniquely determined and strongly stable in the sense of Kojima [12], and a global solution is considered. Without the convexity assumption problem (4.3) has a larger feasible set. Hence even the global solution of the aforementioned problems may not coincide. The following simple relation can be established between problem (1.1),(1.2) and problem (4.3). Theorem 4.4. Let ( x, ȳ, λ) be a feasible point of (4.3). If the lower level problem (1.2) is convex satisfying MFCQ then the primal part, ( x, ȳ) is feasible in (1.1). Proof. Let ( x, ȳ, λ) be a feasible point of problem (4.3). Then for x, G( x) 0, validity of MFCQ in the lower level problem, the pair (ȳ, λ) solves the system y L( x, y, λ) = 0 g( x, y) 0 λ 0 λ i g i ( x, y) = 0, i = 1,..., p. Due to the convexity of the parametric lower level problem the above system is both sufficient and necessary condition for ȳ Ψ( x). Hence ( x, ȳ) is a feasible point for problem (1.1). Problem (4.3) is one element of special classes of mathematical programming problems called Mathematical Programs with Equilibrium Constraints (MPEC). MPECs are special classes of nonlinear programming problems with equilibrium conditions in the constraints. This means an expression of this type: θ 1 (x) 0, θ 2 (x) 0, θ i 1(x)θ i 2(x) = 0, i = 1,..., m appear in the constraints, where θ k : R n R m, k = 1, 2 are differentiable functions. MPECs are considered as one of the ill conditioned problems in optimization theory due to the existence of the complementarity term. Neither known regularity conditions like MFCQ, LICQ and Slater Constraint Qualification nor their nonsmooth variants are satisfied at any point of the feasible set of MPEC [19]. MPECs are used to model many economics and mechanic applications[17]. There are a number of investigations done during the last couples of decades. Consequently, many results were reported. The notable contribution we need to mention here is the successful attempt to solve MPEC as an ordinary nonlinear programming problem [7, 21]. Also a number of papers address question of deriving optimality conditions (see for instance [19]). The interested reader can consult the monographs [14, 17].

9 6 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE Scholtes [21] investigated a general version of problem (4.3) via a regularization method. He relaxed the complementary constraint and then solved the resulting parameterized nonlinear programming problem for decreasing sequence of the relaxation parameter. The validity of MFCQ in the lower level problem implies the multiplier set is a nonempty convex compact polyhedron[10]. However, it is not necessarily a singleton. To obtain a stationary solution of an optimistic bilevel programming problem with convex lower level problem, we have to show the nonexistence of descent directions at the current incumbent stationary solution of the corresponding MPEC. As it is clearly shown in [4, Theorem 5.15], a stationary solution for the corresponding MPEC may not be a local solution for the optimistic bilevel programming problem even under strong assumptions: convexity and validity of regularity conditions for the lower level problem. We approach the ill-conditioned problem (4.3) by relaxing the stationarity condition y L(x, y, λ) = 0 and the complementarity condition λ i g i (x, y) = 0, i = 1, 2,..p. Consequently, we consider the following parametric auxiliary problem NLP(ɛ) for ɛ > 0: (4.5) F (x, y) min x,y,λ G(x) 0 y L(x, y, λ) ɛ g(x, y) 0 λ 0 λ i g i (x, y) ɛ, i = 1,..., p Before we proceed, we need to establish the relation between problem (1.1) and problem (4.5). Theorem 4.5. Let the parametric lower level problem (1.1) be convex satisfying MFCQ. If (x, y ) is a local optimal solution for problem (1.1), (1.2) then (x, y, λ ) is local optimal solution for problem (4.5) for ɛ = 0 and for all λ Λ(x, y ) := {λ : λ 0, λ T g(x, y ) = 0, y L(x, y, λ) = 0}. Proof. Assume that there exists λ Λ(x, y ) such that (x, y, λ) is not a local optimal solution of problem (4.5) with ɛ = 0. Then there exists a sequence {(x k, y k, λ k )} converging to (x, y, λ) with (4.6) F (x k, y k ) < F (x, y ) G(x k ) 0 g(x k, y k ) 0 y L(x k, y k, λ k ) = 0 λ k Λ(x k, y k ). Since the KKT conditions are necessary and sufficient for problem (1.2), we have y k Ψ(x k ), k. This implies (x k, y k ) is feasible in (1.1), (1.2). Therefore (x, y ) is not a local optimal solution of problem (1.1), (1.2) which is a contradiction to the hypothesis of the theorem. We make the following assumption for the upper level constraint.

10 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 7 Assumption 3. The upper level constraint qualification is said to be satisfied at a point x if {d : d G i (x) < 0, i : G i (x) = 0} The existence of descent direction at a point in the neighborhood of a nonstationary point is shown in the next theorem. Theorem 4.6. Let ( x, ȳ, λ) be a feasible solution of problem (4.5) for ɛ = 0. Let the lower level problem (1.2) be convex satisfying MFCQ, SSOC and CRCQ. Assume that the upper level constraint set satisfies Assumption 3. If ( x, ȳ) is not a Bouligand stationary point of the problem (1.1) then there exists d with d x G i ( x) < 0, i : G i ( x) = 0 and δ > 0 such that for some j = 1,..., w. x F (x, y)d + y F (x, y) y j (x)d < 0, (x, y) B δ ( x, ȳ) Proof. Let the lower level problem be convex satisfying MFCQ, SSOC and CRCQ. Let ( x, ȳ, λ) be a feasible solution of problem (4.5) for ɛ = 0. If ( x, ȳ) is not a Bouligand stationary point of the problem (1.1) then there exists a direction d satisfying and G i ( x) d 0, i : G i ( x) = 0 x F ( x, ȳ)d + y F ( x, ȳ)y ( x; d) < 0. Due to MFCQ, SSOC and CRCQ the function y( ) {y 1 ( ),..., y w ( )} is a P C 1 function and hence directionally differentiable (Theorem 3.3). Therefore, there is j {1,..., w} such that Hence for this j we have y ( x; d) = y j ( x)d. x F ( x, ȳ)d + y F ( x, ȳ) y j ( x)d < 0. Since y j ( ) is continuously differentiable there exists δ > 0 such that x F (x, y)d + y F (x, y) y j (x)d < 0, (x, y) B δ ( x, ȳ). The following theorem is one of the key results that relates the MPEC (Problem 4.3) and its relaxed nonlinear programming problem NLP(ɛ): Theorem 4.7. Let {(x k, y k, λ k )} be a local optimal point of problem (4.5) for ɛ = ɛ k where ɛ k 0. Let (x, y, λ ) be a limit point of the sequence {(x k, y k, λ k )}. If { (d, r) G } i(x )d < 0, i : G i (x ) = 0, (4.7) g j (x, y )(d, r) < 0, j : g j (x, y ) = 0 holds, then (x, y, λ ) is a Bouligand stationary solution for problem (4.5) for ɛ = 0.

11 8 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE Proof. Let (x, y, λ ) be not a Bouligand stationary solution for problem (4.5) for ɛ = 0. Then by definition the system x F (x, y )d + y F (x, y )r < 0 x G i (x )d 0, i : G i (x ) = 0 ( y L(x, y, λ ))(d, r, γ) = 0 x g j (x, y )d + y g j (x, y )r = 0, j : g j (x, y ) = 0 γ i 0, i : λ i = 0 (λ i g i (x, y ))(d, r, γ) = 0 has solution (d, r, γ). Let ɛ > 0 be arbitrary small. Due to the assumption (4.7) there is ( ˆd, ˆr, ˆγ) x F (x, y ) ˆd + y F (x, y )ˆr < 0 G i (x ) + x G i (x ) ˆd < 0, i ( y L(x, y, λ ))( ˆd, ˆr, ˆγ) ɛ/2 g j (x, y ) + x g j (x, y ) ˆd + y g j (x, y )ˆr < 0, j : g j (x, y ) = 0 λ i + ˆγ i > 0, i (λ i g i (x, y ))( ˆd, ˆr, ˆγ) ɛ/2 where (d, r) ( ˆd, ˆr) sufficiently small. Since {(x k, y k, λ k )} converges to (x, y, λ ) there exist k > 0 such that the following holds for k k x F (x k, y k ) ˆd + y F (x k, y k )ˆr < 0 G i (x k ) + x G i (x k ) ˆd < 0, i ( y L(x k, y k, λ k ))( ˆd, ˆr, ˆγ) ɛ/2 g j (x k, y k ) + x g j (x k, y k ) ˆd + y g j (x k, y k )ˆr < 0, j : g j (x, y ) = 0 λ i + ˆγ i > 0, i ((λ i ) k g i (x k, y k ))( ˆd, ˆr, ˆγ) ɛ/2 which implies (x k, y k, λ k ) is not a local solution of problem (4.5) resulting in a contradiction. The converse of the above theorem may not be true as the following simple example shows. Example 4.8. [21]: Consider the following simple example. min y s.t xy ɛ x, y 0 Clearly, every point on the positive x-axis is a local minimizer of NLP (0). But the relaxed problem NLP (ɛ), ɛ > 0 has no local solution. It is clear that the feasible set of NLP (ɛ) does not expand as ɛ 0. Hence the following theorem follows: Theorem 4.9. Let {(x k, y k, λ k )} be a sequence of local optimal solutions of problem (4.5) corresponding to ɛ k. If ɛ k 0 then there is a monotone increasing subsequence of the sequence {F (x k, y k )}.

12 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 9 Proof. If the sequence {ɛ k } converges to zero then there is a subsequence {k i } of {k} such that ɛ ki > ɛ ki+1, i. which implies S(ɛ ki ) S(ɛ ki+1 ), i where S(ɛ k ) is the feasible set of problem (4.5) corresponding to ɛ k. The theorem follows from the property of min operator. Those presented results are the base of convergence analysis of the algorithm given next. The verification of optimality or stationarity is based on the size of (4.8) (4.9) (4.10) Initialization: Initialize with (x, y ) {(x, y) : G(x) 0, g(x, y) 0}, k := 0, s < w < 0, δ, σ, αɛ (0, 1) Step 1: Choose a maximal set of indices I k {i : g i (x k, y k ) = 0} with { y g i (x k, y k ) = 0, i I k } is linearly independent. Go to Step 2. Step 2: Solve the following system of inequalities to find a multiplier: L y (x k, y k, λ) ɛ λ i g i (x k, y k ) ɛ, j = 1,..., p λ 0, λ j = 0, j / I k. Step 3 : Find direction a (d k, r k ) using the direction finding problem (4.11) (see the next page) with s k < s. If there is no solution goto Step 6 otherwise proceed to Step 4. Step 4 : Using Armijo rule, find a step size which satisfies the following: F (x k + td k, y k + tr k ) F (x k, y k ) + δts k G i (x k + td k ) 0, i = 1,..., l g j (x k + td k, y k + tr k ) 0, j = 1,..., p. Step 5: Set x k+1 = x k + t k d k, y k+1 = y k + t k r k, put k:=k+1 and goto step 1. Step 6: Put s := s /2. If s w goto Step 3 otherwise goto Step 7. Step 7: Choose new set I k (σ) {i : σ g i (x k, y k ) = 0} with { y g i (x k, y k ), i I k (σ)} is linearly independent. If all such I k (σ) are already tried then go to Step 9 otherwise go to step 8. Step 8: Solve the following optimization problem y L(x k, y k, λ) 2 min λ λ i = 0, i / I k (σ) goto Step 3. Step 9: Put ɛ := αɛ. If ɛ ɛ then stop other wise goto Step 2. Algorithm 4.1: Feasible direction algorithm applied to bilevel programming problem with convex lower level problem. the directional derivative. If the directional derivative is sufficiently small (in the absolute value), we compute an approximate Lagrange multiplier by considering the near active constraints in the lower level problem. The solution of problem (4.10) is uniquely determined and converges to the true multiplier λ Λ(x, y ) for the sequence {(x k, y k )} produced by the algorithm with limit point (x, y ), if I k (σ) {I : {j : λ j > 0} I I g(x, y ) with { y g i (x k, y k ) : i I k (σ)} is linearly independent. The main purpose of doing Step 7 is to exclude Clarke stationary

13 10 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE points that are not Bouligand stationary points. However, the choice of σ in Step 7 seems to be difficult. The direction in Step 3 of Algorithm 4.1 is found by solving the following linear optimization problem: (4.11a) (4.11b) (4.11c) (4.11d) (4.11e) (4.11f) (4.11g) (4.11h) (4.11i) (4.11j) (4.11k) (4.11l) s min d,r,s,γ xf (x k, y k )d + yf (x k, y k )r s xg i(x k )d + G i(x) s, i = 1,..., l yl(x k, y k, λ k ) + 2 xyl(x k, y k, λ k )d + 2 yyl(x k, y k, λ k )r + γ T yg(x k, y k ) ɛ yl(x k, y k, λ k ) 2 xyl(x k, y k, λ k )d 2 yyl(x k, y k, λ k )r γ T yg(x k, y k ) ɛ λ T g(x k, y k ) γ T g(x k, y k ) λ T xg(x k, y k )d λ T yg(x k, y k )r ɛ xg i(x k, y k )d + yg i(x k, y k )r = 0, i I k g j(x k, y k ) + xg j(x k, y k )d + yg j(x k, y k )r 0, j / I k λ γ 0 d 1 r 1 γ 0. Problem (4.11) has a negative optimal objective function value at all non stationary terms of the feasible sequence {(x k, y k )} provided that a suitable index set I k is chosen in Step 2 of the algorithm. Theorem Let the lower level problem be convex satisfying MFCQ, CRCQ and SSOC. Let ( x, ȳ, λ) be a feasible point of problem (4.3). If (x, y) B δ ( x, ȳ) {(x, y) : y Ψ(x)}, δ > 0, then there exists ɛ > 0 and I {i : g i ( x, ȳ) = 0}, with { y g i ( x, ȳ) : i I} linearly independent such that the following system is satisfied: (4.12) y L(x, y, λ) ɛ λ 0, λ i = 0, i / I. Proof. Since the lower level problem satisfies MFCQ, the Λ( x, ȳ) is nonempty and bounded for ȳ Ψ( x). Moreover, ȳ Ψ( x) and convexity implies that (4.13) y L( x, ȳ, λ) = 0 λ g( x, ȳ) = 0 λ 0 for λ Λ( x, ȳ). From this there exists an index set I such that y L( x, ȳ, λ) = 0 λ i = 0, i / I with {i : λ i > 0} I I g ( x, ȳ). Consider the sequence {(x k, y k )} B δ ( x, ȳ) {(x, y) : y Ψ(x)}

14 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 11 with {(x k, y k )} converging to ( x, ȳ). Fix such an index set I. Then we have y L(x k, y k, λ) = y f(x k, y k ) + λ i y g i (x k, y k ) i I = y f( x, ȳ) + λ i y g i ( x, ȳ) i I + ( y f( x, ȳ) + i I λ i y g i ( x, ȳ)((x k, y k ) ( x, ȳ) ) +o ((x k, y k ) ( x, ȳ) ) y f( x, ȳ) + λ i y g i ( x, ȳ) i I } {{ } =0 + ( y f( x, ȳ) + i I λ i y g i ( x, ȳ))((x k, y k ) ( x, ȳ) ) Put +o ((x k, y k ) ( x, ȳ) ). ɛ = ( y f( x, ȳ) + i I +o ((x k, y k ) ( x, ȳ) ). λ i g i ( x, ȳ)((x k, y k ) ( x, ȳ) ) The following result is adapted from a result reported in [5]. It shows that the sequence of step size parameters is bounded away from zero if the sequence of optimal function values of problem (4.11) converges to a non zero number. Theorem Let the lower level problem be convex satisfying MFCQ. Suppose {x k, y k, λ k, d k, r k, ɛ k, s k, t k, I k } be a sequence produced by the algorithm with ɛ k 0. If the sequence {x k, y k, λ k } converges to (x, y, λ ) and lim k s k = s < 0 then lim k t k > 0. Proof. Assume that aforementioned positive step size converges to zero. Then, the sequence of maximal possible step sizes τ k must tend to zero too. Define τ k by τ k = min{max{t : G i (x k + td k ) 0}, max{t : g j (x k + td k, y k + td k ) 0}}. Then there exists or i : G i (x k ) = G i (x k + τ k d k ) = 0 (4.14) j : g j (x k, y k ) = g j (x k + τ k d k, y k + τ k d k ) = 0. If the sequence {x k }, {d k } with d k = 1 converges respectively to x and d then the sequence { Gi(x k+τ k d k ) G i(x k ) τ k } converges to G i ( x) d. Observe that for the above taken i we have Gi(x k+τ k d k ) G i(x k ) τ k = 0. But G i (x k )d k s k for G i ( x) = 0. This implies G i (x k + τ k d k ) G i (x k ) 0 = lim = G i ( x) k,τ k 0 τ d s < 0 k which is a contradiction. The other case, namely for condition (4.14) can be shown analogously.

15 12 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE The next result shows that the directional derivative converges to zero. Theorem Let F be bounded below, the lower level problem be convex satisfying MFCQ, SSOC, CRCQ and the upper level constraint satisfy condition {d : d G i (x) < 0} =. Let {x k, y k, λ k, d k, r k, ɛ k, s k, t k, I k } be a sequence computed by the algorithm where ɛ k 0. Then the sequence {s k } has zero as the only accumulation point. Proof. Suppose there exist a subsequence that does not converge to zero. i.e it converges to a number say s < 0. Assume without loss of generality that such a subsequence be {s k }. From Theorem 4.11, we have lim k 0 t k > 0. But from Step 4 of the algorithm we have F (x k + t k d k, y k + t k d k ) F (x k, y k ) δs k t k, k which implies F (x k, y k ) as k. This is a contradiction to the fact that F is bounded below. Assume that λ is obtained from Step 2 of the proposed algorithm. Let ( x, ȳ, λ) be a solution of (4.3) and let s = 0 be the corresponding optimal objective function value of problem (4.11). Then we have the following result. Theorem Let (x, y) be a point satisfying G(x) 0, y Ψ(x). Let the feasible set of the lower level problem satisfy MFCQ, SSOC, and CRCQ. Moreover, let the optimal objective function value of problem (4.11) with ɛ = 0 be zero and λ be obtained by solving (4.8). Then (x, y) is a Clarke stationary point for problem (2.3). Proof. From y Ψ(x) we have Λ(x, y). That means the system λ T g(x, y) = 0 y L(x, y, λ) = 0 λ 0 has a solution. Moreover, if ɛ = 0 then from (4.11d) we have (4.15) 2 yyl(x, y, λ)r + γ T y g(x, y) = 2 xyl(x, y, λ)d. Also λ i > 0 implies g i (x, y) = 0. In this case (4.11f) gives us and hence we get λ i x g i (x, y)d + λ i y g i (x, y)r 0 x g i (x, y)d + y g i (x, y)r 0 for ɛ = 0. On the other hand from (4.11h) and since the optimal objective function value is zero we have that g i (x, y) = 0 implies x g i (x, y)d + y g i (x, y)r 0. Therefore (4.16) x g i (x, y)d = y g i (x, y)r i : λ i > 0. From (4.15) and (4.16) we derive λ arg max{ x L(x, y, λ)d : λ Λ(x, y)} and r = y (x; d). Again from g i (x, y) < 0 we have λ i = 0. This means, from (4.11i) we get γ i 0. But in (4.11f) ɛ = 0 results in γ i g i (x, y) 0 giving γ i 0.

16 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 13 Therefore γ i = 0. Since d, G i (x) + x G i (x)d < 0 there is λ Λ(x, y) such that r = y (x; d) is computed by solving (4.15), (4.16) such that the system { x F (x, y)d + y F (x, y)y (x; d) < 0 G i (x) + x G i (x)d < 0 has no solution. Or in other words d, G i (x) + x G i (x)d < 0, i : 1,..., l, we have x F (x, y)d + y F (x, y)y (x; d) 0. Hence (x, y) is a Clarke stationary point (2.3) in the sense of Definition 4.2. Remark In the above theorem one can not guarantee that ( x, ȳ) is a Bouligand stationary point because the computed direction depends on the fixed multiplier. If the multiplier changes we can not guarantee the nonexistence of descent direction. Theorem Let ɛ k 0, {x k, y k, λ k, γ k, s k } be a solution of problem (4.5) for ɛ = ɛ k. Let {x k, y k, λ k, γ k, s k } converge to (x, y, λ, γ, 0). If (x, y ) satisfies SSOC, MFCQ, CRCQ and there exists d {d : x G i (x )d < 0, i : G i (x ) = 0} then (x, y ) is a Bouligand stationary point of the optimistic bilevel programming problem (2.3). Proof. Let (x, y ) be not a Bouligand stationary point. Then there must exist with d {d : x G i (x )d < 0, i : G i (x ) = 0} x F (x, y )d + y F (x, y )y (x ; d ) < 0. Since the lower level problem satisfies MFCQ, SSOC, and CRCQ then there is λ Λ(x, y ) with (r, γ ), r = y (x ; d ) is the solution of the following system of linear equations and inequalities: 2 yyl(x, y, λ )r + γ T y g(x, y ) + 2 xyl(x, y, λ )d = 0 x g i (x, y )d + y g i (x, y )r = 0, i : λ i > 0 x g i (x, y )d + y g i (x, y )r 0, i : g i (x, y ) = λ = 0 γ i 0, γ i ( x g i (x, y )d + y g i (x, y )r) = 0, i : g i (x, y ) = λ i = 0 Since the (x, y ) satisfies MFCQ and from the definition of Λ(.,.), y L(x, y, λ ) = 0 λ Λ(x, y ) λ 0 λ T g(x, y ) = 0 which implies that the optimal objective function value of (4.11) for ɛ = 0 is s < 0. Since (x k, y k ) (x, y ) for ɛ 0, ɛ > 0 then there exist k: ɛ y L(x k, y k, λ ) ɛ λ 0 k k. λ T g(x k, y k ) ɛ

17 14 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE Then the following is satisfied x g i (x k, y k )d + y g i (x k, y k )r = 0, i I k g i (x k, y k ) + x g i (x k, y k )d + y g i (x k, y k )r s, i / I k ɛ 2 yyl(x k, y k, λ ) + 2 xyl(x k, y k, λ )d + γ T y g i (x k, y k ) ɛ ɛ λ i g i (x k, y k ) + γi g i(x k, y k ) + λ i ( yg i (x k, y k )d + y g i (x k, y k )r) g i (x k, y k ) + γi ( xg i (x k, y k )d + y g i (x k, y k )r) 0 γi 0, i : λ i = 0, g i(x k, y k ) ɛ which implies that the point (d, r, λ, γ ) is feasible in problem (4.11). But this means that optimal objective function value of (4.11) at the limit is s s < 0, k k which implies the sequence {s k } does not converge to zero. This is a contradiction to the fact that s k 0. Hence the limit (x, y ) is B-stationary for the bilevel programming problem(2.3). 5. Numerical Experience In this section we consider the relaxed KKT reformulation NLP (ɛ) to numerically verify our claim in the previous sections of this paper. In this section, we present our experience with numerical experiments of the proposed algorithm in order to substantiate the theory we have established in previous sections. We present two instances of numerical experiments. The first concerns verifying literature solution by directly solving problem (4.5) for a decreasing sequence of the relaxation parameter ɛ. The second concerns the application of the feasible direction method. The latter is purely the implementation of our algorithm. The numerical test is conducted on some exemplary test problems from Leyffer s MacMPEC collection [13]. MacMPEC collection is a collection of mathematical programs with equilibrium constraints coded in AMPL. We have recoded them in AMPL since we consider the relaxed problem (4.5). The experiment was conducted on Medion MD 95800, Intel R Pentium R M Processor 730, 1.6 GHz with 512 MB RAM under Microsoft Windows operating system. We used the optimization solvers SNOPT, DONLP2, LOQO, and MINOS that can be downloaded along with AMPL student edition Relaxation method. We consider problem (4.5) and the solution approach in Scholtes[21] solved by the same solver for a decreasing sequence of relaxation parameters ɛ. On most instances of test problems our relaxation method and Scholtes regularization produce a sequence of iterates that converges to a local solution of the corresponding optimistic bilevel programming reported in the literature with the exception of problems ex9.1.3, ex9.1.7 and ex Our method produced a sequence of iterates that converge to the local solution of the optimistic bilevel programming ex9.1.7 reported in the original literature. However, with the same initial point, the regularization approach in Scholtes[21] fails to converge to the literature solution of the optimistic bilevel programming problem. We obtained a better solution for problems ex9.1.3, ex9.1.8 and bard3. Table 1 gives information regarding the comparison of best function values. The first column shows serial number of considered problems, the second column indicates the name of the problem. The third, the forth and the fifth columns show respectively best function values from literature, using our approach and using Scholtes regularization scheme [21]. The result seems to confirm the relevance of relaxing the stationarity constraint of problem (4.5).

18 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 15 Best function value obtained from ours Scholtes No Name literature approach approach source comment 1 bard [1] 2 bard [1] 3 ex [8] 4 ex [8] 5 ex [8] 6 ex [8] 7 ex [8] 8 ex [8] 9 ex [8] better solution 10 ex [8] 11 ex [8] 12 ex [8] 13 ex [8] 14 ex [8] 15 ex [8] 16 ex [8] 17 ex [8] 18 ex [8] 19 ex [8] 20 ex [8] 21 TuMiPh [23] 22 MD [15] 23 FLZl [9] Table 1. Comparison our approach with Scholtes regularization scheme Feasible direction method. Initially Algorithm 4.1 was implemented in Matlab 7. It is known that Matlab s programming facility made it suitable for such an iterative method. However, Matlab s optimization solvers performed poorly at least for our problems. It was observed that a wrong multiplier is calculated from problems (4.8) and (4.10). Having observed this limitation in Matlab solvers, the algorithm was re-implemented in AMPL using AMPL s script facility. The algorithm solves three problems alternately. The multiplier problem, the direction finding problem and the choice of the step size parameter. The step size t is chosen as the maximum of { 1 2, 1 4, 1 8,...} such that the system F (x k + td k, y k + tr k ) F (x k, y k ) + δts k G i (x k + td k ) 0, i = 1,..., l g j (x k + td k, y k + tr k ) 0, j = 1,..., p is satisfied. For the current incumbent primal point (x k, y k ) the algorithm first computes the corresponding dual point λ either from linear inequalities (4.8) or quadratic optimization problem (4.10) depending on the size of s at hand. Once the appropriate triplets (x k, y k, λ k ) found, the direction finding problem (4.11) is solved by CPLEX, a linear programming solver, which can be downloaded from AMPL web

19 16 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE site, and problem (4.10) is solved by Spellucci s DONLP2, a nonlinear programming solver. In our AMPL implementation, this is done by alternatively solving different problems. For the test problems listed in Table 1, literature solutions are obtained. However, for some problems such as ex9.2.7 the solution slow convergence was observed. This indicates that some researches and more experiments has to be done in order to improve the performance of this algorithm. To better illustrate this we consider the following example. This example is taken from the monograph [8]. Example 5.1. Consider (5.1) where y solves min (x x 0,y 0 5)2 + (2y + 1) 2 s.t. min (y y 0 1)2 1.5xy 3x + y 3 x.5y 4 x + y 7. The convex lower level problem is replaced by the its KKT condition and then the stationarity condition is relaxed. The feasible direction method is applied to the problem (5.2) s.t. min (x 5) 2 + (2y + 1) 2 2(y 1) 1.5x + λ 1.5λ 2 + λ 3 ɛ 2(y 1) + 1.5x λ 1 +.5λ 2 λ 3 ɛ 3x + y x + 0.5y 4 0 x + y 7 0 λ 1 ( 3x + y + 3) ɛ λ 2 (x + 0.5y 4) ɛ λ 3 (x + y 7) ɛ

20 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 17 The direction finding problem is given by s min d,r,s,γ 2(x 5)d + 4(2y + 1)r s x d s 2(y 1) 1.5x + λ 1.5λ 2 + λ 3 λ 4 1.5d + 2r + γ 1 0.5γ 2 + γ 3 γ 4 ɛ 2(y 1) + 1.5x λ 1 +.5λ 2 λ 3 + λ d 2r γ γ 2 γ 3 + γ 4 ɛ { = 0 if 3x + y + 3 = 0 3x + y + 3 3d + r s if 3x + y { = 0 if x 0.5y 4 = 0 x 0.5y 4 + d 0.5r s if x 0.5y 4 0 { = 0 if x + y 7 = 0 x + y 7 + d + r s if x + y 7 0 λ 1 ( 3x + y + 3) γ 1 ( 3x + y + 3) λ 1 ( 3d + r) ɛ λ 2 (x 0.5y 4) γ 2 (x 0.5y 4) λ 2 (d.5r) ɛ λ 3 (x + y 7) γ 3 (x + y 7) λ 3 (d + r) ɛ λ 4 y + γ 4 y + λ 4 r ɛ λ 1 γ 1 s 0 λ 2 γ 2 s 0 λ 3 γ 3 s 0 λ 4 γ 4 s 0 For this particular problem the initial point is taken to be (x, y ) = (1.1, 0.3), (λ 1, λ 2, λ 3, λ 4 ) = (0, 0, 0, 0), σ = 10 1, δ = 0.3, ɛ = The solution reported in the literature is (1, 0)(see [8, Problem 9.3.8]). The first iteration is the computation of the multiplier. Then the direction finding problem is solved using CPLEX. The step size parameter is chosen in the sense of Armijo. These three alternating phase of computation continues until the stopping criterion (ɛ = 10 6 for this problem) satisfied. For this particular problem, the algorithm converges to the above mentioned literature solution after several iterations. 6. Conclusion In this paper, a descent like algorithm based on the framework of a feasible direction method is proposed to compute Bouligand stationary points of optimistic bilevel programming problems with convex lower level problem. It has been shown and numerically evidenced that in addition to relaxing complementarity conditions, relaxing the stationarity conditions of the auxiliary MPEC problem (4.3) has a significant advantage on computing the appropriate multiplier, and hence finding Bouligand stationary points of the optimistic bilevel programming problems. The algorithm is implemented for some bilevel programming problems. Its performance, though in its infant stage, is found promising. References [1] J. F. Bard. Convex two-level optimization. Mathematical Programming, 40:15 27, 1988.

21 18 AYALEW GETACHEW MERSHA AND STEPHAN DEMPE [2] J. F. Bard. Some properties of the bilevel programming problem. Journal of Optimization Theory and Applications, 68: , [3] J. F. Bard. Practical Bilevel Optimization: Algorithms and Applications. Kluwer Academic Publishers, Dordrecht, [4] S. Dempe. Foundations of Bilevel programming. Kluwer Academic Publishers, Dordrecht et al., [5] S. Dempe and H. Schmidt. On an algorithm solving two-level programming problems with nonunique lower level solutions. Computational Optimization and Applications, 6: , [6] J. Dutta and S. Dempe. Bilevel programming with convex lower level problem. In S. Dempe and V. Kalashnikov, editors, Optimization with Multivalued Mappings: Theory, Applications and Algorithms, volume 2 of Springer Optimization and Its Applications, pages Springer US, [7] R. Fletcher and S. Leyffer. Numerical experience with solving MPECs as NLPs. Technical Report NA/210, Department of Mathematics, University of Dundee, [8] C. A. Floudas, P. M. Pardalos, C. S. Adjiman, W. R. Esposito, Z. H. Gümüs, S. T. Harding, J. L. Klepeis, C. A. Meyer, and C. A. Schweiger. Handbook of test problems in local and global optimization. Number 33 in Nonconvex Optimization and Its Applications. Kluwer Academic Publishers, Dordrecht, [9] C. A. Floudas and S Zlobec. Multilevel Optimization: Algorithms and Applications, volume 20, chapter Optimality and duality in parametric convex lexographic programming, pages Kluwer Academic Publishers, Dordrecht, [10] J. Gauvin. A necessary and sufficient regularity condition to have bounded multiplier in non convex programming. Mathematical Programming, 12: , [11] W.W. Hager. Lipschitz continuity for constrained processes. SIAM Journal on Control and Optimization, 17: , [12] M. Kojima. Strongly stable stationary solutions in nonlinear programs. In S.M. Robinson, editor, Analysis and Computation of Fixed Points, pages Academic Press, New York, [13] S. Leyffer. MacMPEC ampl collection of Mathematical Programs with Equilibrium Constraints. Argonne National Laboratory, MacMPEC/. [14] Z.-Q. Luo, J.-S. Pang, and D. Ralph. Mathematical Programs with Equilibrium Constraints. Cambridge University Press, Cambridge, [15] A.Getachew Mersha and S. Dempe. Linear bilevel programming with upper level constraints depending on the lower level solution. Applied Mathematics and Computation, 180: , [16] R. Mifflin. Semismooth and semiconvex functions in constrained optimization. SIAM Journal on Control and Optimization, 15: , [17] J. Outrata, M. Kočvara, and J. Zowe. Nonsmooth Approach to Optimization Problems with Equilibrium Constraints. Kluwer Academic Publishers, Dordrecht, [18] D. Ralph and S. Dempe. Directional derivatives of the solution of a parametric nonlinear program. Mathematical Programming, 70: , [19] H. Scheel and S. Scholtes. Mathematical programs with equilibrium constraints: stationarity, optimality, and sensitivity. Mathematics of Operations Research, 25:1 22, [20] S. Scholtes. Introduction to piecewise differentaible equations. Preprint , Universität Karlsruhe, Karlsruhe, [21] S. Scholtes. Convergence properties of a regularization scheme for mathematical programs with complementarity constraints. SIAM Journal on Optimization, 11: , [22] D. M. Topkis and A.F. Veinott. On the convergence of of some feasible direction algorithms for nonlinear programming. SIAM Journal on Control and Optimization, 5: , [23] H. Tuy, A. Migdalas, and N.T. Hoai-Phuong. A novel approach to bilevel nonlinear programming. Journal of Global Optimization, 38: , 2007.

22 FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM 19 Department of Mathematics and Computer Science, Technical University Bergakademie Freiberg, Germany address: Department of Mathematics and Computer Science, Technical University Bergakademie Freiberg, Germany address:

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen (für Informatiker) M. Grepl J. Berger & J.T. Frings Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2010/11 Problem Statement Unconstrained Optimality Conditions Constrained

More information

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 J. Zhang Institute of Applied Mathematics, Chongqing University of Posts and Telecommunications, Chongqing

More information

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION 1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of

More information

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results

More information

Date: April 12, 2001. Contents

Date: April 12, 2001. Contents 2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........

More information

Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725

Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725 Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T

More information

2.3 Convex Constrained Optimization Problems

2.3 Convex Constrained Optimization Problems 42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions

More information

Nonlinear Optimization: Algorithms 3: Interior-point methods

Nonlinear Optimization: Algorithms 3: Interior-point methods Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris Jean-Philippe.Vert@mines.org Nonlinear optimization c 2006 Jean-Philippe Vert,

More information

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d). 1. Line Search Methods Let f : R n R be given and suppose that x c is our current best estimate of a solution to P min x R nf(x). A standard method for improving the estimate x c is to choose a direction

More information

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the

More information

GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1

GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1 (R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter MWetter@lbl.gov February 20, 2009 Notice: This work was supported by the U.S.

More information

Duality of linear conic problems

Duality of linear conic problems Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least

More information

GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS. 1. Introduction

GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS. 1. Introduction GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS N.V. THOAI, Y. YAMAMOTO, AND A. YOSHISE Abstract. We propose a method for finding a global optimal solution

More information

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}. Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian

More information

10. Proximal point method

10. Proximal point method L. Vandenberghe EE236C Spring 2013-14) 10. Proximal point method proximal point method augmented Lagrangian method Moreau-Yosida smoothing 10-1 Proximal point method a conceptual algorithm for minimizing

More information

Some stability results of parameter identification in a jump diffusion model

Some stability results of parameter identification in a jump diffusion model Some stability results of parameter identification in a jump diffusion model D. Düvelmeyer Technische Universität Chemnitz, Fakultät für Mathematik, 09107 Chemnitz, Germany Abstract In this paper we discuss

More information

Convex Programming Tools for Disjunctive Programs

Convex Programming Tools for Disjunctive Programs Convex Programming Tools for Disjunctive Programs João Soares, Departamento de Matemática, Universidade de Coimbra, Portugal Abstract A Disjunctive Program (DP) is a mathematical program whose feasible

More information

Big Data - Lecture 1 Optimization reminders

Big Data - Lecture 1 Optimization reminders Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics

More information

Nonlinear Programming Methods.S2 Quadratic Programming

Nonlinear Programming Methods.S2 Quadratic Programming Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective

More information

4.6 Linear Programming duality

4.6 Linear Programming duality 4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal

More information

24. The Branch and Bound Method

24. The Branch and Bound Method 24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no

More information

Duality in Linear Programming

Duality in Linear Programming Duality in Linear Programming 4 In the preceding chapter on sensitivity analysis, we saw that the shadow-price interpretation of the optimal simplex multipliers is a very useful concept. First, these shadow

More information

Separation Properties for Locally Convex Cones

Separation Properties for Locally Convex Cones Journal of Convex Analysis Volume 9 (2002), No. 1, 301 307 Separation Properties for Locally Convex Cones Walter Roth Department of Mathematics, Universiti Brunei Darussalam, Gadong BE1410, Brunei Darussalam

More information

Several Views of Support Vector Machines

Several Views of Support Vector Machines Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min

More information

Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach

Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach Outline Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach The University of New South Wales SPOM 2013 Joint work with V. Jeyakumar, B.S. Mordukhovich and

More information

Some representability and duality results for convex mixed-integer programs.

Some representability and duality results for convex mixed-integer programs. Some representability and duality results for convex mixed-integer programs. Santanu S. Dey Joint work with Diego Morán and Juan Pablo Vielma December 17, 2012. Introduction About Motivation Mixed integer

More information

ALMOST COMMON PRIORS 1. INTRODUCTION

ALMOST COMMON PRIORS 1. INTRODUCTION ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type

More information

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem

More information

ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING. 0. Introduction

ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING. 0. Introduction ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING Abstract. In [1] there was proved a theorem concerning the continuity of the composition mapping, and there was announced a theorem on sequential continuity

More information

Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization

Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 2010, Article ID 510838, 15 pages doi:10.1155/2010/510838 Research Article Stability Analysis for Higher-Order Adjacent Derivative

More information

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics Undergraduate Notes in Mathematics Arkansas Tech University Department of Mathematics An Introductory Single Variable Real Analysis: A Learning Approach through Problem Solving Marcel B. Finan c All Rights

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

FACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z

FACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z FACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z DANIEL BIRMAJER, JUAN B GIL, AND MICHAEL WEINER Abstract We consider polynomials with integer coefficients and discuss their factorization

More information

Numerical methods for American options

Numerical methods for American options Lecture 9 Numerical methods for American options Lecture Notes by Andrzej Palczewski Computational Finance p. 1 American options The holder of an American option has the right to exercise it at any moment

More information

Practical Guide to the Simplex Method of Linear Programming

Practical Guide to the Simplex Method of Linear Programming Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear

More information

Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms

Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms G. Wang, Z. Wan and X. Wang Abstract The paper studies and designs an genetic algorithm (GA) of the bilevel linear programming

More information

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical

More information

Numerical Verification of Optimality Conditions in Optimal Control Problems

Numerical Verification of Optimality Conditions in Optimal Control Problems Numerical Verification of Optimality Conditions in Optimal Control Problems Dissertation zur Erlangung des naturwissenschaftlichen Doktorgrades der Julius-Maximilians-Universität Würzburg vorgelegt von

More information

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand Notes V General Equilibrium: Positive Theory In this lecture we go on considering a general equilibrium model of a private ownership economy. In contrast to the Notes IV, we focus on positive issues such

More information

Lecture 7: Finding Lyapunov Functions 1

Lecture 7: Finding Lyapunov Functions 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.243j (Fall 2003): DYNAMICS OF NONLINEAR SYSTEMS by A. Megretski Lecture 7: Finding Lyapunov Functions 1

More information

Mathematical finance and linear programming (optimization)

Mathematical finance and linear programming (optimization) Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may

More information

On the D-Stability of Linear and Nonlinear Positive Switched Systems

On the D-Stability of Linear and Nonlinear Positive Switched Systems On the D-Stability of Linear and Nonlinear Positive Switched Systems V. S. Bokharaie, O. Mason and F. Wirth Abstract We present a number of results on D-stability of positive switched systems. Different

More information

International Doctoral School Algorithmic Decision Theory: MCDA and MOO

International Doctoral School Algorithmic Decision Theory: MCDA and MOO International Doctoral School Algorithmic Decision Theory: MCDA and MOO Lecture 2: Multiobjective Linear Programming Department of Engineering Science, The University of Auckland, New Zealand Laboratoire

More information

A Branch and Bound Algorithm for Solving the Binary Bi-level Linear Programming Problem

A Branch and Bound Algorithm for Solving the Binary Bi-level Linear Programming Problem A Branch and Bound Algorithm for Solving the Binary Bi-level Linear Programming Problem John Karlof and Peter Hocking Mathematics and Statistics Department University of North Carolina Wilmington Wilmington,

More information

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)! Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following

More information

(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties

(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties Lecture 1 Convex Sets (Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties 1.1.1 A convex set In the school geometry

More information

On Minimal Valid Inequalities for Mixed Integer Conic Programs

On Minimal Valid Inequalities for Mixed Integer Conic Programs On Minimal Valid Inequalities for Mixed Integer Conic Programs Fatma Kılınç Karzan June 27, 2013 Abstract We study mixed integer conic sets involving a general regular (closed, convex, full dimensional,

More information

An interval linear programming contractor

An interval linear programming contractor An interval linear programming contractor Introduction Milan Hladík Abstract. We consider linear programming with interval data. One of the most challenging problems in this topic is to determine or tight

More information

MATH 425, PRACTICE FINAL EXAM SOLUTIONS.

MATH 425, PRACTICE FINAL EXAM SOLUTIONS. MATH 45, PRACTICE FINAL EXAM SOLUTIONS. Exercise. a Is the operator L defined on smooth functions of x, y by L u := u xx + cosu linear? b Does the answer change if we replace the operator L by the operator

More information

BANACH AND HILBERT SPACE REVIEW

BANACH AND HILBERT SPACE REVIEW BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but

More information

TOPIC 4: DERIVATIVES

TOPIC 4: DERIVATIVES TOPIC 4: DERIVATIVES 1. The derivative of a function. Differentiation rules 1.1. The slope of a curve. The slope of a curve at a point P is a measure of the steepness of the curve. If Q is a point on the

More information

Cyber-Security Analysis of State Estimators in Power Systems

Cyber-Security Analysis of State Estimators in Power Systems Cyber-Security Analysis of State Estimators in Electric Power Systems André Teixeira 1, Saurabh Amin 2, Henrik Sandberg 1, Karl H. Johansson 1, and Shankar Sastry 2 ACCESS Linnaeus Centre, KTH-Royal Institute

More information

Notes on metric spaces

Notes on metric spaces Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.

More information

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS KEITH CONRAD 1. Introduction The Fundamental Theorem of Algebra says every nonconstant polynomial with complex coefficients can be factored into linear

More information

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

More information

BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION

BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION Ş. İlker Birbil Sabancı University Ali Taylan Cemgil 1, Hazal Koptagel 1, Figen Öztoprak 2, Umut Şimşekli

More information

CONSTRAINED NONLINEAR PROGRAMMING

CONSTRAINED NONLINEAR PROGRAMMING 149 CONSTRAINED NONLINEAR PROGRAMMING We now turn to methods for general constrained nonlinear programming. These may be broadly classified into two categories: 1. TRANSFORMATION METHODS: In this approach

More information

Proximal mapping via network optimization

Proximal mapping via network optimization L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:

More information

Linear Programming in Matrix Form

Linear Programming in Matrix Form Linear Programming in Matrix Form Appendix B We first introduce matrix concepts in linear programming by developing a variation of the simplex method called the revised simplex method. This algorithm,

More information

Nonlinear Algebraic Equations Example

Nonlinear Algebraic Equations Example Nonlinear Algebraic Equations Example Continuous Stirred Tank Reactor (CSTR). Look for steady state concentrations & temperature. s r (in) p,i (in) i In: N spieces with concentrations c, heat capacities

More information

Solving polynomial least squares problems via semidefinite programming relaxations

Solving polynomial least squares problems via semidefinite programming relaxations Solving polynomial least squares problems via semidefinite programming relaxations Sunyoung Kim and Masakazu Kojima August 2007, revised in November, 2007 Abstract. A polynomial optimization problem whose

More information

Adaptive Online Gradient Descent

Adaptive Online Gradient Descent Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650

More information

Max-Min Representation of Piecewise Linear Functions

Max-Min Representation of Piecewise Linear Functions Beiträge zur Algebra und Geometrie Contributions to Algebra and Geometry Volume 43 (2002), No. 1, 297-302. Max-Min Representation of Piecewise Linear Functions Sergei Ovchinnikov Mathematics Department,

More information

Linear Programming Notes V Problem Transformations

Linear Programming Notes V Problem Transformations Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material

More information

CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e.

CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e. CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e. This chapter contains the beginnings of the most important, and probably the most subtle, notion in mathematical analysis, i.e.,

More information

Constrained optimization.

Constrained optimization. ams/econ 11b supplementary notes ucsc Constrained optimization. c 2010, Yonatan Katznelson 1. Constraints In many of the optimization problems that arise in economics, there are restrictions on the values

More information

Robust Geometric Programming is co-np hard

Robust Geometric Programming is co-np hard Robust Geometric Programming is co-np hard André Chassein and Marc Goerigk Fachbereich Mathematik, Technische Universität Kaiserslautern, Germany Abstract Geometric Programming is a useful tool with a

More information

Practice with Proofs

Practice with Proofs Practice with Proofs October 6, 2014 Recall the following Definition 0.1. A function f is increasing if for every x, y in the domain of f, x < y = f(x) < f(y) 1. Prove that h(x) = x 3 is increasing, using

More information

The Envelope Theorem 1

The Envelope Theorem 1 John Nachbar Washington University April 2, 2015 1 Introduction. The Envelope Theorem 1 The Envelope theorem is a corollary of the Karush-Kuhn-Tucker theorem (KKT) that characterizes changes in the value

More information

CHAPTER 9. Integer Programming

CHAPTER 9. Integer Programming CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral

More information

Support Vector Machines Explained

Support Vector Machines Explained March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),

More information

Randomization Approaches for Network Revenue Management with Customer Choice Behavior

Randomization Approaches for Network Revenue Management with Customer Choice Behavior Randomization Approaches for Network Revenue Management with Customer Choice Behavior Sumit Kunnumkal Indian School of Business, Gachibowli, Hyderabad, 500032, India sumit kunnumkal@isb.edu March 9, 2011

More information

Discrete Optimization

Discrete Optimization Discrete Optimization [Chen, Batson, Dang: Applied integer Programming] Chapter 3 and 4.1-4.3 by Johan Högdahl and Victoria Svedberg Seminar 2, 2015-03-31 Todays presentation Chapter 3 Transforms using

More information

4 Lyapunov Stability Theory

4 Lyapunov Stability Theory 4 Lyapunov Stability Theory In this section we review the tools of Lyapunov stability theory. These tools will be used in the next section to analyze the stability properties of a robot controller. We

More information

Convex analysis and profit/cost/support functions

Convex analysis and profit/cost/support functions CALIFORNIA INSTITUTE OF TECHNOLOGY Division of the Humanities and Social Sciences Convex analysis and profit/cost/support functions KC Border October 2004 Revised January 2009 Let A be a subset of R m

More information

Optimal shift scheduling with a global service level constraint

Optimal shift scheduling with a global service level constraint Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The

More information

Fuzzy Differential Systems and the New Concept of Stability

Fuzzy Differential Systems and the New Concept of Stability Nonlinear Dynamics and Systems Theory, 1(2) (2001) 111 119 Fuzzy Differential Systems and the New Concept of Stability V. Lakshmikantham 1 and S. Leela 2 1 Department of Mathematical Sciences, Florida

More information

A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems

A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems myjournal manuscript No. (will be inserted by the editor) A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems Eduardo Álvarez-Miranda Ivana Ljubić Paolo Toth Received:

More information

Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

More information

x a x 2 (1 + x 2 ) n.

x a x 2 (1 + x 2 ) n. Limits and continuity Suppose that we have a function f : R R. Let a R. We say that f(x) tends to the limit l as x tends to a; lim f(x) = l ; x a if, given any real number ɛ > 0, there exists a real number

More information

Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points

Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points Introduction to Algebraic Geometry Bézout s Theorem and Inflection Points 1. The resultant. Let K be a field. Then the polynomial ring K[x] is a unique factorisation domain (UFD). Another example of a

More information

What is Linear Programming?

What is Linear Programming? Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to

More information

Metric Spaces. Chapter 7. 7.1. Metrics

Metric Spaces. Chapter 7. 7.1. Metrics Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some

More information

CONTINUED FRACTIONS AND PELL S EQUATION. Contents 1. Continued Fractions 1 2. Solution to Pell s Equation 9 References 12

CONTINUED FRACTIONS AND PELL S EQUATION. Contents 1. Continued Fractions 1 2. Solution to Pell s Equation 9 References 12 CONTINUED FRACTIONS AND PELL S EQUATION SEUNG HYUN YANG Abstract. In this REU paper, I will use some important characteristics of continued fractions to give the complete set of solutions to Pell s equation.

More information

Optimal energy trade-off schedules

Optimal energy trade-off schedules Optimal energy trade-off schedules Neal Barcelo, Daniel G. Cole, Dimitrios Letsios, Michael Nugent, Kirk R. Pruhs To cite this version: Neal Barcelo, Daniel G. Cole, Dimitrios Letsios, Michael Nugent,

More information

Interior Point Methods and Linear Programming

Interior Point Methods and Linear Programming Interior Point Methods and Linear Programming Robert Robere University of Toronto December 13, 2012 Abstract The linear programming problem is usually solved through the use of one of two algorithms: either

More information

A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS

A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors

More information

15 Kuhn -Tucker conditions

15 Kuhn -Tucker conditions 5 Kuhn -Tucker conditions Consider a version of the consumer problem in which quasilinear utility x 2 + 4 x 2 is maximised subject to x +x 2 =. Mechanically applying the Lagrange multiplier/common slopes

More information

A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets

A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets Abstract Adilson Elias Xavier Vinicius Layter Xavier Dept. of Systems Engineering and Computer Science Graduate School of Engineering

More information

Scheduling Parallel Jobs with Monotone Speedup 1

Scheduling Parallel Jobs with Monotone Speedup 1 Scheduling Parallel Jobs with Monotone Speedup 1 Alexander Grigoriev, Marc Uetz Maastricht University, Quantitative Economics, P.O.Box 616, 6200 MD Maastricht, The Netherlands, {a.grigoriev@ke.unimaas.nl,

More information

GI01/M055 Supervised Learning Proximal Methods

GI01/M055 Supervised Learning Proximal Methods GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators

More information

constraint. Let us penalize ourselves for making the constraint too big. We end up with a

constraint. Let us penalize ourselves for making the constraint too big. We end up with a Chapter 4 Constrained Optimization 4.1 Equality Constraints (Lagrangians) Suppose we have a problem: Maximize 5, (x 1, 2) 2, 2(x 2, 1) 2 subject to x 1 +4x 2 =3 If we ignore the constraint, we get the

More information

DIFFERENTIABILITY OF COMPLEX FUNCTIONS. Contents

DIFFERENTIABILITY OF COMPLEX FUNCTIONS. Contents DIFFERENTIABILITY OF COMPLEX FUNCTIONS Contents 1. Limit definition of a derivative 1 2. Holomorphic functions, the Cauchy-Riemann equations 3 3. Differentiability of real functions 5 4. A sufficient condition

More information

Online Learning of Optimal Strategies in Unknown Environments

Online Learning of Optimal Strategies in Unknown Environments 1 Online Learning of Optimal Strategies in Unknown Environments Santiago Paternain and Alejandro Ribeiro Abstract Define an environment as a set of convex constraint functions that vary arbitrarily over

More information

3. Linear Programming and Polyhedral Combinatorics

3. Linear Programming and Polyhedral Combinatorics Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the

More information

Multiple Solutions of Mixed Variable Optimization by Multistart Hooke and Jeeves Filter Method

Multiple Solutions of Mixed Variable Optimization by Multistart Hooke and Jeeves Filter Method Applied Mathematical Sciences, Vol. 8, 2014, no. 44, 2163-2179 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.4143 Multiple Solutions of Mixed Variable Optimization by Multistart Hooke

More information

Shape Optimization Problems over Classes of Convex Domains

Shape Optimization Problems over Classes of Convex Domains Shape Optimization Problems over Classes of Convex Domains Giuseppe BUTTAZZO Dipartimento di Matematica Via Buonarroti, 2 56127 PISA ITALY e-mail: buttazzo@sab.sns.it Paolo GUASONI Scuola Normale Superiore

More information

Equilibrium computation: Part 1

Equilibrium computation: Part 1 Equilibrium computation: Part 1 Nicola Gatti 1 Troels Bjerre Sorensen 2 1 Politecnico di Milano, Italy 2 Duke University, USA Nicola Gatti and Troels Bjerre Sørensen ( Politecnico di Milano, Italy, Equilibrium

More information

A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA: A FIRST REPORT

A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA: A FIRST REPORT New Mathematics and Natural Computation Vol. 1, No. 2 (2005) 295 303 c World Scientific Publishing Company A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA:

More information

Equilibria and Dynamics of. Supply Chain Network Competition with Information Asymmetry. and Minimum Quality Standards

Equilibria and Dynamics of. Supply Chain Network Competition with Information Asymmetry. and Minimum Quality Standards Equilibria and Dynamics of Supply Chain Network Competition with Information Asymmetry in Quality and Minimum Quality Standards Anna Nagurney John F. Smith Memorial Professor and Dong (Michelle) Li Doctoral

More information