An interval linear programming contractor

Save this PDF as:

Size: px
Start display at page:

Transcription

1 An interval linear programming contractor Introduction Milan Hladík Abstract. We consider linear programming with interval data. One of the most challenging problems in this topic is to determine or tight approximate the set of all optimal solutions subject to all perturbations within the given intervals. We propose an iterative method that finds an enclosure of the set of optimal solutions. he method is based on a linear approximation and sequential refinement. It runs in polynomial time, so, naturally, convergence to the ideal set cannot be ensured. We apply the method in a simple portfolio selection problem with uncertain data. Keywords: interval linear programming, interval analysis JEL classification: C44 AMS classification: 90C3 Since linear programming is widely used in modelling real-life problems, it must take into account inaccuracies and data measurement errors, which are common in most of the problems. here are plenty of various approaches to handle uncertainty in linear systems, see e.g. [4]. In this paper, we deal with an interval linear programming, in which we assume that there are a priori known some intervals in which inexact quantities may perturb. Linear programming with interval data has been studied for forty years; see a survey paper [0]. he problems discussed are the optimal value range [4], [7], [2], basis stability, and duality [5], among others. Interval linear programming was applied in portfolio selection problems[6], environmental management [5] interval matrix games [6], and can also serve in fuzzy linear regression as an alternative to traditional approaches [, 3]. In this paper, we focus on the optimal solution set. he problem of calculating the set of all possible solutions over all data perturbations is considered to be very difficult. It becomes tractable in the special case of the so called basis stability [8], [4], [7], meaning that there is a basis that is optimal under any admissible perturbation. However, basis stability is not so easy to verify; indeed, it is an NP-hard problem [0]. Moreover, since many practical problems suffer from degeneracy, one cannot expect that basis stability holds true in general. hus, the research was driven to calculate an enclosure (interval superset) of the optimal solution set. Such enclosures can be computed e.g. by using interval arithmetic [2], [3], instead of the real one, but the results are usually very overestimated. Our aim is to propose an efficient algorithm for computing an enclosure of the optimal solution set. We present an iterative algorithm that starts with an initial enclosure and sequentially makes it tighter. Naturally, it doesn t converge to the optimal enclosure in general, but (in view of the performed examples) it gives bounds that are sufficiently accurate for many purposes. Let us introduce some notations. An interval matrix is a family of matrices A := {A R m n A A A}, wherea,a R m n aregiven. he midpoint andthe radiusofanintervalmatrixaisdefined respectively as A c := 2 (A+A), A := 2 (A A), Charles University, Faculty of Mathematics and Physics, Department of Applied Mathematics, Malostranské nám. 25, 8 00, Prague, Czech Republic, and University of Economics, Faculty of Informatics and Statistics, nám. W. Churchilla 4, 3067, Prague, Czech Republic,

2 and similarly for interval vectors. he set of all interval m n matrices is denoted by IR m n. A solution to the interval system Ax = b, x 0 means any solution to any scenario Ax = b, x 0 with A A, b b. Similarlyfor interval systemsofinequalities. he diagonalmatrixwith entries z,...,z n is denoted by diag(z). 2 Interval linear programming Consider a linear programming problem and its dual minc x subject to Ax = b, x 0, () maxb y subject to A y c. Suppose that (possibly all) input values are inexact, and we have lower and upper limits for the ranges of variations. hus, we are given A IR m n, b IR m, c IR n, and we focus on the family of linear programs () with A A, b b, c c. Denote by S(A,b,c) the set of optimal solutions to (). By the set of optimal solutions to an interval linear program we understand the set S := S(A,b,c). A A, b b, c c he set S needn t be a polyhedron, and it is hard to determine it in general. We want to give a cheap calculation for an enclosure to S. By using duality theory, we have that x S if and only if there is some y R m, A A, b b, and c c such that Relaxing the dependency we obtain a superset described by Ax = b, x 0, A y c, c x = b y. (2) Ax = b, x 0, A y c, c x = b y. (3) Notice that the solution sets are not equal, since in the latter the intervals vary independently within A and A, while in the former they are related. By [0], the solution set to (3) is described as Ax b, Ax b, x 0, A c y A y c, c c x b c y c x+b y. (4) It represents a non-convex polyhedron, which becomes convex when we restrict the signs of y i, i =,...,m. A simple method introduced in [0] was based on decomposing (4) into 2 m sub-problems according to the signs of y i, i =,...,m. Each sub-problem has a linear description and is solved by ordinary linear programming. However, the method is very time consuming as m grows. In the following, we employ a kind of linearization of (4), which is based on a result by Beaumont []. It needs an initial enclosure, which is used for refinement. heorem (Beaumont, 998). For every y y R with y < y one has where α = y y y y Moreover, if y 0 or y 0 then (5) holds as equation. y αy +β, (5) and β = y y y y. y y Let x IR n and y IR m be an enclosure to (4). By the Beaumont theorem, we will linearize the absolute value y as follows. Define vectors α,β R m componentwise as y i y i α i := y i y if y i < y i i, sgn(y i ) if y i = y i, y i y i y i y i β i := y i y if y i < y i i, 0 if y i = y i

3 hen the linearization of (4) reads Ax b, Ax b, x 0, (6a) ( A c A diag(α)) y c+a β, (6b) c x+ ( b c b diag(α)) y b β, c x+ ( b c b diag(α)) y b β. Now, we compute the interval hull to (6). If it is smallerthan the initial enclosure x,y then we can iterate the process to obtain more tight enclosure. It is a basic idea of our method described in Algorithm. he initial enclosure x 0,y 0 from step is taken as x 0 := ([0,K],...,[0,K]), y 0 := ([ K,K],...,[ K,K]), where K 0 is large enough. he stopping criterion used in step 7 is n j= (xi ) j + m j= (yi ) j n j= (xi ) j + m j= (yi ) j (6c) (6d) Algorithm (Optimal solution set contractor) : Compute an initial interval enclosure x 0,y 0 of (4); 2: i := 0; 3: repeat 4: compute α and β by using y i ; 5: i := i+; 6: compute the interval hull x i,y i of (6); 7: until improvement is nonsignificant; 8: return x i ; Remark. here is a natural problem how to choose a sufficiently large K for the initial enclosure. Even though the algorithm works well for very conservative K (cf. Example ), in some cases, we can validate that K is large enough. Suppose that x 0,y 0 contain at least one optimal solution for some scenario (which is easy to satisfy). Next, suppose that (2) is solvable for each A A, b b, and c c; this issue is discussed e.g. in [0]. ake the initial enclosure x 0,y 0 as x 0 := ([,K],...,[,K]), y 0 := ([ K,K],...,[ K,K]), If the interval vectors x,y from the next iteration are strictly inside the initial ones, that is, if x 0 < x x < x 0 and y 0 < y y < y 0, then x,y comprise all optimal solutions (this is why we put the lower limits of x 0 as instead of 0). his is easily seen from the continuity reasons since, under our assumption, the optimal solution set is connected. Example. Consider an interval linear program min [5,6]x [7,8]x 2 subject to x [0,], x +[5,6]x 2 [25,26], [6,6.5]x +[3,4.5]x 2 [8,82], x, x [0,2]x 2 [,2]. Even though it is not in the standard form (), the associated primal dual pair is the same as for (). We take the initial enclosure x 0 = 000 ([,], [,]), y 0 = 000 ([0,], [0,], [0,], [0,], [0,])

4 x x Figure : Intersection of all feasible sets in dark gray; union in light gray; the resulting enclosure of optimal solutions represented by the dotted rectangle. he iterations of the procedure go as follows x = ([,], [ 568,96]), y = ([0,000], [0,936], [0,358], [0,000], [0,572]), x 2 = ([,], [ 7.2,72]), y 2 = ([0,90], [0,58.5], [0,24.3], [0,76], [0,34.6]), x 3 = ([3.78,], [.9,9.80]), y 3 = ([0,30.6], [0,6.98], [4.7], [0,7.], [0,3.09]), x 4 = ([6.65,], [2.66,7.2]), y 4 = ([0,22.5], [0.08,4.33], [0,3.67], [0,8.8], [0,.47]). he fifth iteration does no improvement, so we return the enclosure x 4. Notice that the same result is obtained by the exponential decomposition procedure discussed above, but with a much more computational effort. An illustration is given in Figure. We see that the upper bounds are very tight, but the lower bounds are quite conservative. Example 2. Let us apply our approach in a simple portfolio selection model. Consider we have J possible investments for time periods and r tj, t =,...,, j =,...,J, stands for return on investment j in time period t. Estimated reward on investment j using historical means is defined as W j := r tj, j =,...,J, t= or in matrix form W = R e, where e is the vector of ones (with convenient dimension). In order to get a linear programming formulation of the problem we measure risk of investment j by sum of absolute values instead of the sum of the squares: r tj W j. t= Let µ be a risk aversion parameter (upper bound for risk) given by a user, and the variables x j, j =,...,J, denotes a fraction of portfolio to invest in j. hen the maximal allowed risk is expressed by the constraint J (r tj W j )x j µ, t= or, by converting to a linear inequality system j= J y j (r tj W j )x j y t, t =,...,, j= y t µ. t=

5 In a compact matrix form, it reads y (I E ) Rx y, e y µ, where E is the matrix of ones and I the identity matrix. In order to obtain a robust formulation of the portfolio selection problem, we state the linear programming problem as max J W j x j = e Rx j= subject to y e x =, (I E ) z y, Rx = z, e y µ, x,y 0. For concreteness, consider a portfolio selection problem with J = 4 investments and = 5 time periods. he risk aversion parameter is set as µ := 2. he returns are displayed below: time period t reward on investment he optimal solution is x = (0, , , 0) and the corresponding optimal return is Supposethatthereturnsarenotknownprecisely. Weextendthevaluesofr tj tointervals[0.99r tj,.0r tj ], that is, the returns may vary independently and simultaneously within % tolerance. We calculate the following enclosure of the optimal solutions x () = ([0,0.699], [0.762,], [0,0.8], [0,0.2379]). Even though the result is very conservative, we can conclude some interesting properties. Regardless the setting of values from intervals, we can be sure that at least 75% of the optimal portfolio is directed to the second investment. he remaining investments constitute at most 7%, 8% and 24% of the portfolio, respectively. Now, let study 5% perturbations of all returns except for the second investment. hus, we replace the values r jt by intervals [0.95r tj,.05r tj ], j =,3,4, t =,...,. he calculated enclosure for the optimal solutions is very tight: x (2) = ([0,0.0495], [0.9276,0.972], [0,0.053], [0,0.0724]). Notice that this problem setting is not B-stable [9], that is, there is no basis being optimal for each interval realization. For instance, putting the fourth investment returns to the upper limit and the others to the lower limit, we get a solution x = (0, , 0, ). In this setting, it is optimal to invest in the fourth subject instead of the third one. 3 Conclusion We proposed a method for contracting an interval enclosure of the optimal solution set. It was based on a linearization and iterative refinement. Even though it doesn t converge to the optimal bounds in general, it gives a sufficiently tight enclosure in short time. hus it can be used as a method for solving interval linear programming itself, or a first step in more involved algorithms. he method seems to converge quickly in spite of a very huge initial enclosure. However, to decrease the number of iterations, we will address the future research to finding a more appropriate initial enclosure. We will also carry out more numerical experiments to empirically verify the convergence speed

6 Acknowledgements he author was partially supported by the Czech Science Foundation Grant P403/2/947. References [] Beaumont, O.: Solving interval linear systems with linear programming techniques. Linear Algebra and its Applications 28 (998), [2] Beeck, H.: Linear programming with inexact data. technical report UM-ISU-7830, echnical University of Munich, Munich, 978. [3] Černý, M.: Computational aspects of regression analysis of interval data. Proceedings of World Academy of Science, Engineering and echnology 8 (20), [4] Fiedler, M., Nedoma, J., Ramík, J., Rohn, J., and Zimmermann, K.: Linear optimization problems with inexact data. Springer, New York, [5] Gabrel, V. and Murat, C.: Robustness and duality in linear programming. Journal of the Operational Research Society 6 (200), [6] Hladík, M.: olerances in portfolio selection via interval linear programming. In: CD-ROM Proceedings 26-th Int. Conf. Mathematical Methods in Economics MME08, Liberec, Czech Republic, 2008, [7] Hladík, M.: Optimal value range in interval linear programming. Fuzzy Optimization and Decision Making 8 (2009), [8] Hladík, M.: How to determine basis stability in interval linear programming. echnical report KAM- DIMAIA Series( ), Department of Applied Mathematics, Charles University, Prague, 200. [9] Hladík, M.: Interval linear programming: A survey. echnical report KAM-DIMAIA Series (200-98), Department of Applied Mathematics, Charles University, Prague, 200. [0] Hladík, M.: Interval linear programming: A survey. In: Linear Programming - New Frontiers in heory and Applications (Z. A. Mann, ed.), Nova Science Publishers, New York, 202, chapter 2, [] Hladík, M. and Černý, M.: Interval regression by tolerance analysis approach. Fuzzy Sets and Systems 93 (202), [2] Jansson, C.: Rigorous lower and upper bounds in linear programming. SIAM Journal on Optimization 4 (2004), [3] Jansson, C.: A self-validating method for solving linear programming problems with interval input data. In: Scientific computation with automatic result verification (U. Kulisch, and H. J. Stetter, eds.), Springer, Wien, 988, [4] Koníčková, J.: Sufficient condition of basis stability of an interval linear programming problem. ZAMM, Zeitschrift fuer Angewandte Mathematik und Mechanik 8 (200), [5] Li, Y. P., Huang, G. H., Guo, P., Yang, Z. F., and Nie, S. L.: A dual-interval vertex analysis method and its application to environmental decision making under uncertainty. European Journal of Operational Research 200 (200), [6] Liu, S.-. and Kao, C.: Matrix games with interval data. Computers & Industrial Engineering 56 (2009), [7] Rohn, J.: Stability of the optimal basis of a linear program under uncertainty. Operations Research Letters 3 (993),

Mathematical finance and linear programming (optimization)

Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may

Minimize subject to. x S R

Chapter 12 Lagrangian Relaxation This chapter is mostly inspired by Chapter 16 of [1]. In the previous chapters, we have succeeded to find efficient algorithms to solve several important problems such

Interval Pseudo-Inverse Matrices and Interval Greville Algorithm

Interval Pseudo-Inverse Matrices and Interval Greville Algorithm P V Saraev Lipetsk State Technical University, Lipetsk, Russia psaraev@yandexru Abstract This paper investigates interval pseudo-inverse

Absolute Value Programming

Computational Optimization and Aplications,, 1 11 (2006) c 2006 Springer Verlag, Boston. Manufactured in The Netherlands. Absolute Value Programming O. L. MANGASARIAN olvi@cs.wisc.edu Computer Sciences

Lecture 3: Linear Programming Relaxations and Rounding

Lecture 3: Linear Programming Relaxations and Rounding 1 Approximation Algorithms and Linear Relaxations For the time being, suppose we have a minimization problem. Many times, the problem at hand can

4.6 Linear Programming duality

4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal

Duality of linear conic problems

Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least

What is Linear Programming?

Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to

Definition of a Linear Program

Definition of a Linear Program Definition: A function f(x 1, x,..., x n ) of x 1, x,..., x n is a linear function if and only if for some set of constants c 1, c,..., c n, f(x 1, x,..., x n ) = c 1 x 1

Institute of Computer Science. An Iterative Method for Solving Absolute Value Equations

Institute of Computer Science Academy of Sciences of the Czech Republic An Iterative Method for Solving Absolute Value Equations Vahideh Hooshyarbakhsh, Taher Lotfi, Raena Farhadsefat, and Jiri Rohn Technical

New approach to interval linear regression

New approach to interval linear regression Milan Hladík 1 Michal Černý 2 1 Faculty of Mathematics and Physics, Charles University in Prague, Czech Republic 2 Faculty of Computer Science and Statistics,

Integrating Benders decomposition within Constraint Programming

Integrating Benders decomposition within Constraint Programming Hadrien Cambazard, Narendra Jussien email: {hcambaza,jussien}@emn.fr École des Mines de Nantes, LINA CNRS FRE 2729 4 rue Alfred Kastler BP

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.

This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra

3. Linear Programming and Polyhedral Combinatorics

Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the

1 Solving LPs: The Simplex Algorithm of George Dantzig

Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725

Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T

24. The Branch and Bound Method

24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

1. Introduction Linear Programming for Optimization Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1.1 Definition Linear programming is the name of a branch of applied mathematics that

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT

CHAPTER 9. Integer Programming

CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral

A Log-Robust Optimization Approach to Portfolio Management

A Log-Robust Optimization Approach to Portfolio Management Dr. Aurélie Thiele Lehigh University Joint work with Ban Kawas Research partially supported by the National Science Foundation Grant CMMI-0757983

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S

LINEAR PROGRAMMING WITH ONLINE LEARNING

LINEAR PROGRAMMING WITH ONLINE LEARNING TATSIANA LEVINA, YURI LEVIN, JEFF MCGILL, AND MIKHAIL NEDIAK SCHOOL OF BUSINESS, QUEEN S UNIVERSITY, 143 UNION ST., KINGSTON, ON, K7L 3N6, CANADA E-MAIL:{TLEVIN,YLEVIN,JMCGILL,MNEDIAK}@BUSINESS.QUEENSU.CA

Robust Geometric Programming is co-np hard

Robust Geometric Programming is co-np hard André Chassein and Marc Goerigk Fachbereich Mathematik, Technische Universität Kaiserslautern, Germany Abstract Geometric Programming is a useful tool with a

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.

Regression Using Support Vector Machines: Basic Foundations

Regression Using Support Vector Machines: Basic Foundations Technical Report December 2004 Aly Farag and Refaat M Mohamed Computer Vision and Image Processing Laboratory Electrical and Computer Engineering

Electric Company Portfolio Optimization Under Interval Stochastic Dominance Constraints

4th International Symposium on Imprecise Probabilities and Their Applications, Pittsburgh, Pennsylvania, 2005 Electric Company Portfolio Optimization Under Interval Stochastic Dominance Constraints D.

Discrete Optimization

Discrete Optimization [Chen, Batson, Dang: Applied integer Programming] Chapter 3 and 4.1-4.3 by Johan Högdahl and Victoria Svedberg Seminar 2, 2015-03-31 Todays presentation Chapter 3 Transforms using

2.3 Convex Constrained Optimization Problems

42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions

The Conference Call Search Problem in Wireless Networks

The Conference Call Search Problem in Wireless Networks Leah Epstein 1, and Asaf Levin 2 1 Department of Mathematics, University of Haifa, 31905 Haifa, Israel. lea@math.haifa.ac.il 2 Department of Statistics,

A Robust Formulation of the Uncertain Set Covering Problem

A Robust Formulation of the Uncertain Set Covering Problem Dirk Degel Pascal Lutter Chair of Management, especially Operations Research Ruhr-University Bochum Universitaetsstrasse 150, 44801 Bochum, Germany

Module1. x 1000. y 800.

Module1 1 Welcome to the first module of the course. It is indeed an exciting event to share with you the subject that has lot to offer both from theoretical side and practical aspects. To begin with,

Minimizing the Number of Machines in a Unit-Time Scheduling Problem

Minimizing the Number of Machines in a Unit-Time Scheduling Problem Svetlana A. Kravchenko 1 United Institute of Informatics Problems, Surganova St. 6, 220012 Minsk, Belarus kravch@newman.bas-net.by Frank

Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach

Outline Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach The University of New South Wales SPOM 2013 Joint work with V. Jeyakumar, B.S. Mordukhovich and

Perron vector Optimization applied to search engines

Perron vector Optimization applied to search engines Olivier Fercoq INRIA Saclay and CMAP Ecole Polytechnique May 18, 2011 Web page ranking The core of search engines Semantic rankings (keywords) Hyperlink

Permutation Betting Markets: Singleton Betting with Extra Information

Permutation Betting Markets: Singleton Betting with Extra Information Mohammad Ghodsi Sharif University of Technology ghodsi@sharif.edu Hamid Mahini Sharif University of Technology mahini@ce.sharif.edu

JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS. Received December May 12, 2003; revised February 5, 2004

Scientiae Mathematicae Japonicae Online, Vol. 10, (2004), 431 437 431 JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS Ondřej Čepeka and Shao Chin Sung b Received December May 12, 2003; revised February

An Approximation Algorithm for Bounded Degree Deletion

An Approximation Algorithm for Bounded Degree Deletion Tomáš Ebenlendr Petr Kolman Jiří Sgall Abstract Bounded Degree Deletion is the following generalization of Vertex Cover. Given an undirected graph

Minkowski Sum of Polytopes Defined by Their Vertices

Minkowski Sum of Polytopes Defined by Their Vertices Vincent Delos, Denis Teissandier To cite this version: Vincent Delos, Denis Teissandier. Minkowski Sum of Polytopes Defined by Their Vertices. Journal

Scheduling Home Health Care with Separating Benders Cuts in Decision Diagrams

Scheduling Home Health Care with Separating Benders Cuts in Decision Diagrams André Ciré University of Toronto John Hooker Carnegie Mellon University INFORMS 2014 Home Health Care Home health care delivery

Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650

Week 5 Integral Polyhedra

Week 5 Integral Polyhedra We have seen some examples 1 of linear programming formulation that are integral, meaning that every basic feasible solution is an integral vector. This week we develop a theory

Nan Kong, Andrew J. Schaefer. Department of Industrial Engineering, Univeristy of Pittsburgh, PA 15261, USA

A Factor 1 2 Approximation Algorithm for Two-Stage Stochastic Matching Problems Nan Kong, Andrew J. Schaefer Department of Industrial Engineering, Univeristy of Pittsburgh, PA 15261, USA Abstract We introduce

Convex Programming Tools for Disjunctive Programs

Convex Programming Tools for Disjunctive Programs João Soares, Departamento de Matemática, Universidade de Coimbra, Portugal Abstract A Disjunctive Program (DP) is a mathematical program whose feasible

Equilibrium computation: Part 1

Equilibrium computation: Part 1 Nicola Gatti 1 Troels Bjerre Sorensen 2 1 Politecnico di Milano, Italy 2 Duke University, USA Nicola Gatti and Troels Bjerre Sørensen ( Politecnico di Milano, Italy, Equilibrium

Arrangements And Duality

Arrangements And Duality 3.1 Introduction 3 Point configurations are tbe most basic structure we study in computational geometry. But what about configurations of more complicated shapes? For example,

Equilibria and Dynamics of. Supply Chain Network Competition with Information Asymmetry. and Minimum Quality Standards

Equilibria and Dynamics of Supply Chain Network Competition with Information Asymmetry in Quality and Minimum Quality Standards Anna Nagurney John F. Smith Memorial Professor and Dong (Michelle) Li Doctoral

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

Practical Guide to the Simplex Method of Linear Programming

Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear

Big Data - Lecture 1 Optimization reminders

Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics

Convex Optimization SVM s and Kernel Machines

Convex Optimization SVM s and Kernel Machines S.V.N. Vishy Vishwanathan vishy@axiom.anu.edu.au National ICT of Australia and Australian National University Thanks to Alex Smola and Stéphane Canu S.V.N.

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number

ARTICLE IN PRESS. European Journal of Operational Research xxx (2004) xxx xxx. Discrete Optimization. Nan Kong, Andrew J.

A factor 1 European Journal of Operational Research xxx (00) xxx xxx Discrete Optimization approximation algorithm for two-stage stochastic matching problems Nan Kong, Andrew J. Schaefer * Department of

Proximal mapping via network optimization

L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:

Online Convex Optimization

E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting

Algebra I Credit Recovery

Algebra I Credit Recovery COURSE DESCRIPTION: The purpose of this course is to allow the student to gain mastery in working with and evaluating mathematical expressions, equations, graphs, and other topics,

Precalculus REVERSE CORRELATION. Content Expectations for. Precalculus. Michigan CONTENT EXPECTATIONS FOR PRECALCULUS CHAPTER/LESSON TITLES

Content Expectations for Precalculus Michigan Precalculus 2011 REVERSE CORRELATION CHAPTER/LESSON TITLES Chapter 0 Preparing for Precalculus 0-1 Sets There are no state-mandated Precalculus 0-2 Operations

Nonlinear Optimization: Algorithms 3: Interior-point methods

Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris Jean-Philippe.Vert@mines.org Nonlinear optimization c 2006 Jean-Philippe Vert,

Introduction to Convex Optimization for Machine Learning

Introduction to Convex Optimization for Machine Learning John Duchi University of California, Berkeley Practical Machine Learning, Fall 2009 Duchi (UC Berkeley) Convex Optimization for Machine Learning

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical

LAGRANGIAN RELAXATION TECHNIQUES FOR LARGE SCALE OPTIMIZATION

LAGRANGIAN RELAXATION TECHNIQUES FOR LARGE SCALE OPTIMIZATION Kartik Sivaramakrishnan Department of Mathematics NC State University kksivara@ncsu.edu http://www4.ncsu.edu/ kksivara SIAM/MGSA Brown Bag

MATH 095, College Prep Mathematics: Unit Coverage Pre-algebra topics (arithmetic skills) offered through BSE (Basic Skills Education)

MATH 095, College Prep Mathematics: Unit Coverage Pre-algebra topics (arithmetic skills) offered through BSE (Basic Skills Education) Accurately add, subtract, multiply, and divide whole numbers, integers,

MATH BOOK OF PROBLEMS SERIES. New from Pearson Custom Publishing!

MATH BOOK OF PROBLEMS SERIES New from Pearson Custom Publishing! The Math Book of Problems Series is a database of math problems for the following courses: Pre-algebra Algebra Pre-calculus Calculus Statistics

Applied Algorithm Design Lecture 5

Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design

2004 Networks UK Publishers. Reprinted with permission.

Riikka Susitaival and Samuli Aalto. Adaptive load balancing with OSPF. In Proceedings of the Second International Working Conference on Performance Modelling and Evaluation of Heterogeneous Networks (HET

2. INEQUALITIES AND ABSOLUTE VALUES

2. INEQUALITIES AND ABSOLUTE VALUES 2.1. The Ordering of the Real Numbers In addition to the arithmetic structure of the real numbers there is the order structure. The real numbers can be represented by

On the representability of the bi-uniform matroid

On the representability of the bi-uniform matroid Simeon Ball, Carles Padró, Zsuzsa Weiner and Chaoping Xing August 3, 2012 Abstract Every bi-uniform matroid is representable over all sufficiently large

Section 1.1. Introduction to R n

The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to

Overview of Math Standards

Algebra 2 Welcome to math curriculum design maps for Manhattan- Ogden USD 383, striving to produce learners who are: Effective Communicators who clearly express ideas and effectively communicate with diverse

Dantzig-Wolfe bound and Dantzig-Wolfe cookbook

Dantzig-Wolfe bound and Dantzig-Wolfe cookbook thst@man.dtu.dk DTU-Management Technical University of Denmark 1 Outline LP strength of the Dantzig-Wolfe The exercise from last week... The Dantzig-Wolfe

Lecture 7: Approximation via Randomized Rounding

Lecture 7: Approximation via Randomized Rounding Often LPs return a fractional solution where the solution x, which is supposed to be in {0, } n, is in [0, ] n instead. There is a generic way of obtaining

Imprecise probabilities, bets and functional analytic methods in Łukasiewicz logic.

Imprecise probabilities, bets and functional analytic methods in Łukasiewicz logic. Martina Fedel joint work with K.Keimel,F.Montagna,W.Roth Martina Fedel (UNISI) 1 / 32 Goal The goal of this talk is to

Linear Programming. March 14, 2014

Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS KEITH CONRAD 1. Introduction The Fundamental Theorem of Algebra says every nonconstant polynomial with complex coefficients can be factored into linear

Lecture 2: August 29. Linear Programming (part I)

10-725: Convex Optimization Fall 2013 Lecture 2: August 29 Lecturer: Barnabás Póczos Scribes: Samrachana Adhikari, Mattia Ciollaro, Fabrizio Lecci Note: LaTeX template courtesy of UC Berkeley EECS dept.

Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method

Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming

Week 1: Introduction to Online Learning

Week 1: Introduction to Online Learning 1 Introduction This is written based on Prediction, Learning, and Games (ISBN: 2184189 / -21-8418-9 Cesa-Bianchi, Nicolo; Lugosi, Gabor 1.1 A Gentle Start Consider

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Chapter 1 Vocabulary identity - A statement that equates two equivalent expressions. verbal model- A word equation that represents a real-life problem. algebraic expression - An expression with variables.

Portfolio selection based on upper and lower exponential possibility distributions

European Journal of Operational Research 114 (1999) 115±126 Theory and Methodology Portfolio selection based on upper and lower exponential possibility distributions Hideo Tanaka *, Peijun Guo Department

General Framework for an Iterative Solution of Ax b. Jacobi s Method

2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,

Inverse Optimization by James Orlin

Inverse Optimization by James Orlin based on research that is joint with Ravi Ahuja Jeopardy 000 -- the Math Programming Edition The category is linear objective functions The answer: When you maximize

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation

Approximation Algorithms

Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2

4. Basic feasible solutions and vertices of polyhedra Due to the fundamental theorem of Linear Programming, to solve any LP it suffices to consider the vertices (finitely many) of the polyhedron P of the

Minimizing costs for transport buyers using integer programming and column generation. Eser Esirgen

MASTER STHESIS Minimizing costs for transport buyers using integer programming and column generation Eser Esirgen DepartmentofMathematicalSciences CHALMERS UNIVERSITY OF TECHNOLOGY UNIVERSITY OF GOTHENBURG

5.1 Bipartite Matching

CS787: Advanced Algorithms Lecture 5: Applications of Network Flow In the last lecture, we looked at the problem of finding the maximum flow in a graph, and how it can be efficiently solved using the Ford-Fulkerson

Support Vector Machines with Clustering for Training with Very Large Datasets

Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano

Solving polynomial least squares problems via semidefinite programming relaxations

Solving polynomial least squares problems via semidefinite programming relaxations Sunyoung Kim and Masakazu Kojima August 2007, revised in November, 2007 Abstract. A polynomial optimization problem whose

Study Guide 2 Solutions MATH 111

Study Guide 2 Solutions MATH 111 Having read through the sample test, I wanted to warn everyone, that I might consider asking questions involving inequalities, the absolute value function (as in the suggested

Efficient and Robust Allocation Algorithms in Clouds under Memory Constraints

Efficient and Robust Allocation Algorithms in Clouds under Memory Constraints Olivier Beaumont,, Paul Renaud-Goud Inria & University of Bordeaux Bordeaux, France 9th Scheduling for Large Scale Systems

Completely Positive Cone and its Dual

On the Computational Complexity of Membership Problems for the Completely Positive Cone and its Dual Peter J.C. Dickinson Luuk Gijben July 3, 2012 Abstract Copositive programming has become a useful tool

Approximation Algorithms: LP Relaxation, Rounding, and Randomized Rounding Techniques. My T. Thai

Approximation Algorithms: LP Relaxation, Rounding, and Randomized Rounding Techniques My T. Thai 1 Overview An overview of LP relaxation and rounding method is as follows: 1. Formulate an optimization

Scheduling and (Integer) Linear Programming

Scheduling and (Integer) Linear Programming Christian Artigues LAAS - CNRS & Université de Toulouse, France artigues@laas.fr Master Class CPAIOR 2012 - Nantes Christian Artigues Scheduling and (Integer)

Fuzzy Probability Distributions in Bayesian Analysis

Fuzzy Probability Distributions in Bayesian Analysis Reinhard Viertl and Owat Sunanta Department of Statistics and Probability Theory Vienna University of Technology, Vienna, Austria Corresponding author:

Understanding the Impact of Weights Constraints in Portfolio Theory

Understanding the Impact of Weights Constraints in Portfolio Theory Thierry Roncalli Research & Development Lyxor Asset Management, Paris thierry.roncalli@lyxor.com January 2010 Abstract In this article,

We shall turn our attention to solving linear systems of equations. Ax = b

59 Linear Algebra We shall turn our attention to solving linear systems of equations Ax = b where A R m n, x R n, and b R m. We already saw examples of methods that required the solution of a linear system

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology

FE670 Algorithmic Trading Strategies Lecture 6. Portfolio Optimization: Basic Theory and Practice Steve Yang Stevens Institute of Technology 10/03/2013 Outline 1 Mean-Variance Analysis: Overview 2 Classical