Chap 8. Linear Programming and Game Theory
|
|
- Mildred Harmon
- 7 years ago
- Views:
Transcription
1 Chap 8. Linear Programming and Game Theory KAIST wit Lab 유인철
2 I. Linear Inequalities
3 Introduction Linear Programming (from wikipedia) Mathematical method for determining a way to achieve the best outcome (such as maximum profit or lowest cost) in a given mathematical model for some list of requirements A technique for the optimization of a linear objective function, subject to some constraints Objective of this chapter To see the geometric meaning of linear inequalities To find the optimal soloution Hyperplane and Halfspace Hyperplane : Halfspace : 3
4 The Feasible Set and the Cost Function Feasible set the intersection of halfspaces composition of the solutions to a family of linear inequalities like Example) 24,, Fundamental constraint to linear programming
5 Problem in linear programming Find the point that lies in the feasible set The point is feasible minimizes or maximize the cost The point is optimal Cost increases Example : in the previous figure Cost : 2 3 Cost decreases Minimum cost : 2 3 6, 2 Vector (,2) is feasible and optimal The optimal vector occurs at a corner of the feasible set Possible categories. The feasible set is empty 2. The cost function is unbounded on the feasible set 3. The cost reaches its minimum (or maximum) on the feasible set 5
6 Slack Variables Slack Variable To change the inequality to an equation Result : equation + nonnegativity constraints Primal problem : Constraints : Example : in the previous example Inequality : 24 Slack variable : 24 Nonnegativity constraint : 6
7 The Diet Problem and Its Dual Diet problem in linear programming 2 sources of protein : A pound of peanut butter : a unit of protein A steak : 2 unit of protein Diet : at least 4 units are required Contains Cost : x pounds of peanut butter y steaks A pound of peanut butter : $2 A steak : $3 Constraints : 24,, the cost of whole diet : 2 3 optimal diet for minimum cost = 2 steaks (, 2) 7
8 Duality theorem The minimum in the given primal problem equals the maximum in its dual. Every linear program has dual. Example : Dual problem of previous one 2,23 optimal price : $.5 maximum revenue : 4 $6 8
9 Typical Applications. Production Planning : General motors Chevrolet Buick Cadillac profit $2 $3 $5 miles per gallon Assemble duration minute 2 minutes 3 minutes Also, the average car must get 8 miles per gallon What is the maximum profit in 8 hours (48 minutes)? 9
10 2. Portfolio Selection federal bonds municipals junk bonds interest 5% 6% 9% We can buy amounts,, not exceeding a total of $, No more than $2, can be invested in junk bonds The portfolio s average quality must be no lower than municipals What is the maximum interest?
11 Problem Set 8. Cost :. 4 2 Cost : 3 2,2 2 4, Cost : 2,2 4 4,
12 II. The Simplex Method
13 Introduction Linear Programming with n unknowns m constraints Minimum problem feasible vectors meet conditions Optimal occurs at the point where the planes first touch the feasible set Compute : Find all the corners of the feasible set Simplex method A systematic way to solve linear programs 3
14 The Geometry : Movement Along Edges Corner in linear algebra The meeting point of different planes Subspace of Possible intersection point of linear programming Review the feasible set : the intersection space of given by equation equations (by slack var.) + fundamental constraints equations (,, ) + fundamental constraints equations, but only equations are needed 4
15 How equations are chosen for intersection point? constraints : 26, 26,, 26 26,,, P : intersection of & 2 6 equations,, can be chosen out of equations The number of possible intersection points : C!!! 5
16 The intersection point is a genuine corner if It also satisfies the remaining inequality constraints. Otherwise, it is a complete fake. Ex) in the previous figure P : & 6, 6 J : & 3, 3 J 6
17 Key idea of simplex methods Go from corner to corner along the edges of the feasible set Edge of the feasible set If one of the intersecting planes is removed, remaining equations form an edge that comes out of the corner P : & 6, 6 : : 7
18 Newly defined linear programming by slack variable Slack variable : Equality constraints and nonnegativity Minimize, subject to, Newly defined Minimize, subject to, 8
19 Example. 9
20 The Simplex Algorithm Free variables vs. Basic variables in constraints Free variables : n components Basic variables (Pivot variables) : m components : equations pivots Basic feasible solutions of A solution is basic when of its components are zero, and it is feasible when it satisfies Setting the free variables to zero, basic solution is a genuine corner if its nonzero components are positive Phase of simplex method Phase I : find one basic feasible solution Phase II : move step by step to the optimal 2
21 Which corner do we go to next? basic variables remain basic Only basic variables will become free ( become zero), while other basic variables will stay positive Terminology Entering variable : the free variable which will become basic Leaving variable : the basic variable which will become free Ex) in the previous example If P moves to Q, Entering variable, : Leaving variable 2
22 Example 2 ),, 8, 9: genuine 7 3 There is possibility that cost can be minimized 2) As is increased, : entering variable : leaving variable 3) At this point,
23 Observation of Example 2 Quick Way, The right sides divided by the coefficients of the entering variable Ratio : 4, 3 hits zero first 23
24 The next step of Example 2 The current step ends at the new corner 2,,,,3 Free variable :,, Leaving variable Basic variable :, Entering variable Pivot by substituting 9 Repeat Phase II,,,,
25 Tableau (or large matrix) The Tableau One way to organize the simplex step into matrix Applying row operation to tableau Renumbering if necessary,, Tableau at corner Cost : Constraints :, :, : :, : 25
26 Ex) For Example 2 cost : 7 3 constraints :
27 Reduced tableau The basic variables will stand alone when elimination multiplies by Ex)
28 Fully reduced : R=rref(T) Subtract times the top block row from bottom row 2 2 Ex)
29 Meaning of matrix R Cost : Constraints : At corner : & Stopping test (optimality condition) r : coefficients of 29
30 Example 3. Figure 8.3 Minimize Subject to 26,26 <solve problem using tableau> Slack variable : , 3
31 ) Find initial feasible point Let basis variables(free variable) be zero 26 26, 6, 6: fake Find other feasible corner as initial point, 6, 6: genuine We choose this point as initial one 3
32 2) Renumbering Take basic variable(=non zeros) firstly This means that problem is changed as below , , 32
33 3) Make tableau and take row operation ,, at point 6 6, 33
34 4) Stopping test There is possibility that cost can be minimized Let be variable which becomes basic (=non zero) 3 34
35 5) Find the variable which becomes free(=zero) Let : a column of : the index of smallest ratio,, Then, becomes free variable (=zero) Ratio : : becomes zero 35
36 6) Find next solution Tableau is reconstructed by changing column and ,, /3 2/3 2/3 / /3 2/3 /3 2/3 /3 /
37 7) Stopping test again /3 2/3 /3 2/3 /3 / Since, this point is optimal. 4 at the point 2 2, 37
38 The Organization of a Simplex Step Reduced simplex method For computational purposes, only,, are used ) 2) 38
39 reduce the computation since has the form as below 39
40 Karmarkar s Method Wonho Kang 4
41 Definition Karmarkar s Method An iterative algorithm that, given an initial point and parameter, generates a sequence,,, which are the solutions of linear program (LP) Assumption. LP has a strictly feasible point, and the set of optimal point is bounded 2. LP has a special canonical form: minimize z subject to,, 3. The value of the objective at the optimum is known and is equal to zero, 4
42 Minimize Subject to Karmarkar s Method Minimize,,, Subject to, z Minimize,,, Subject to y,,,, 42
43 Minimize Subject to y Karmarkar s Method 43
44 Karmarkar s Method Canonical form minimize z subject to,, Denote nullspace of Ω Define simplex,,,, Denote center of the simplex,,, 44
45 Rewritten canonical form minimize z subject to Ω Karmarkar s Method Note that the constraint set (or feasible set) Ω can be represented as Ω,, Ω, Ω, 45
46 Karmarkar s Method Procedure. Initialize: Set ; /; 2. Update: Set Ψ where Ψ is an update map 3. Check stopping criterion: If the condition 2 is satisfied, then stop where is a given termination parameter; 4. Iterate: Set, go to 2 46
47 Karmarkar s Method First update step in the algorithm. Compute the orthogonal projector onto the nullspace of 2. Compute the normalized orthogonal projection of onto the nullspace of 3. Compute the steepest descent direction vector, / 4. Compute using where is the prespecified step size,, Karmarkar recommands a value of ¼ in his original paper 47
48 General update step. Compute the matrices Karmarkar s Method,, and is, in general, not at the center of the simplex, so whose diagonal entries are the components of the vector is used to transform this point to the center 2. Compute the orthogonal projector onto the nullspace of 48
49 General update step Karmarkar s Method 3. Compute the normalized orthogonal projection of onto the nullspace of 4. Compute the steepest descent direction vector 5. Compute using 6. Compute by applying the inverse transformation 49
50 Karmarkar s Method Example Minimize 3 3 subject to 2 3,, 3 3 Ω 2 3 Δ, 3 5
51 Solution. Initialize: 2. Set ; Karmarkar s Method 5. / /3 /3 /3 6.,, 7. 8.,, / / / / 5
52 Karmarkar s Method 2. Update: Set Ψ where Ψ is an update map ) Compute the orthogonal projector onto the nullspace of
53 Karmarkar s Method 2. Update: Set Ψ where Ψ is an update map 2) Compute the normalized orthogonal projection of onto the nullspace of
54 Karmarkar s Method 2. Update: Set Ψ where Ψ is an update map 3) Compute the steepest descent direction vector / / 6 7 is the radius of the largest sphere 8) inscribed in the set 9, 54
55 Karmarkar s Method 2. Update: Set Ψ where Ψ is an update map 3) Compute the steepest descent direction vector /
56 Karmarkar s Method 2. Update: Set Ψ where Ψ is an update map 4) Compute 5 /3 /3 / ) Note: makes be guaranteed 5) to lie in the constraint set 6,, 7 8) (a strictly interior point of the set), 56
57 3. Check stopping criterion: Karmarkar s Method 4. If the condition 2 is satisfied, then stop where is a given termination parameter /3 /3 / cf) /4 3/ if 2.4, then stop, or 4. Iterate: Set, go to 2 57
58 Karmarkar s Method c c d x () x ()
59 Karmarkar s Method /3 /3 /3 Minimize Minimize Subject to Subject to 59
60 Karmarkar s Method
61 Karmarkar s Method c c d x () x ()
62 Karmarkar s Method c c d x' () x' (2).8 x () x (2)
63 III. The Dual Problem
64 The Dual Problem Minimize Subject to, n feasible set Dual Maximize Subject to, y feasible set Ex) In Section 8., Minimize 2 Subject to Dual Maximize 4 Subject to
65 Duality theorem When both problems have feasible vectors, the minimum cost c the maximum income Weak duality If x and y are feasible in the primal and dual problems, then (proof) x and y are feasible and y 65
66 Reverse of duality theorem If the vectors x and y are feasible and, then x and y are optimal Ex) Minimize 4 subject to 2 6, , 6 7, 4 Dual Maximize 6 7 subject to 2 5, 3 4 : /2 66
67 Complementary slackness condition Relationship between the primal and dual Assume (P) means primal problem and have a optimal solution and (D) means dual problem and have a optimal solution, then. If in (P), then 2. If in (D), then (proof) c c should be always satisfied. However, from this equation, if, is 67
68 The proof of Duality From simplex method,, by reordering, In dual problem, if y is feasible If, optimal 68
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
More information4.6 Linear Programming duality
4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal
More informationChapter 6. Linear Programming: The Simplex Method. Introduction to the Big M Method. Section 4 Maximization and Minimization with Problem Constraints
Chapter 6 Linear Programming: The Simplex Method Introduction to the Big M Method In this section, we will present a generalized version of the simplex method that t will solve both maximization i and
More information1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
More informationLinear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.
1. Introduction Linear Programming for Optimization Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1.1 Definition Linear programming is the name of a branch of applied mathematics that
More informationLECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method
LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method Introduction to dual linear program Given a constraint matrix A, right
More informationIEOR 4404 Homework #2 Intro OR: Deterministic Models February 14, 2011 Prof. Jay Sethuraman Page 1 of 5. Homework #2
IEOR 4404 Homework # Intro OR: Deterministic Models February 14, 011 Prof. Jay Sethuraman Page 1 of 5 Homework #.1 (a) What is the optimal solution of this problem? Let us consider that x 1, x and x 3
More information3. Evaluate the objective function at each vertex. Put the vertices into a table: Vertex P=3x+2y (0, 0) 0 min (0, 5) 10 (15, 0) 45 (12, 2) 40 Max
SOLUTION OF LINEAR PROGRAMMING PROBLEMS THEOREM 1 If a linear programming problem has a solution, then it must occur at a vertex, or corner point, of the feasible set, S, associated with the problem. Furthermore,
More information1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.
Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S
More informationUsing the Simplex Method to Solve Linear Programming Maximization Problems J. Reeb and S. Leavengood
PERFORMANCE EXCELLENCE IN THE WOOD PRODUCTS INDUSTRY EM 8720-E October 1998 $3.00 Using the Simplex Method to Solve Linear Programming Maximization Problems J. Reeb and S. Leavengood A key problem faced
More informationSimplex method summary
Simplex method summary Problem: optimize a linear objective, subject to linear constraints 1. Step 1: Convert to standard form: variables on right-hand side, positive constant on left slack variables for
More informationSpecial Situations in the Simplex Algorithm
Special Situations in the Simplex Algorithm Degeneracy Consider the linear program: Maximize 2x 1 +x 2 Subject to: 4x 1 +3x 2 12 (1) 4x 1 +x 2 8 (2) 4x 1 +2x 2 8 (3) x 1, x 2 0. We will first apply the
More informationChapter 2 Solving Linear Programs
Chapter 2 Solving Linear Programs Companion slides of Applied Mathematical Programming by Bradley, Hax, and Magnanti (Addison-Wesley, 1977) prepared by José Fernando Oliveira Maria Antónia Carravilla A
More informationNonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
More informationLinear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
More informationLinear Programming Problems
Linear Programming Problems Linear programming problems come up in many applications. In a linear programming problem, we have a function, called the objective function, which depends linearly on a number
More informationOPRE 6201 : 2. Simplex Method
OPRE 6201 : 2. Simplex Method 1 The Graphical Method: An Example Consider the following linear program: Max 4x 1 +3x 2 Subject to: 2x 1 +3x 2 6 (1) 3x 1 +2x 2 3 (2) 2x 2 5 (3) 2x 1 +x 2 4 (4) x 1, x 2
More informationLinear Programming I
Linear Programming I November 30, 2003 1 Introduction In the VCR/guns/nuclear bombs/napkins/star wars/professors/butter/mice problem, the benevolent dictator, Bigus Piguinus, of south Antarctica penguins
More informationSolving Linear Programs
Solving Linear Programs 2 In this chapter, we present a systematic procedure for solving linear programs. This procedure, called the simplex method, proceeds by moving from one feasible solution to another,
More informationSolving Systems of Linear Equations
LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how
More informationMathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
More informationLinear Programming: Theory and Applications
Linear Programming: Theory and Applications Catherine Lewis May 11, 2008 1 Contents 1 Introduction to Linear Programming 3 1.1 What is a linear program?...................... 3 1.2 Assumptions.............................
More informationLinear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.
Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.
More informationWhat is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
More informationDuality in Linear Programming
Duality in Linear Programming 4 In the preceding chapter on sensitivity analysis, we saw that the shadow-price interpretation of the optimal simplex multipliers is a very useful concept. First, these shadow
More informationThis exposition of linear programming
Linear Programming and the Simplex Method David Gale This exposition of linear programming and the simplex method is intended as a companion piece to the article in this issue on the life and work of George
More information26 Linear Programming
The greatest flood has the soonest ebb; the sorest tempest the most sudden calm; the hottest love the coldest end; and from the deepest desire oftentimes ensues the deadliest hate. Th extremes of glory
More informationLinear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
More information3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the
More informationStandard Form of a Linear Programming Problem
494 CHAPTER 9 LINEAR PROGRAMMING 9. THE SIMPLEX METHOD: MAXIMIZATION For linear programming problems involving two variables, the graphical solution method introduced in Section 9. is convenient. However,
More informationChapter 6: Sensitivity Analysis
Chapter 6: Sensitivity Analysis Suppose that you have just completed a linear programming solution which will have a major impact on your company, such as determining how much to increase the overall production
More information9.4 THE SIMPLEX METHOD: MINIMIZATION
SECTION 9 THE SIMPLEX METHOD: MINIMIZATION 59 The accounting firm in Exercise raises its charge for an audit to $5 What number of audits and tax returns will bring in a maximum revenue? In the simplex
More informationSolutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
More information3.1 Solving Systems Using Tables and Graphs
Algebra 2 Chapter 3 3.1 Solve Systems Using Tables & Graphs 3.1 Solving Systems Using Tables and Graphs A solution to a system of linear equations is an that makes all of the equations. To solve a system
More informationOperation Research. Module 1. Module 2. Unit 1. Unit 2. Unit 3. Unit 1
Operation Research Module 1 Unit 1 1.1 Origin of Operations Research 1.2 Concept and Definition of OR 1.3 Characteristics of OR 1.4 Applications of OR 1.5 Phases of OR Unit 2 2.1 Introduction to Linear
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationLinear Programming in Matrix Form
Linear Programming in Matrix Form Appendix B We first introduce matrix concepts in linear programming by developing a variation of the simplex method called the revised simplex method. This algorithm,
More informationSensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS
Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and
More information4.1 Learning algorithms for neural networks
4 Perceptron Learning 4.1 Learning algorithms for neural networks In the two preceding chapters we discussed two closely related models, McCulloch Pitts units and perceptrons, but the question of how to
More informationLecture 2: August 29. Linear Programming (part I)
10-725: Convex Optimization Fall 2013 Lecture 2: August 29 Lecturer: Barnabás Póczos Scribes: Samrachana Adhikari, Mattia Ciollaro, Fabrizio Lecci Note: LaTeX template courtesy of UC Berkeley EECS dept.
More information10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method
578 CHAPTER 1 NUMERICAL METHODS 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS As a numerical technique, Gaussian elimination is rather unusual because it is direct. That is, a solution is obtained after
More informationMAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =
MAT 200, Midterm Exam Solution. (0 points total) a. (5 points) Compute the determinant of the matrix 2 2 0 A = 0 3 0 3 0 Answer: det A = 3. The most efficient way is to develop the determinant along the
More information2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system
1. Systems of linear equations We are interested in the solutions to systems of linear equations. A linear equation is of the form 3x 5y + 2z + w = 3. The key thing is that we don t multiply the variables
More information0.1 Linear Programming
0.1 Linear Programming 0.1.1 Objectives By the end of this unit you will be able to: formulate simple linear programming problems in terms of an objective function to be maximized or minimized subject
More informationHow To Understand And Solve A Linear Programming Problem
At the end of the lesson, you should be able to: Chapter 2: Systems of Linear Equations and Matrices: 2.1: Solutions of Linear Systems by the Echelon Method Define linear systems, unique solution, inconsistent,
More informationMath 215 HW #6 Solutions
Math 5 HW #6 Solutions Problem 34 Show that x y is orthogonal to x + y if and only if x = y Proof First, suppose x y is orthogonal to x + y Then since x, y = y, x In other words, = x y, x + y = (x y) T
More informationSolving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
More informationDate: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
More informationLinear Programming. April 12, 2005
Linear Programming April 1, 005 Parts of this were adapted from Chapter 9 of i Introduction to Algorithms (Second Edition) /i by Cormen, Leiserson, Rivest and Stein. 1 What is linear programming? The first
More informationLU Factorization Method to Solve Linear Programming Problem
Website: wwwijetaecom (ISSN 2250-2459 ISO 9001:2008 Certified Journal Volume 4 Issue 4 April 2014) LU Factorization Method to Solve Linear Programming Problem S M Chinchole 1 A P Bhadane 2 12 Assistant
More informationRow Echelon Form and Reduced Row Echelon Form
These notes closely follow the presentation of the material given in David C Lay s textbook Linear Algebra and its Applications (3rd edition) These notes are intended primarily for in-class presentation
More informationSolving Systems of Linear Equations
LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how
More informationLinear Programming. Solving LP Models Using MS Excel, 18
SUPPLEMENT TO CHAPTER SIX Linear Programming SUPPLEMENT OUTLINE Introduction, 2 Linear Programming Models, 2 Model Formulation, 4 Graphical Linear Programming, 5 Outline of Graphical Procedure, 5 Plotting
More informationLinear Equations ! 25 30 35$ & " 350 150% & " 11,750 12,750 13,750% MATHEMATICS LEARNING SERVICE Centre for Learning and Professional Development
MathsTrack (NOTE Feb 2013: This is the old version of MathsTrack. New books will be created during 2013 and 2014) Topic 4 Module 9 Introduction Systems of to Matrices Linear Equations Income = Tickets!
More information1 2 3 1 1 2 x = + x 2 + x 4 1 0 1
(d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which
More informationLinear Programming Notes VII Sensitivity Analysis
Linear Programming Notes VII Sensitivity Analysis 1 Introduction When you use a mathematical model to describe reality you must make approximations. The world is more complicated than the kinds of optimization
More informationDuality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
More informationLinearly Independent Sets and Linearly Dependent Sets
These notes closely follow the presentation of the material given in David C. Lay s textbook Linear Algebra and its Applications (3rd edition). These notes are intended primarily for in-class presentation
More informationQuestion 2: How do you solve a matrix equation using the matrix inverse?
Question : How do you solve a matrix equation using the matrix inverse? In the previous question, we wrote systems of equations as a matrix equation AX B. In this format, the matrix A contains the coefficients
More information1 Linear Programming. 1.1 Introduction. Problem description: motivate by min-cost flow. bit of history. everything is LP. NP and conp. P breakthrough.
1 Linear Programming 1.1 Introduction Problem description: motivate by min-cost flow bit of history everything is LP NP and conp. P breakthrough. general form: variables constraints: linear equalities
More informationLinear Programming II: Minimization 2006 Samuel L. Baker Assignment 11 is on page 16.
LINEAR PROGRAMMING II 1 Linear Programming II: Minimization 2006 Samuel L. Baker Assignment 11 is on page 16. Introduction A minimization problem minimizes the value of the objective function rather than
More informationMATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar
More information1.5 SOLUTION SETS OF LINEAR SYSTEMS
1-2 CHAPTER 1 Linear Equations in Linear Algebra 1.5 SOLUTION SETS OF LINEAR SYSTEMS Many of the concepts and computations in linear algebra involve sets of vectors which are visualized geometrically as
More informationRecall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
More informationLargest Fixed-Aspect, Axis-Aligned Rectangle
Largest Fixed-Aspect, Axis-Aligned Rectangle David Eberly Geometric Tools, LLC http://www.geometrictools.com/ Copyright c 1998-2016. All Rights Reserved. Created: February 21, 2004 Last Modified: February
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationStudy Guide 2 Solutions MATH 111
Study Guide 2 Solutions MATH 111 Having read through the sample test, I wanted to warn everyone, that I might consider asking questions involving inequalities, the absolute value function (as in the suggested
More information. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2
4. Basic feasible solutions and vertices of polyhedra Due to the fundamental theorem of Linear Programming, to solve any LP it suffices to consider the vertices (finitely many) of the polyhedron P of the
More informationOrthogonal Projections
Orthogonal Projections and Reflections (with exercises) by D. Klain Version.. Corrections and comments are welcome! Orthogonal Projections Let X,..., X k be a family of linearly independent (column) vectors
More informationAn Introduction to Linear Programming
An Introduction to Linear Programming Steven J. Miller March 31, 2007 Mathematics Department Brown University 151 Thayer Street Providence, RI 02912 Abstract We describe Linear Programming, an important
More informationLeast-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
More informationSolving Systems of Linear Equations Using Matrices
Solving Systems of Linear Equations Using Matrices What is a Matrix? A matrix is a compact grid or array of numbers. It can be created from a system of equations and used to solve the system of equations.
More informationThe Graphical Method: An Example
The Graphical Method: An Example Consider the following linear program: Maximize 4x 1 +3x 2 Subject to: 2x 1 +3x 2 6 (1) 3x 1 +2x 2 3 (2) 2x 2 5 (3) 2x 1 +x 2 4 (4) x 1, x 2 0, where, for ease of reference,
More informationChapter 3: Section 3-3 Solutions of Linear Programming Problems
Chapter 3: Section 3-3 Solutions of Linear Programming Problems D. S. Malik Creighton University, Omaha, NE D. S. Malik Creighton University, Omaha, NE Chapter () 3: Section 3-3 Solutions of Linear Programming
More information7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
More informationReduced echelon form: Add the following conditions to conditions 1, 2, and 3 above:
Section 1.2: Row Reduction and Echelon Forms Echelon form (or row echelon form): 1. All nonzero rows are above any rows of all zeros. 2. Each leading entry (i.e. left most nonzero entry) of a row is in
More informationMATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More informationThese axioms must hold for all vectors ū, v, and w in V and all scalars c and d.
DEFINITION: A vector space is a nonempty set V of objects, called vectors, on which are defined two operations, called addition and multiplication by scalars (real numbers), subject to the following axioms
More information2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
More informationLecture Notes 2: Matrices as Systems of Linear Equations
2: Matrices as Systems of Linear Equations 33A Linear Algebra, Puck Rombach Last updated: April 13, 2016 Systems of Linear Equations Systems of linear equations can represent many things You have probably
More informationMATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column
More informationInternational Doctoral School Algorithmic Decision Theory: MCDA and MOO
International Doctoral School Algorithmic Decision Theory: MCDA and MOO Lecture 2: Multiobjective Linear Programming Department of Engineering Science, The University of Auckland, New Zealand Laboratoire
More information24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
More informationArrangements And Duality
Arrangements And Duality 3.1 Introduction 3 Point configurations are tbe most basic structure we study in computational geometry. But what about configurations of more complicated shapes? For example,
More information56:171 Operations Research Midterm Exam Solutions Fall 2001
56:171 Operations Research Midterm Exam Solutions Fall 2001 True/False: Indicate by "+" or "o" whether each statement is "true" or "false", respectively: o_ 1. If a primal LP constraint is slack at the
More informationIntroductory Notes on Demand Theory
Introductory Notes on Demand Theory (The Theory of Consumer Behavior, or Consumer Choice) This brief introduction to demand theory is a preview of the first part of Econ 501A, but it also serves as a prototype
More informationAlgebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.
Chapter 1 Vocabulary identity - A statement that equates two equivalent expressions. verbal model- A word equation that represents a real-life problem. algebraic expression - An expression with variables.
More informationLecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method
Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming
More informationLinear Programming Supplement E
Linear Programming Supplement E Linear Programming Linear programming: A technique that is useful for allocating scarce resources among competing demands. Objective function: An expression in linear programming
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
More informationThe Geometry of Polynomial Division and Elimination
The Geometry of Polynomial Division and Elimination Kim Batselier, Philippe Dreesen Bart De Moor Katholieke Universiteit Leuven Department of Electrical Engineering ESAT/SCD/SISTA/SMC May 2012 1 / 26 Outline
More informationChapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
More informationMATH2210 Notebook 1 Fall Semester 2016/2017. 1 MATH2210 Notebook 1 3. 1.1 Solving Systems of Linear Equations... 3
MATH0 Notebook Fall Semester 06/07 prepared by Professor Jenny Baglivo c Copyright 009 07 by Jenny A. Baglivo. All Rights Reserved. Contents MATH0 Notebook 3. Solving Systems of Linear Equations........................
More informationLECTURE: INTRO TO LINEAR PROGRAMMING AND THE SIMPLEX METHOD, KEVIN ROSS MARCH 31, 2005
LECTURE: INTRO TO LINEAR PROGRAMMING AND THE SIMPLEX METHOD, KEVIN ROSS MARCH 31, 2005 DAVID L. BERNICK dbernick@soe.ucsc.edu 1. Overview Typical Linear Programming problems Standard form and converting
More informationa 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
More informationDecember 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
More informationUnit 1 Equations, Inequalities, Functions
Unit 1 Equations, Inequalities, Functions Algebra 2, Pages 1-100 Overview: This unit models real-world situations by using one- and two-variable linear equations. This unit will further expand upon pervious
More information