11. APPROXIMATION ALGORITHMS

Save this PDF as:

Size: px
Start display at page:

Transcription

1 Coping with NP-completeness 11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Q. Suppose I need to solve an NP-hard problem. What should I do? A. Sacrifice one of three desired features. i. Solve arbitrary instances of the problem. ii. Solve problem to optimality. iii. Solve problem in polynomial time. ρ-approximation algorithm. Guaranteed to run in poly-time. Guaranteed to solve arbitrary instance of the problem Guaranteed to find solution within ratio ρ of true optimum. Lecture slides by Kevin Wayne Copyright 005 Pearson-Addison Wesley Challenge. Need to prove a solution's value is close to optimum, without even knowing what optimum value is Last updated on /15/17 6:00 PM Load balancing 11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Input. m identical machines; n jobs, job j has processing time t j. Job j must run contiguously on one machine. A machine can process at most one job at a time. Def. Let J(i) be the subset of jobs assigned to machine i. The load of machine i is L i = Σ j J(i) t j. Def. The makespan is the maximum load on any machine L = max i L i. Load balancing. Assign each job to a machine to minimize makespan. machine 1 a dmachine 1 f machine b c Machine e g 0 L1 L time

2 Load balancing on machines is NP-hard Load balancing: list scheduling Claim. Load balancing is hard even if only machines. Pf. PARTITION P LOAD-BALANCE. NP-complete by Exercise 8.6 List-scheduling algorithm. Consider n jobs in some fixed order. Assign job j to machine whose load is smallest so far. List-Scheduling(m, n, t 1,t,,t n ) { e a b f c g d for i = 1 to m { L i 0 } J(i) load on machine i jobs assigned to machine i length of job f for j = 1 to n { i = argmin k L k J(i) J(i) {j} machine i has smallest load assign job j to machine i L i L i + t j update load of machine i machine 1 machine a dmachine 1 f b c Machine e g yes } } return J(1),, J(m) Implementation. O(n log m) using a priority queue. 0 L time 5 6 Load balancing: list scheduling analysis Believe it or not Theorem. [Graham 1966] Greedy algorithm is a -approximation. First worst-case analysis of an approximation algorithm. Need to compare resulting solution with optimal makespan L*. Lemma 1. The optimal makespan L* max j t j. Pf. Some machine must process the most time-consuming job. Lemma. The optimal makespan Pf. The total processing time is Σ j t j. L * m 1 j t j. One of m machines must do at least a 1 / m fraction of total work. L * m 1 j t j. 7 8

4 Load balancing: LPT rule Load balancing: LPT rule Longest processing time (LPT). Sort n jobs in descending order of processing time, and then run list scheduling algorithm. Observation. If at most m jobs, then list-scheduling is optimal. Pf. Each job put on its own machine. Lemma 3. If there are more than m jobs, L* t m+1. LPT-List-Scheduling(m, n, t 1,t,,t n ) { Sort jobs so that t 1 t for i = 1 to m { } L i 0 J(i) t n load on machine i jobs assigned to machine i Pf. Consider first m+1 jobs t 1,, t m+1. Since the t i 's are in descending order, each takes at least t m+1 time. There are m + 1 jobs and m machines, so by pigeonhole principle, at least one machine gets two jobs. for j = 1 to n { i = argmin k L k machine i has smallest load Theorem. LPT rule is a 3/-approximation algorithm. Pf. Same basic approach as for list scheduling. } J(i) J(i) {j} L i L i + t j } return J(1),, J(m) assign job j to machine i update load of machine i 13 L i = (L i t j )!# "# \$ + t j % 3 L *. L* 1 L* Lemma 3 ( by observation, can assume number of jobs > m ) 1 Load Balancing: LPT rule Q. Is our 3/ analysis tight? A. No. Theorem. [Graham 1969] LPT rule is a /3-approximation. Pf. More sophisticated analysis of same algorithm. Q. Is Graham's /3 analysis tight? A. Essentially yes. 11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Ex. m machines n = m + 1 jobs jobs of length m, m + 1,, m 1 and one more job of length m. Then, L / L* = (m 1) / (3m) 15

5 Center selection problem Center selection problem Input. Set of n sites s 1,, s n and an integer k > 0. Input. Set of n sites s 1,, s n and an integer k > 0. Center selection problem. Select set of k centers C so that maximum distance r(c) from a site to nearest center is minimized. Center selection problem. Select set of k centers C so that maximum distance r(c) from a site to nearest center is minimized. r(c) k = centers Notation. dist(x, y) = distance between sites x and y. dist(s i, C) = min c C dist(s i, c) = distance from s i to closest center. r(c) = max i dist(s i, C) = smallest covering radius. Goal. Find set of centers C that minimizes r(c), subject to C = k. center site Distance function properties. dist(x, x) = 0 [ identity ] dist(x, y) = dist(y, x) [ symmetry ] dist(x, y) dist(x, z) + dist(z, y) [ triangle inequality ] Center selection example Greedy algorithm: a false start Ex: each site is a point in the plane, a center can be any point in the plane, dist(x, y) = Euclidean distance. Remark: search can be infinite! Greedy algorithm. Put the first center at the best possible location for a single center, and then keep adding centers so as to reduce the covering radius each time by as much as possible. Remark: arbitrarily bad! k = centers k = centers r(c) greedy center 1 center site center site 19 0

6 Center selection: greedy algorithm Center selection: analysis of greedy algorithm Repeatedly choose next center to be site farthest from any existing center. GREEDY-CENTER-SELECTION (k, n, s 1, s,, s n) C. REPEAT k times Select a site s i with maximum distance dist(s i, C). C C s i. RETURN C. site farthest from any center Lemma. Let C* be an optimal set of centers. Then r(c) r(c*). Pf. [by contradiction] Assume r(c*) < ½ r(c). For each site c i C, consider ball of radius ½ r(c) around it. Exactly one c * i in each ball; let c i be the site paired with c i *. Consider any site s and its closest center c * i C*. dist(s, C) dist(s, c i ) dist(s, c i *) + dist(c i *, c i ) r(c*). Thus, r(c) r(c*). r(c*) since c i * is closest center Δ-inequality ½ r(c) ½ r(c) ½ r(c) c i Property. Upon termination, all centers in C are pairwise at least r(c) apart. Pf. By construction of algorithm. C* site s c i * 1 Center selection Dominating set reduces to center selection Lemma. Let C* be an optimal set of centers. Then r(c) r (C*). Theorem. Greedy algorithm is a -approximation for center selection problem. Remark. Greedy algorithm always places centers at sites, but is still within a factor of of best solution that is allowed to place centers anywhere. Question. Is there hope of a 3/-approximation? /3? e.g., points in the plane Theorem. Unless P = NP, there no ρ-approximation for center selection problem for any ρ <. Pf. We show how we could use a ( ε) approximation algorithm for CENTER-SELECTION selection to solve DOMINATING-SET in poly-time. Let G = (V, E), k be an instance of DOMINATING-SET. Construct instance G' of CENTER-SELECTION with sites V and distances dist(u, v) = 1 if (u, v) E dist(u, v) = if (u, v) E Note that G' satisfies the triangle inequality. G has dominating set of size k iff there exists k centers C* with r(c*) = 1. Thus, if G has a dominating set of size k, a ( ε)-approximation algorithm for CENTER-SELECTION would find a solution C* with r(c*) = 1 since it cannot use any edge of distance. 3

7 Weighted vertex cover 11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Definition. Given a graph G = (V, E), a vertex cover is a set S V such that each edge in E has at least one end in S. Weighted vertex cover. Given a graph G with vertex weights, find a vertex cover of minimum weight. 9 9 weight = + + weight = 11 6 Pricing method Pricing method Pricing method. Each edge must be covered by some vertex. Edge e = (i, j) pays price p e 0 to use both vertex i and j. Set prices and find vertex cover simultaneously. Fairness. Edges incident to vertex i should pay w i in total. WEIGHTED-VERTEX-COVER (G, w) for each vertex i : p e e=(i, j) w i S. FOREACH e E p e 0. 9 Fairness lemma. For any vertex cover S and any fair prices pe : e pe w(s). Pf. WHILE (there exists an edge (i, j) such that neither i nor j is tight) Select such an edge e = (i, j). Increase pe as much as possible until i or j tight. S set of all tight nodes. RETURN S. each edge e covered by at least one node in S sum fairness inequalities for each node in S 7 8

8 Pricing method example Pricing method: analysis Theorem. Pricing method is a -approximation for WEIGHTED-VERTEX-COVER. Pf. Algorithm terminates since at least one new node becomes tight after each iteration of while loop. Let S = set of all tight nodes upon termination of algorithm. S is a vertex cover: if some edge (i, j) is uncovered, then neither i nor j is tight. But then while loop would not terminate. vertex weight price of edge a-b Let S* be optimal vertex cover. We show w(s) w(s*). w(s) = w i = i S i S p e e=(i, j) i V p e e=(i, j) = p e w(s*). e E all nodes in S are tight S V, prices 0 each edge counted twice fairness lemma 9 30 Weighted vertex cover 11. APPROXIMATION ALGORITHMS Given a graph G = (V, E) with vertex weights w i 0, find a min weight subset of vertices S V such that every edge is incident to at least one vertex in S. load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem total weight = = 57 3

9 Weighted vertex cover: ILP formulation Weighted vertex cover: ILP formulation Given a graph G = (V, E) with vertex weights w i 0, find a min weight subset of vertices S V such that every edge is incident to at least one vertex in S. Integer linear programming formulation. Model inclusion of each vertex i using a 0/1 variable x i. x i = " # \$ Vertex covers in 1 1 correspondence with 0/1 assignments: S = { i V : x i = 1}. 0 if vertex i is not in vertex cover 1 if vertex i is in vertex cover Objective function: minimize Σ i w i x i. Weighted vertex cover. Integer linear programming formulation. (ILP) i V w i x i x i + x j 1 (i, j) E x i {0, 1} i V Observation. If x* is optimal solution to (ILP), then S = { i V : x i * = 1} is a min weight vertex cover. For every edge (i, j), must take either vertex i or j (or both): x i + x j Integer linear programming Linear programming Given integers a ij, b i, and c j, find integers x j that satisfy: c T x Ax b x 0 x Observation. Vertex cover formulation proves that INTEGER-PROGRAMMING is an NP-hard search problem. n j=1 n j=1 c j x j a ij x j b i 1 i m x j 0 1 j n x j 1 j n Given integers a ij, b i, and c j, find real numbers x j that satisfy: c T x Ax b x 0 Linear. No x, xy, arccos(x), x(1 x), etc. Simplex algorithm. [Dantzig 197] Can solve LP in practice. Ellipsoid algorithm. [Khachian 1979] Can solve LP in poly-time. Interior point algorithms. [Karmarkar 198, Renegar 1988, ] Can solve LP in poly-time and in practice. n j=1 n j=1 c j x j a ij x j b i 1 i m x j 0 1 j n 35 36

10 LP feasible region Weighted vertex cover: LP relaxation LP geometry in D. Linear programming relaxation. x 1 = 0 (LP ) i V w i x i x i + x j 1 (i, j) E x i 0 i V Observation. Optimal value of (LP) is optimal value of (ILP). Pf. LP has fewer constraints. Note. LP is not equivalent to vertex cover. ½ ½ x 1 + x = 6 x = 0 x 1 + x = 6 Q. How can solving LP help us find a small vertex cover? A. Solve LP and round fractional values. ½ Weighted vertex cover: LP rounding algorithm Weighted vertex cover inapproximability Lemma. If x* is optimal solution to (LP), then S = { i V : x i * ½} is a vertex cover whose weight is at most twice the min possible weight. Theorem. [Dinur Safra 00] If P NP, then no ρ-approximation for WEIGHTED-VERTEX-COVER for any ρ < (even if all weights are 1). Pf. [S is a vertex cover] Consider an edge (i, j) E. Since x i * + x j * 1, either x i * ½ or x j * ½ (or both) (i, j) covered. Pf. [S has desired cost] Let S* be optimal vertex cover. Then w i i S* * w i x i 1 w i i S LP is a relaxation xi* ½ i S Theorem. The rounding algorithm is a -approximation algorithm. Pf. Lemma + fact that LP can be solved in poly-time. On the Hardness of Approximating Minimum Vertex Cover Irit Dinur Samuel Safra May 6, 00 Abstract We prove the Minimum Vertex Cover problem to be NP-hard to approximate to within a factor of , extending on previous PCP and hardness of approximation technique. To that end, one needs to develop a new proof framework, and borrow and extend ideas from several fields. Open research problem. Close the gap. 39 0

11 Generalized load balancing 11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Input. Set of m machines M; set of n jobs J. Job j J must run contiguously on an authorized machine in M j M. Job j J has processing time t j. Each machine can process at most one job at a time. Def. Let J(i) be the subset of jobs assigned to machine i. The load of machine i is L i = Σ j J(i) t j. Def. The makespan is the maximum load on any machine = max i L i. Generalized load balancing. Assign each job to an authorized machine to minimize makespan. Generalized load balancing: integer linear program and relaxation Generalized load balancing: lower bounds ILP formulation. x ij = time machine i spends processing job j. (IP) min L s. t. x i j = t j for all j J i L for all i M x i j j x i j {0, t j } for all j J and i M j Lemma 1. The optimal makespan L* max j t j. Pf. Some machine must process the most time-consuming job. Lemma. Let L be optimal value to the LP. Then, optimal makespan L* L. Pf. LP has fewer constraints than IP formulation. x i j = 0 for all j J and i M j LP relaxation. (LP) min L s. t. x i j = t j for all j J i L for all i M x i j j x i j 0 for all j J and i M j x i j = 0 for all j J and i M j 3

12 Generalized load balancing: structure of LP solution Generalized load balancing: rounding Lemma 3. Let x be solution to LP. Let G(x) be the graph with an edge between machine i and job j if x ij > 0. Then G(x) is acyclic. Pf. (deferred) can transform x into another LP solution where G(x) is acyclic if LP solver doesn't return such an x Rounded solution. Find LP solution x where G(x) is a forest. Root forest G(x) at some arbitrary machine node r. If job j is a leaf node, assign j to its parent machine i. If job j is not a leaf node, assign j to any one of its children. Lemma. Rounded solution only assigns jobs to authorized machines. Pf. If job j is assigned to machine i, then x ij > 0. LP solution can only assign positive value to authorized machines. x ij > 0 G(x) acyclic job G(x) cyclic job machine machine 5 6 Generalized load balancing: analysis Generalized load balancing: analysis Lemma 5. If job j is a leaf node and machine i = parent(j), then x ij = t j. Pf. Since i is a leaf, x ij = 0 for all j parent(i). LP constraint guarantees Σ i x ij = t j. Theorem. Rounded solution is a -approximation. Pf. Let J(i) be the jobs assigned to machine i. By LEMMA 6, the load L i on machine i has two components: Lemma 6. At most one non-leaf job is assigned to a machine. Pf. The only possible non-leaf job assigned to machine i is parent(i). leaf nodes: t j j J(i) j is a leaf = x ij x ij L L * Lemma 1 Lemma 5 j J(i) j is a leaf j J LP Lemma (LP is a relaxation) optimal value of LP parent: t parent(i) L * job machine Thus, the overall load L i L*. 7 8

13 Generalized load balancing: flow formulation Generalized load balancing: structure of solution Flow formulation of LP. x i j i x i j j = t j for all j J L for all i M x i j 0 for all j J and i M j x i j = 0 for all j J and i M j Lemma 3. Let (x, L) be solution to LP. Let G(x) be the graph with an edge from machine i to job j if x ij > 0. We can find another solution (x', L) such that G(x') is acyclic. Pf. Let C be a cycle in G(x). Augment flow along the cycle C. flow conservation maintained At least one edge from C is removed (and none are added). Repeat until G(x') is acyclic. 3 3 augment flow along cycle C Observation. Solution to feasible flow problem with value L are in 1-to-1 correspondence with LP solutions of value L. 1 3 G(x) 5 1 G(x') Conclusions Running time. The bottleneck operation in our -approximation is solving one LP with m n + 1 variables. Remark. Can solve LP using flow techniques on a graph with m + n + 1 nodes: given L, find feasible flow if it exists. Binary search to find L*. Extensions: unrelated parallel machines. [Lenstra Shmoys Tardos 1990] Job j takes t ij time if processed on machine i. -approximation algorithm via LP rounding. If P NP, then no no ρ-approximation exists for any ρ < 3/. 11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Mathematical Programming 6 (1990) North-Holland APPROXIMATION ALGORITHMS FOR SCHEDULING UNRELATED PARALLEL MACHINES Jan Karel LENSTRA Eindhoven University of Technology, Eindhoven, The Netherlands, and Centre for Mathematics and Computer Science, Amsterdam, The Netherlands David B. SHMOYS and l~va TARDOS Cornell University, Ithaca, NY, USA 51

14 Polynomial-time approximation scheme Knapsack problem PTAS. (1 + ε)-approximation algorithm for any constant ε > 0. Load balancing. [Hochbaum Shmoys 1987] Euclidean TSP. [Arora, Mitchell 1996] Consequence. PTAS produces arbitrarily high quality solution, but trades off accuracy for time. Knapsack problem. Given n objects and a knapsack. Item i has value v i > 0 and weighs w i > 0. Knapsack has weight limit W. Goal: fill knapsack so as to maximize total value. Ex: { 3, } has value 0. we assume wi W for each i This section. PTAS for knapsack problem via rounding and scaling. item value weight original instance (W = 11) 53 5 Knapsack is NP-complete Knapsack problem: dynamic programming I KNAPSACK. Given a set X, weights w i 0, values v i 0, a weight limit W, and a target value V, is there a subset S X such that: SUBSET-SUM. Given a set X, values u i 0, and an integer U, is there a subset S X whose elements sum to exactly U? Theorem. SUBSET-SUM P KNAPSACK. W w i i S v i i S V Pf. Given instance (u1,, un, U) of SUBSET-SUM, create KNAPSACK instance: v i = w i = u i V = W = U u i i S u i i S U U Def. OPT(i, w) = max value subset of items 1,..., i with weight limit w. Case 1. OPT does not select item i. OPT selects best of 1,, i 1 using up to weight limit w. Case. OPT selects item i. New weight limit = w w i. OPT selects best of 1,, i 1 using up to weight limit w w i. # 0 if i = 0 % OPT(i, w) = \$ OPT(i 1, w) if w i > w % & max{ OPT(i 1, w), v i + OPT(i 1, w w i )} otherwise Theorem. Computes the optimal value in O(n W) time. Not polynomial in input size. Polynomial in input size if weights are small integers

15 Knapsack problem: dynamic programming II Knapsack problem: dynamic programming II Def. OPT(i, v) = min weight of a knapsack for which we can obtain a solution of value v using a subset of items 1,..., i. Note. Optimal value is the largest value v such that OPT(n, v) W. Case 1. OPT does not select item i. OPT selects best of 1,, i 1 that achieves value v. Case. OPT selects item i. Consumes weight w i, need to achieve value v v i. OPT selects best of 1,, i 1 that achieves value v v i. Theorem. Dynamic programming algorithm II computes the optimal value in O(n vmax) time, where vmax is the maximum of any value. Pf. The optimal value V* n v max. There is one subproblem for each item and for each value v V*. It takes O(1) time per subproblem. Remark 1. Not polynomial in input size! Remark. Polynomial time if values are small integers. OPT(i, v) = 0 v 0 i =0 v>0 min {OPT(i 1,v), w i + OPT(i 1,v v i )} Knapsack problem: polynomial-time approximation scheme Knapsack problem: polynomial-time approximation scheme Intuition for approximation algorithm. Round all values up to lie in smaller range. Run dynamic programming algorithm II on rounded/scaled instance. Return optimal items in rounded instance. item value weight item value weight Round up all values: 0 < ε 1 = precision parameter. v max = largest value in original instance. θ = scaling factor = ε v max / n. v i = Observation. Optimal solutions to problem with v are equivalent to optimal solutions to problem with v ˆ. Intuition. v close to v so optimal solution using v is nearly optimal; v ˆ small and integral so dynamic programming algorithm II is fast. v i, ˆv i = v i original instance (W = 11) rounded instance (W = 11) 59 60

16 Knapsack problem: polynomial-time approximation scheme Knapsack problem: polynomial-time approximation scheme Theorem. If S is solution found by rounding algorithm and S* is any other feasible solution, then Pf. Let S* be any feasible solution satisfying weight constraint. v i v i (1 + ) always round up v i v i subset containing only the item of largest value Theorem. For any ε > 0, the rounding algorithm computes a feasible solution whose value is within a (1 + ε) factor of the optimum in O(n 3 / ε) time. Pf. We have already proved the accuracy bound. Dynamic program II running time is O(n v ˆ max ), where i i S S v i (v i + ) solve rounded instance optimally never round up by more than θ choosing S* = { max } v v i + 1 v ˆv max = v max = n = v i + n S n v i + 1 v θ = ε v max / n v i + 1 v thus v v i (1 + ) i S v i v max Σ i S v i 61 6

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.

Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of

11. APPROXIMATION ALGORITHMS

11. APPROXIMATION ALGORITHMS load balancing center selection pricing method: vertex cover LP rounding: vertex cover generalized load balancing knapsack problem Lecture slides by Kevin Wayne Copyright 2005

Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.

Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three

Algorithm Design and Analysis

Algorithm Design and Analysis LECTURE 27 Approximation Algorithms Load Balancing Weighted Vertex Cover Reminder: Fill out SRTEs online Don t forget to click submit Sofya Raskhodnikova 12/6/2011 S. Raskhodnikova;

Applied Algorithm Design Lecture 5

Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design

Approximation Algorithms

Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

2.3 Scheduling jobs on identical parallel machines

2.3 Scheduling jobs on identical parallel machines There are jobs to be processed, and there are identical machines (running in parallel) to which each job may be assigned Each job = 1,,, must be processed

Guessing Game: NP-Complete?

Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES 2. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple

Approximation Algorithms. Scheduling. Approximation algorithms. Scheduling jobs on a single machine

Approximation algorithms Approximation Algorithms Fast. Cheap. Reliable. Choose two. NP-hard problems: choose 2 of optimal polynomial time all instances Approximation algorithms. Trade-off between time

Approximation Algorithms: LP Relaxation, Rounding, and Randomized Rounding Techniques. My T. Thai

Approximation Algorithms: LP Relaxation, Rounding, and Randomized Rounding Techniques My T. Thai 1 Overview An overview of LP relaxation and rounding method is as follows: 1. Formulate an optimization

Minimum Makespan Scheduling

Minimum Makespan Scheduling Minimum makespan scheduling: Definition and variants Factor 2 algorithm for identical machines PTAS for identical machines Factor 2 algorithm for unrelated machines Martin Zachariasen,

Answers to some of the exercises.

Answers to some of the exercises. Chapter 2. Ex.2.1 (a) There are several ways to do this. Here is one possibility. The idea is to apply the k-center algorithm first to D and then for each center in D

5 INTEGER LINEAR PROGRAMMING (ILP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1

5 INTEGER LINEAR PROGRAMMING (ILP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1 General Integer Linear Program: (ILP) min c T x Ax b x 0 integer Assumption: A, b integer The integrality condition

Scheduling Shop Scheduling. Tim Nieberg

Scheduling Shop Scheduling Tim Nieberg Shop models: General Introduction Remark: Consider non preemptive problems with regular objectives Notation Shop Problems: m machines, n jobs 1,..., n operations

Lecture 7: Approximation via Randomized Rounding

Lecture 7: Approximation via Randomized Rounding Often LPs return a fractional solution where the solution x, which is supposed to be in {0, } n, is in [0, ] n instead. There is a generic way of obtaining

Lecture 3: Linear Programming Relaxations and Rounding

Lecture 3: Linear Programming Relaxations and Rounding 1 Approximation Algorithms and Linear Relaxations For the time being, suppose we have a minimization problem. Many times, the problem at hand can

Fairness in Routing and Load Balancing

Fairness in Routing and Load Balancing Jon Kleinberg Yuval Rabani Éva Tardos Abstract We consider the issue of network routing subject to explicit fairness conditions. The optimization of fairness criteria

5.1 Bipartite Matching

CS787: Advanced Algorithms Lecture 5: Applications of Network Flow In the last lecture, we looked at the problem of finding the maximum flow in a graph, and how it can be efficiently solved using the Ford-Fulkerson

CMPSCI611: Approximating MAX-CUT Lecture 20

CMPSCI611: Approximating MAX-CUT Lecture 20 For the next two lectures we ll be seeing examples of approximation algorithms for interesting NP-hard problems. Today we consider MAX-CUT, which we proved to

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming Lecturer: Sanjeev Arora

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming Lecturer: Sanjeev Arora Scribe: One of the running themes in this course is the notion of

The multi-integer set cover and the facility terminal cover problem

The multi-integer set cover and the facility teral cover problem Dorit S. Hochbaum Asaf Levin December 5, 2007 Abstract The facility teral cover problem is a generalization of the vertex cover problem.

8.1 Makespan Scheduling

600.469 / 600.669 Approximation Algorithms Lecturer: Michael Dinitz Topic: Dynamic Programing: Min-Makespan and Bin Packing Date: 2/19/15 Scribe: Gabriel Kaptchuk 8.1 Makespan Scheduling Consider an instance

Scheduling Parallel Machine Scheduling. Tim Nieberg

Scheduling Parallel Machine Scheduling Tim Nieberg Problem P C max : m machines n jobs with processing times p 1,..., p n Problem P C max : m machines n jobs with processing times p 1,..., p { n 1 if job

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.

Topic: Greedy Approximations: Set Cover and Min Makespan Date: 1/30/06

CS880: Approximations Algorithms Scribe: Matt Elder Lecturer: Shuchi Chawla Topic: Greedy Approximations: Set Cover and Min Makespan Date: 1/30/06 3.1 Set Cover The Set Cover problem is: Given a set of

Discrete (and Continuous) Optimization Solutions of Exercises 1 WI4 131

Discrete (and Continuous) Optimization Solutions of Exercises 1 WI4 131 Kees Roos Technische Universiteit Delft Faculteit Informatietechnologie en Systemen Afdeling Informatie, Systemen en Algoritmiek

Introduction to Algorithms Review information for Prelim 1 CS 4820, Spring 2010 Distributed Wednesday, February 24

Introduction to Algorithms Review information for Prelim 1 CS 4820, Spring 2010 Distributed Wednesday, February 24 The final exam will cover seven topics. 1. greedy algorithms 2. divide-and-conquer algorithms

Scheduling Parallel Jobs with Monotone Speedup 1

Scheduling Parallel Jobs with Monotone Speedup 1 Alexander Grigoriev, Marc Uetz Maastricht University, Quantitative Economics, P.O.Box 616, 6200 MD Maastricht, The Netherlands, {a.grigoriev@ke.unimaas.nl,

CSC2420 Spring 2015: Lecture 3

CSC2420 Spring 2015: Lecture 3 Allan Borodin January 22, 2015 1 / 1 Announcements and todays agenda Assignment 1 due next Thursday. I may add one or two additional questions today or tomorrow. Todays agenda

Good luck, veel succes!

Final exam Advanced Linear Programming, May 7, 13.00-16.00 Switch off your mobile phone, PDA and any other mobile device and put it far away. No books or other reading materials are allowed. This exam

Approximated Distributed Minimum Vertex Cover Algorithms for Bounded Degree Graphs

Approximated Distributed Minimum Vertex Cover Algorithms for Bounded Degree Graphs Yong Zhang 1.2, Francis Y.L. Chin 2, and Hing-Fung Ting 2 1 College of Mathematics and Computer Science, Hebei University,

Partitioning edge-coloured complete graphs into monochromatic cycles and paths

arxiv:1205.5492v1 [math.co] 24 May 2012 Partitioning edge-coloured complete graphs into monochromatic cycles and paths Alexey Pokrovskiy Departement of Mathematics, London School of Economics and Political

Using the Simplex Method in Mixed Integer Linear Programming

Integer Using the Simplex Method in Mixed Integer UTFSM Nancy, 17 december 2015 Using the Simplex Method in Mixed Integer Outline Mathematical Programming Integer 1 Mathematical Programming Optimisation

Scheduling Parallel Jobs with Linear Speedup

Scheduling Parallel Jobs with Linear Speedup Alexander Grigoriev and Marc Uetz Maastricht University, Quantitative Economics, P.O.Box 616, 6200 MD Maastricht, The Netherlands. Email: {a.grigoriev,m.uetz}@ke.unimaas.nl

6. Mixed Integer Linear Programming

6. Mixed Integer Linear Programming Javier Larrosa Albert Oliveras Enric Rodríguez-Carbonell Problem Solving and Constraint Programming (RPAR) Session 6 p.1/40 Mixed Integer Linear Programming A mixed

Duplicating and its Applications in Batch Scheduling

Duplicating and its Applications in Batch Scheduling Yuzhong Zhang 1 Chunsong Bai 1 Shouyang Wang 2 1 College of Operations Research and Management Sciences Qufu Normal University, Shandong 276826, China

Constant Factor Approximation Algorithm for the Knapsack Median Problem

Constant Factor Approximation Algorithm for the Knapsack Median Problem Amit Kumar Abstract We give a constant factor approximation algorithm for the following generalization of the k-median problem. We

Ecient approximation algorithm for minimizing makespan. on uniformly related machines. Chandra Chekuri. November 25, 1997.

Ecient approximation algorithm for minimizing makespan on uniformly related machines Chandra Chekuri November 25, 1997 Abstract We obtain a new ecient approximation algorithm for scheduling precedence

Scheduling and (Integer) Linear Programming

Scheduling and (Integer) Linear Programming Christian Artigues LAAS - CNRS & Université de Toulouse, France artigues@laas.fr Master Class CPAIOR 2012 - Nantes Christian Artigues Scheduling and (Integer)

LINEAR PROGRAMMING PROBLEM: A GEOMETRIC APPROACH

59 LINEAR PRGRAMMING PRBLEM: A GEMETRIC APPRACH 59.1 INTRDUCTIN Let us consider a simple problem in two variables x and y. Find x and y which satisfy the following equations x + y = 4 3x + 4y = 14 Solving

Example 5.5. The heuristics are demonstrated with this example of n=5 vertices and distances in the following matrix:

39 5.3. HEURISTICS FOR THE TSP Notation: f = length of the tour given by the algortihm a f = length of the optimal tour min Most of the following heuristics are designed to solve symmetric instances of

Planar Tree Transformation: Results and Counterexample

Planar Tree Transformation: Results and Counterexample Selim G Akl, Kamrul Islam, and Henk Meijer School of Computing, Queen s University Kingston, Ontario, Canada K7L 3N6 Abstract We consider the problem

THE PROBLEM WORMS (1) WORMS (2) THE PROBLEM OF WORM PROPAGATION/PREVENTION THE MINIMUM VERTEX COVER PROBLEM

1 THE PROBLEM OF WORM PROPAGATION/PREVENTION I.E. THE MINIMUM VERTEX COVER PROBLEM Prof. Tiziana Calamoneri Network Algorithms A.y. 2014/15 2 THE PROBLEM WORMS (1)! A computer worm is a standalone malware

Single machine parallel batch scheduling with unbounded capacity

Workshop on Combinatorics and Graph Theory 21th, April, 2006 Nankai University Single machine parallel batch scheduling with unbounded capacity Yuan Jinjiang Department of mathematics, Zhengzhou University

Week 5 Integral Polyhedra

Week 5 Integral Polyhedra We have seen some examples 1 of linear programming formulation that are integral, meaning that every basic feasible solution is an integral vector. This week we develop a theory

Near Optimal Solutions

Near Optimal Solutions Many important optimization problems are lacking efficient solutions. NP-Complete problems unlikely to have polynomial time solutions. Good heuristics important for such problems.

max cx s.t. Ax c where the matrix A, cost vector c and right hand side b are given and x is a vector of variables. For this example we have x

Linear Programming Linear programming refers to problems stated as maximization or minimization of a linear function subject to constraints that are linear equalities and inequalities. Although the study

Inverse Optimization by James Orlin

Inverse Optimization by James Orlin based on research that is joint with Ravi Ahuja Jeopardy 000 -- the Math Programming Edition The category is linear objective functions The answer: When you maximize

A constant-factor approximation algorithm for the k-median problem

A constant-factor approximation algorithm for the k-median problem Moses Charikar Sudipto Guha Éva Tardos David B. Shmoys July 23, 2002 Abstract We present the first constant-factor approximation algorithm

Computer Algorithms. NP-Complete Problems. CISC 4080 Yanjun Li

Computer Algorithms NP-Complete Problems NP-completeness The quest for efficient algorithms is about finding clever ways to bypass the process of exhaustive search, using clues from the input in order

Chapter 4. Trees. 4.1 Basics

Chapter 4 Trees 4.1 Basics A tree is a connected graph with no cycles. A forest is a collection of trees. A vertex of degree one, particularly in a tree, is called a leaf. Trees arise in a variety of applications.

2.3 Convex Constrained Optimization Problems

42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions

Ph.D. Thesis. Judit Nagy-György. Supervisor: Péter Hajnal Associate Professor

Online algorithms for combinatorial problems Ph.D. Thesis by Judit Nagy-György Supervisor: Péter Hajnal Associate Professor Doctoral School in Mathematics and Computer Science University of Szeged Bolyai

Definition 11.1. Given a graph G on n vertices, we define the following quantities:

Lecture 11 The Lovász ϑ Function 11.1 Perfect graphs We begin with some background on perfect graphs. graphs. First, we define some quantities on Definition 11.1. Given a graph G on n vertices, we define

Solving Mixed Integer Linear Programs Using Branch and Cut Algorithm

1 Solving Mixed Integer Linear Programs Using Branch and Cut Algorithm by Shon Albert A Project Submitted to the Graduate Faculty of North Carolina State University in Partial Fulfillment of the Requirements

TU e. Advanced Algorithms: experimentation project. The problem: load balancing with bounded look-ahead. Input: integer m 2: number of machines

The problem: load balancing with bounded look-ahead Input: integer m 2: number of machines integer k 0: the look-ahead numbers t 1,..., t n : the job sizes Problem: assign jobs to machines machine to which

Why? A central concept in Computer Science. Algorithms are ubiquitous.

Analysis of Algorithms: A Brief Introduction Why? A central concept in Computer Science. Algorithms are ubiquitous. Using the Internet (sending email, transferring files, use of search engines, online

Optimal Binary Space Partitions in the Plane

Optimal Binary Space Partitions in the Plane Mark de Berg and Amirali Khosravi TU Eindhoven, P.O. Box 513, 5600 MB Eindhoven, the Netherlands Abstract. An optimal bsp for a set S of disjoint line segments

Permutation Betting Markets: Singleton Betting with Extra Information

Permutation Betting Markets: Singleton Betting with Extra Information Mohammad Ghodsi Sharif University of Technology ghodsi@sharif.edu Hamid Mahini Sharif University of Technology mahini@ce.sharif.edu

GRAPH THEORY and APPLICATIONS. Trees

GRAPH THEORY and APPLICATIONS Trees Properties Tree: a connected graph with no cycle (acyclic) Forest: a graph with no cycle Paths are trees. Star: A tree consisting of one vertex adjacent to all the others.

Triangle deletion. Ernie Croot. February 3, 2010

Triangle deletion Ernie Croot February 3, 2010 1 Introduction The purpose of this note is to give an intuitive outline of the triangle deletion theorem of Ruzsa and Szemerédi, which says that if G = (V,

4 Basics of Trees. Petr Hliněný, FI MU Brno 1 FI: MA010: Trees and Forests

4 Basics of Trees Trees, actually acyclic connected simple graphs, are among the simplest graph classes. Despite their simplicity, they still have rich structure and many useful application, such as in

COLORED GRAPHS AND THEIR PROPERTIES

COLORED GRAPHS AND THEIR PROPERTIES BEN STEVENS 1. Introduction This paper is concerned with the upper bound on the chromatic number for graphs of maximum vertex degree under three different sets of coloring

Linear Programming: Introduction

Linear Programming: Introduction Frédéric Giroire F. Giroire LP - Introduction 1/28 Course Schedule Session 1: Introduction to optimization. Modelling and Solving simple problems. Modelling combinatorial

A counter-example to the Hirsch conjecture

A counter-example to the Hirsch conjecture Francisco Santos Universidad de Cantabria, Spain http://personales.unican.es/santosf/hirsch Bernoulli Conference on Discrete and Computational Geometry Lausanne,

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma Please Note: The references at the end are given for extra reading if you are interested in exploring these ideas further. You are

Online Adwords Allocation Shoshana Neuburger May 6, 2009 1 Overview Many search engines auction the advertising space alongside search results. When Google interviewed Amin Saberi in 2004, their advertisement

RANDOMIZATION IN APPROXIMATION AND ONLINE ALGORITHMS. Manos Thanos Randomized Algorithms NTUA

RANDOMIZATION IN APPROXIMATION AND ONLINE ALGORITHMS 1 Manos Thanos Randomized Algorithms NTUA RANDOMIZED ROUNDING Linear Programming problems are problems for the optimization of a linear objective function,

Collinear Points in Permutations

Collinear Points in Permutations Joshua N. Cooper Courant Institute of Mathematics New York University, New York, NY József Solymosi Department of Mathematics University of British Columbia, Vancouver,

Two General Methods to Reduce Delay and Change of Enumeration Algorithms

ISSN 1346-5597 NII Technical Report Two General Methods to Reduce Delay and Change of Enumeration Algorithms Takeaki Uno NII-2003-004E Apr.2003 Two General Methods to Reduce Delay and Change of Enumeration

8.1 Min Degree Spanning Tree

CS880: Approximations Algorithms Scribe: Siddharth Barman Lecturer: Shuchi Chawla Topic: Min Degree Spanning Tree Date: 02/15/07 In this lecture we give a local search based algorithm for the Min Degree

CHAPTER 9. Integer Programming

CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral

Resource Allocation with Time Intervals

Resource Allocation with Time Intervals Andreas Darmann Ulrich Pferschy Joachim Schauer Abstract We study a resource allocation problem where jobs have the following characteristics: Each job consumes

GENERATING LOW-DEGREE 2-SPANNERS

SIAM J. COMPUT. c 1998 Society for Industrial and Applied Mathematics Vol. 27, No. 5, pp. 1438 1456, October 1998 013 GENERATING LOW-DEGREE 2-SPANNERS GUY KORTSARZ AND DAVID PELEG Abstract. A k-spanner

CSC 373: Algorithm Design and Analysis Lecture 16

CSC 373: Algorithm Design and Analysis Lecture 16 Allan Borodin February 25, 2013 Some materials are from Stephen Cook s IIT talk and Keven Wayne s slides. 1 / 17 Announcements and Outline Announcements

Scheduling to Minimize Power Consumption using Submodular Functions

Scheduling to Minimize Power Consumption using Submodular Functions Erik D. Demaine MIT edemaine@mit.edu Morteza Zadimoghaddam MIT morteza@mit.edu ABSTRACT We develop logarithmic approximation algorithms

Introduction to Scheduling Theory

Introduction to Scheduling Theory Arnaud Legrand Laboratoire Informatique et Distribution IMAG CNRS, France arnaud.legrand@imag.fr November 8, 2004 1/ 26 Outline 1 Task graphs from outer space 2 Scheduling

Solutions to Homework 6

Solutions to Homework 6 Debasish Das EECS Department, Northwestern University ddas@northwestern.edu 1 Problem 5.24 We want to find light spanning trees with certain special properties. Given is one example

JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS. Received December May 12, 2003; revised February 5, 2004

Scientiae Mathematicae Japonicae Online, Vol. 10, (2004), 431 437 431 JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS Ondřej Čepeka and Shao Chin Sung b Received December May 12, 2003; revised February

A simpler and better derandomization of an approximation algorithm for Single Source Rent-or-Buy

A simpler and better derandomization of an approximation algorithm for Single Source Rent-or-Buy David P. Williamson Anke van Zuylen School of Operations Research and Industrial Engineering, Cornell University,

The Graphical Simplex Method: An Example

The Graphical Simplex Method: An Example Consider the following linear program: Max 4x 1 +3x Subject to: x 1 +3x 6 (1) 3x 1 +x 3 () x 5 (3) x 1 +x 4 (4) x 1, x 0. Goal: produce a pair of x 1 and x that

Jianlin Cheng, PhD Computer Science Department University of Missouri, Columbia Fall, 2013

Jianlin Cheng, PhD Computer Science Department University of Missouri, Columbia Fall, 2013 Princeton s class notes on linear programming MIT s class notes on linear programming Xian Jiaotong University

4.4 The recursion-tree method

4.4 The recursion-tree method Let us see how a recursion tree would provide a good guess for the recurrence = 3 4 ) Start by nding an upper bound Floors and ceilings usually do not matter when solving

WORST-CASE PERFORMANCE ANALYSIS OF SOME APPROXIMATION ALGORITHMS FOR MINIMIZING MAKESPAN AND FLOWTIME

WORST-CASE PERFORMANCE ANALYSIS OF SOME APPROXIMATION ALGORITHMS FOR MINIMIZING MAKESPAN AND FLOWTIME PERUVEMBA SUNDARAM RAVI, LEVENT TUNÇEL, MICHAEL HUANG Abstract. In 1976, Coffman and Sethi conjectured

CMSC 451: Graph Properties, DFS, BFS, etc.

CMSC 451: Graph Properties, DFS, BFS, etc. Slides By: Carl Kingsford Department of Computer Science University of Maryland, College Park Based on Chapter 3 of Algorithm Design by Kleinberg & Tardos. Graphs

Improved Algorithms for Data Migration

Improved Algorithms for Data Migration Samir Khuller 1, Yoo-Ah Kim, and Azarakhsh Malekian 1 Department of Computer Science, University of Maryland, College Park, MD 20742. Research supported by NSF Award

Linear Programming, Lagrange Multipliers, and Duality Geoff Gordon

lp.nb 1 Linear Programming, Lagrange Multipliers, and Duality Geoff Gordon lp.nb 2 Overview This is a tutorial about some interesting math and geometry connected with constrained optimization. It is not

MapReduce and Distributed Data Analysis. Sergei Vassilvitskii Google Research

MapReduce and Distributed Data Analysis Google Research 1 Dealing With Massive Data 2 2 Dealing With Massive Data Polynomial Memory Sublinear RAM Sketches External Memory Property Testing 3 3 Dealing With

Minimizing the Number of Machines in a Unit-Time Scheduling Problem

Minimizing the Number of Machines in a Unit-Time Scheduling Problem Svetlana A. Kravchenko 1 United Institute of Informatics Problems, Surganova St. 6, 220012 Minsk, Belarus kravch@newman.bas-net.by Frank

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S

THE SCHEDULING OF MAINTENANCE SERVICE

THE SCHEDULING OF MAINTENANCE SERVICE Shoshana Anily Celia A. Glass Refael Hassin Abstract We study a discrete problem of scheduling activities of several types under the constraint that at most a single

Nan Kong, Andrew J. Schaefer. Department of Industrial Engineering, Univeristy of Pittsburgh, PA 15261, USA

A Factor 1 2 Approximation Algorithm for Two-Stage Stochastic Matching Problems Nan Kong, Andrew J. Schaefer Department of Industrial Engineering, Univeristy of Pittsburgh, PA 15261, USA Abstract We introduce

Load Balancing Backtracking, branch & bound and alpha-beta pruning: how to assign work to idle processes without much communication? Additionally for alpha-beta pruning: implementing the young-brothers-wait

Introduction and message of the book

1 Introduction and message of the book 1.1 Why polynomial optimization? Consider the global optimization problem: P : for some feasible set f := inf x { f(x) : x K } (1.1) K := { x R n : g j (x) 0, j =

1 Approximating Set Cover

CS 05: Algorithms (Grad) Feb 2-24, 2005 Approximating Set Cover. Definition An Instance (X, F ) of the set-covering problem consists of a finite set X and a family F of subset of X, such that every elemennt

CSE 101: Design and Analysis of Algorithms Homework 4 Winter 2016 Due: February 25, 2016

CSE 101: Design and Analysis of Algorithms Homework 4 Winter 2016 Due: February 25, 2016 Exercises (strongly recommended, do not turn in) 1. (DPV textbook, Problem 7.11.) Find the optimal solution to the

Complexity Theory. IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar

Complexity Theory IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar Outline Goals Computation of Problems Concepts and Definitions Complexity Classes and Problems Polynomial Time Reductions Examples

Analysis of Approximation Algorithms for k-set Cover using Factor-Revealing Linear Programs

Analysis of Approximation Algorithms for k-set Cover using Factor-Revealing Linear Programs Stavros Athanassopoulos, Ioannis Caragiannis, and Christos Kaklamanis Research Academic Computer Technology Institute

Linear Programming. March 14, 2014

Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1