Lecture 7: Approximation via Randomized Rounding


 Bonnie Maxwell
 1 years ago
 Views:
Transcription
1 Lecture 7: Approximation via Randomized Rounding Often LPs return a fractional solution where the solution x, which is supposed to be in {0, } n, is in [0, ] n instead. There is a generic way of obtaining an {0, } n vector from a x [0, ] n one: let the ith variable be with probability x i, 0 with probability ( x i ). The resulting random vector x int has the desirable property that for any i, E[x int i] = x i. In particular, if x minimized a certain linear function c x, then by linearity of expectation, the expected cost of the random {0, } n vector is also equal to c x. Of course, often the resulting x int won t satisfy the constraints, and therein lies the nontriviality. We will need some facts in probability theory, mostly in the concentration of random variables around their means. I will collect these facts down in the appendix; the proofs of which can be found in texts on randomized algorithms or the web. The Set Cover Problem. Recall the LP relaxation for the set cover problem. min subject to m c(s j )x j () j= x j i U (2) x j [0, ] j =...m (3) Let x be a solution to the LP. Consider the experiment where we randomly select set S j with probability x j, independently of each other. Let X j be the indicator random variable which is if we picked S j and 0 otherwise. The random variable m j= c(s j)x j indicates the cost of the sets picked, and thus the expected cost of our solution is precisely m j= c(s j)e[x j ] which is the LP solution. Are all elements covered? Fix an element i. The following claim shows that each element is covered with at least a constant probability. Claim. The probability that an element i is covered is at least ( /e). Proof. Element i is covered iff a set containing i is picked. The probability that i is not covered, thus, is at most ( x j ) due to independence. But this is at most where the inequality follows from (2). e xj = e P x j e
2 So our randomized algorithm returns a collection of sets whose expected cost is the LP cost, and covers each element with probability at least ( /e). Suppose we repeated this 2 ln n times and returned the union of the sets picked in each round. The expected cost of our solution is going to be 2 ln n times the LP solution. What s the probability that an element i is not covered in any of the rounds? Since the rounds are independent, this probability is at most ( e )2 ln n = /n 2. By the union bound (Fact 2), the probability that there exists some element i which is not covered is at most n = /n. Thus, we get the following theorem. n 2 Theorem. The algorithm described above returns a set cover with probability at least ( n ) whose expected cost is 2 ln n times the LP solution. Randomized approximation algorithms often look like the theorem above. They succeed with high probability and return a solution whose expected cost is at most (or at least) a factor of the optimum solution. By taking a slight hit at the approximation factor, the guarantee on the expected cost can often be converted to a guarantee on the cost with high probability. For instance, Markov s inequality (Fact 3) tells us that the probability that the cost of the sets picked exceeds 4 ln n times the optimum is at most /2. Thus, the algorithm returns a set cover of cost 4 ln n times the optimum with probability at least 2 n /3. If we repeat this algorithm t times, then with probability 3 (really, really high probability if t is some polynomial in n) we will encounter such a set cover. t Suppose we look at the set multicover problem where each element i needs to be covered at least b i times for some positive integer b i? The above analysis needs to be modified a bit; in particular, the probability that an element i is not covered in any iteration goes down from /e to /e b i, although, it now needs to be covered b i times. This is true, and I ll leave this calculation checking as an exercise. I now show a slightly different (but similar in spirit) randomized algorithm. It runs in a single round by sampling sets with higher probabilities. Let ˆx j = min(, 6 ln n x j ). Sample each set S j independently with probability ˆx j. That s it. The expected cost now is j c(s j)ˆx j 6 ln n lp. To argue that we cover each element sufficient number of times, we use the Chernoff bound (Fact 4). Let X j be the indicator random variable of the event whether S j is picked or not. For an element i, define Z i = X j. If Z i b i, then i is covered, otherwise not. The following claim, using a union bound, gives that with probability ( /n), the sets picked form a set cover. Claim 2. Pr[Z i < b i ] n 2. Proof. By reordering suppose i S j for j k. We may assume ˆx j < for all such j; since all others (say l of them) are picked with probability, we can move to the residual problem where the element i needs to be covered b i l times. Now, E[Z i ] = k j= ˆx j = 6 ln n b i. Thus, Fact 4 gives us (with δ = 6 ln n 5/6) Pr[Z i < b i ] exp( 6 ln n ) n 2 Facility Location Problem Let s recall the LP relaxation for the facility location problem and it s dual. Let (x, y) and (α, β) be a pair of optimal primaldual solutions. Recall the following claim from last class 2
3 min f i y i + c(i, j)x ij (4) i F i F,j C x ij =, j C (5) i F y i x ij, i F, j C (6) x, y 0 max α j (7) j C β ij f i, i F (8) j C α j β ij c(i, j), i F, j C (9) β 0 Claim 3. If x ij > 0, then c(i, j) α j. Before we describe the algorithm, we are going to make a structural assumption about (x, y). I will leave it as an exercise to prove that the assumption is without loss of generality. Assumption: If x ij > 0, then y i = x ij. Algorithm. (In class, we looked at the clustering algorithm which ordered clients in increasing order of α j, and gave us a (2+2/e) approximation. Below we give a refined argument which attains a ( + 2/e)approximation.) The algorithm proceeds as follows. Let Cj := i c ijx ij. Order the clients in increasing order of (α j + Cj ). Let U denote the set of unassigned clients initialized to C. Let j be the first client in U in the order. Form the cluster N 2 (j) consisting of j, all facilities {i : x ij > 0}, and all clients {j U : x ij > 0 for some i N 2 (j)}. j is called the cluster center of N 2 (j). Remove all clients of N 2 (j) from U and repeat till U is empty. At the end of this step all clients are assigned some cluster N 2 (j), for some client j. Note that in any cluster N 2 (j), i N 2 (j) y i = i N 2 (j) x ij =. In each such cluster, open exactly one facility with probability y i  this is valid since i y i =. Let X be the set of facilities opened. The clients are assigned their nearest facility. Claim 4. The expected facility opening cost is precisely i F f iy i. Proof. The probability a facility i is in X is precisely y i. Let us now calculate the expected connection cost for a client j. Note that a cluster center j pays expected connection cost i N 2 (j) c(i, j)x ij = Cj. Let s take a noncenter client j N 2(j ). Let Γ(j) := {i : x ij > 0}, and let {j,..., j r } be the centers such that N 2 (j l ) Γ(j) is nonempty. For l =,..., r, let p l := i N 2 (j l ) x ij be the probability a facility in N 2 (j l ) Γ(j) is opened. Note that l p l =, and that these r events are independent. Let Cj (l) be the expected cost of connecting to a client in N 2(j l ) Γ(j) conditioned on the event that such a facility is open. Thus, C j (l) := i N 2 (j l ) c(i, j) xij p l Let s order j,..., j l so that Cj () C j (l). Lastly, let s define C j ( ) to be the expected cost of connecting j conditioned on the event that no client in Γ(j) is opened. We ll bound this shortly. Now we are ready to bound the expected connection cost of client j. 3
4 Claim 5. The expected connection cost of client j is at most p C j () + ( p )p 2 C j (2) + + ( p )( p 2 ) ( p r )p r C j (r) + C j ( ) r ( p l ) Proof. We are conditioning on mutually exclusive events: event l is when no facility in N 2 (j l ) Γ(j) has been opened for l < l, and a facility in N 2 (j l ) Γ(j) has been opened; and the last event when no facility in Γ(j) has been opened. Claim 6. C j ( ) C j + 2α j. Proof. If no facility in Γ(j) is opened, then connect j to the facility i opened in N 2 (j ): recall j N 2 (j ) and thus there exists i N 2 (j ) with x ij > 0. The connection cost to i is c(i, j) c(i, j) + c(i, j ) + c(i, j ) 2α j + c(i, j ), where we have used Claim 3 and the fact that α j α j. Taking expectations, we get Cj ( ) 2α j + c(i, j )x i j. i N 2 (j) l= Putting it all together, we get the total connection cost of j is at most r [ pl ( p l ) ( p )Cj (l) ] + (Cj + 2α j ) l= r ( p l ) (0) Since r l= p l =, we have the product r l= ( p l) /e. As a first approximation, note that ( p l ) and r l= p lcj (l) = C j. This gives us total expected facility + connection cost of at most f i y i + Cj + (Cj + 2α j ) opt( + 3/e) e i F j C giving a approximation. One can analyze better. Now we state the following inequality: Lemma. Let A A 2 A r, and let r p i =. Then, ( r ) ( p A + p 2 ( p )A p r ( p )( p 2 ) ( p r )A r p i A i j C l= ) r ( p i ) Proof. Define q i := p i( p )...( p i ) Q r ( p. We need to prove that r i) q ia i r p ia i. Note that r q i = = r p i. Furthermore, there exists a < k < r such that q i p i for all i k and q i p i for all i > k; it s the largest i for which i l= ( p l) exceeds r ( p i). Let δ i = q i p i. Thus, δ i 0 for i k, δ i 0 for i > k and i δ i = 0. ( r r k ) ( r k ) (q i p i )A i = δ i A i A k δ i + A k+ δ i = δ i (A k A k+ ) 0 i=(k+) 4
5 Given the above fact, we can get a stronger approximation factor. Let q := r l= ( p l). Then, the expected connection cost is at most ( q)c j + qc j ( ) C j + 2α j /e Putting it together gives a ( + 2/e)approximation. The Group Steiner Tree Problem (on a tree) In this problem, the input is a rooted tree T on n vertices. There are k disjoint groups g,..., g k, where each group is a subset of leaves. Each edge of the tree has a cost, and the goal is to find the cheapest rooted subtree which contains a vertex from each group. LP Relaxation. min e E(T ) c e x e x 0 () subject to x(δ(s)) S V : g i S for some i, r / S (2) Algorithm. Let the root be r. Given a nonroot vertex v in the tree, let e v denote the edge connecting v to its parent. We ll abuse notation and let x v denote x ev. We call an edge f an ancestor of edge e if the path from r to the end points of e contains f. f is the parent of e if its an ancestor and shares an end point with e. Preprocessing x. We process the solution x as follows. Firstly, remove all vertices v with x v /2g. For any group g i, we still have v g i x v /2. Secondly, increase x e for every edge to the nearest power of (/2). Note that the new LP cost at most doubles. Thirdly, for an edge e with parent f, if x e = x f, then merge the two edges into a single edge with the same x e value. This doesn t change the feasibility of the solution. Note that the height of the tree now is at most 2 log g: any path from root to a vertex v has strictly decreasing x e in powers of /2 ranging from at most to at least /2g. Finally, decrease all the x ev s such that for any group g i, v g i x v =. Since these are leaf edges and groups are disjoint, this can be done without changing the feasibility of the solution. Independently, sample all edges e incident on the root with probability x e. For every other edge, e with parent f, sample e independently with probability xe x f. The resulting graph can have many components  throw away all except the component containing the root. The above step is repeated (6 ln n log g) times independently and the final solution is the union of all edges picked; g is the maximum size of a group. Theorem 2. The above algorithm has an expected cost of O(log n log g) opt and is a feasible group Steiner tree with high probability. Proof. In each iteration the probability that an edge e is sampled is exactly x e. This is because, let the path from root to e be e p, e p,..., e, e. The probability e survives this iteration is at most x e x e x e x e2 xe p x ep x ep = x e 5
6 By linearity of expectation, the expected cost of the edges in any iteration is opt, and the first claim in the theorem follows. Fix a group g i and a iteration. We claim that the probability that the subtree sampled contains a vertex from g is at least 8 log g group g i is not covered in 6 ln n log g iterations is at most. This will prove the second claim of the theorem since the probability ( ) 6 ln n log g 8 log g. Union bound n 2 over all groups (there are n many at most) ends the argument. For a vertex u g i, let E u be the event that u is contained in the subtree. We are trying to lower bound the event Pr[ u g i E u ]. Let s calculate Pr[E u ]. u is contained in the tree if and only if all the edges on the path from u to r are sampled. This probability is x u. Furthermore, the constraint in the LP gives us u g i x u =. So, we have µ := u g i Pr[E u ] =. (Thus, if all the E u s were independent in a group g i, then in fact the probability that none of the vertices are contained in the tree would have been at most /e.) Let u v if the events E u and E v are not independent. Define := u v Pr[E u E v ]. Claim 7. 2 log g Proof. For u and v in the tree, let w be the least common ancestor. Note that conditioned on the event E w, the event E u and E v are independent. Thus the probability Pr[E u E v ] = Pr[E u E v E w ] Pr[E w ] = Pr[E u E w ] Pr[E v E w ] Pr[E w ] = x ux v Now for any interior node w, let T w be the subtree rooted at w. Note that u g i T w x u. This is because of the feasibility of the LP constraint with S containing all vertices in T w and all vertices in g i \ T w : x(δ(s)) = + x u = + x u u g i \T w u g i T w Let A u denote the set of ancestors of u in the tree. Since the height of the tree is at most 2 log g, A u 2 log g. Thus, = u v x u x v u g i w A u x u x v x u v g i T w u g i w A u x v x u A u 2 log g v g i T w u g i Now by Janson s inequality (Fact 5), we get that Pr[ ) ( E u ] exp ( µ2 exp ) µ + 4 log g 8 log g u g i since e x x + x2 2 x/2 when 0 x. So we get that in any iteration and any group g i, the probability we pick a vertex from the group is at least 8 log g. This completes the proof. 6
7 Some Probabilistic Facts. Fact (Linearity of Expectation). Let X,..., X n be random variables and let X := n X i. Then, E[X] = n E[X i]. Fact 2 (The Union Bound). Given n events E,..., E n, the probability that one of them occurs is at most the sum of their probabilities. Pr[E E 2 E n ] n Pr[E i ] Fact 3 (Markov s Inequality). Let X be a nonnegative random variable. Then Pr[X te[x] ] t Fact 4 (Chernoff Bound). X,, X n be n independent {0, } random variables with Pr[X i = ] = p i. Let X := n X i and let µ = n p i = E[X]. Then ( e δ ) µ Pr[X > ( + δ)µ] ( + δ) (+δ) ( e δ ) µ Pr[X < ( δ)µ] ( δ) ( δ) Useful versions of the above: For any δ > 0, For t > 0, For t > 4µ, Pr[X > ( + δ)µ] exp( µδ 2 /3) Pr[X < ( δ)µ] exp( µδ 2 /2) Pr[ X µ > t] 2 exp( 2t 2 /n) Pr[X > t] 2 t Fact 5 (Janson s Inequality). Let X U obtained by sampling element i U with probability p i independently. A,..., A t be subsets of U and let E i be the event that A i X. Then, if µ = t Pr[E i] and = i,j:a i A j Pr[E i E j ], then Pr[ t ) E i ] exp ( µ2. µ + 7
Lecture 3: Linear Programming Relaxations and Rounding
Lecture 3: Linear Programming Relaxations and Rounding 1 Approximation Algorithms and Linear Relaxations For the time being, suppose we have a minimization problem. Many times, the problem at hand can
More information10.1 Integer Programming and LP relaxation
CS787: Advanced Algorithms Lecture 10: LP Relaxation and Rounding In this lecture we will design approximation algorithms using linear programming. The key insight behind this approach is that the closely
More informationApproximation Algorithms: LP Relaxation, Rounding, and Randomized Rounding Techniques. My T. Thai
Approximation Algorithms: LP Relaxation, Rounding, and Randomized Rounding Techniques My T. Thai 1 Overview An overview of LP relaxation and rounding method is as follows: 1. Formulate an optimization
More informationLecture 6: Approximation via LP Rounding
Lecture 6: Approximation via LP Rounding Let G = (V, E) be an (undirected) graph. A subset C V is called a vertex cover for G if for every edge (v i, v j ) E we have v i C or v j C (or both). In other
More informationLecture Steiner Tree Approximation June 15
Approximation Algorithms Workshop June 1317, 2011, Princeton Lecture Steiner Tree Approximation June 15 Fabrizio Grandoni Scribe: Fabrizio Grandoni 1 Overview In this lecture we will summarize our recent
More informationMarkov Chains, part I
Markov Chains, part I December 8, 2010 1 Introduction A Markov Chain is a sequence of random variables X 0, X 1,, where each X i S, such that P(X i+1 = s i+1 X i = s i, X i 1 = s i 1,, X 0 = s 0 ) = P(X
More information1 Unrelated Parallel Machine Scheduling
IIIS 014 Spring: ATCS  Selected Topics in Optimization Lecture date: Mar 13, 014 Instructor: Jian Li TA: Lingxiao Huang Scribe: Yifei Jin (Polishing not finished yet.) 1 Unrelated Parallel Machine Scheduling
More informationAnswers to some of the exercises.
Answers to some of the exercises. Chapter 2. Ex.2.1 (a) There are several ways to do this. Here is one possibility. The idea is to apply the kcenter algorithm first to D and then for each center in D
More informationprinceton univ. F 13 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming Lecturer: Sanjeev Arora
princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming Lecturer: Sanjeev Arora Scribe: One of the running themes in this course is the notion of
More informationThe multiinteger set cover and the facility terminal cover problem
The multiinteger set cover and the facility teral cover problem Dorit S. Hochbaum Asaf Levin December 5, 2007 Abstract The facility teral cover problem is a generalization of the vertex cover problem.
More informationApproximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NPCompleteness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
More information5.1 Bipartite Matching
CS787: Advanced Algorithms Lecture 5: Applications of Network Flow In the last lecture, we looked at the problem of finding the maximum flow in a graph, and how it can be efficiently solved using the FordFulkerson
More informationCSL851: Algorithmic Graph Theory Semester I Lecture 1: July 24
CSL851: Algorithmic Graph Theory Semester I 20132014 Lecture 1: July 24 Lecturer: Naveen Garg Scribes: Suyash Roongta Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have
More informationTHEORY OF SIMPLEX METHOD
Chapter THEORY OF SIMPLEX METHOD Mathematical Programming Problems A mathematical programming problem is an optimization problem of finding the values of the unknown variables x, x,, x n that maximize
More information2.3 Scheduling jobs on identical parallel machines
2.3 Scheduling jobs on identical parallel machines There are jobs to be processed, and there are identical machines (running in parallel) to which each job may be assigned Each job = 1,,, must be processed
More informationCS5314 Randomized Algorithms. Lecture 16: Balls, Bins, Random Graphs (Random Graphs, Hamiltonian Cycles)
CS5314 Randomized Algorithms Lecture 16: Balls, Bins, Random Graphs (Random Graphs, Hamiltonian Cycles) 1 Objectives Introduce Random Graph Model used to define a probability space for all graphs with
More informationWeek 5 Integral Polyhedra
Week 5 Integral Polyhedra We have seen some examples 1 of linear programming formulation that are integral, meaning that every basic feasible solution is an integral vector. This week we develop a theory
More informationSteiner Tree Approximation via IRR. Randomized Rounding
Steiner Tree Approximation via Iterative Randomized Rounding Graduate Program in Logic, Algorithms and Computation μπλ Network Algorithms and Complexity June 18, 2013 Overview 1 Introduction Scope Related
More informationCS268: Geometric Algorithms Handout #5 Design and Analysis Original Handout #15 Stanford University Tuesday, 25 February 1992
CS268: Geometric Algorithms Handout #5 Design and Analysis Original Handout #15 Stanford University Tuesday, 25 February 1992 Original Lecture #6: 28 January 1991 Topics: Triangulating Simple Polygons
More informationApplied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
More informationCOS 511: Foundations of Machine Learning. Rob Schapire Lecture #14 Scribe: Qian Xi March 30, 2006
COS 511: Foundations of Machine Learning Rob Schapire Lecture #14 Scribe: Qian Xi March 30, 006 In the previous lecture, we introduced a new learning model, the Online Learning Model. After seeing a concrete
More information! Solve problem to optimality. ! Solve problem in polytime. ! Solve arbitrary instances of the problem. #approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NPhard problem What should I do? A Theory says you're unlikely to find a polytime algorithm Must sacrifice one of three
More informationChapter 4. Trees. 4.1 Basics
Chapter 4 Trees 4.1 Basics A tree is a connected graph with no cycles. A forest is a collection of trees. A vertex of degree one, particularly in a tree, is called a leaf. Trees arise in a variety of applications.
More informationCOLORED GRAPHS AND THEIR PROPERTIES
COLORED GRAPHS AND THEIR PROPERTIES BEN STEVENS 1. Introduction This paper is concerned with the upper bound on the chromatic number for graphs of maximum vertex degree under three different sets of coloring
More information1 Polyhedra and Linear Programming
CS 598CSC: Combinatorial Optimization Lecture date: January 21, 2009 Instructor: Chandra Chekuri Scribe: Sungjin Im 1 Polyhedra and Linear Programming In this lecture, we will cover some basic material
More informationLecture Notes 6: Approximations for MAXSAT. 2 Notes about Weighted Vertex Cover Approximation
Algorithmic Methods //00 Lecture Notes 6: Approximations for MAXSAT Professor: Yossi Azar Scribe:Alon Ardenboim Introduction Although solving SAT is nown to be NPComplete, in this lecture we will cover
More information/ Approximation Algorithms Lecturer: Michael Dinitz Topic: Steiner Tree and TSP Date: 01/29/15 Scribe: Katie Henry
600.469 / 600.669 Approximation Algorithms Lecturer: Michael Dinitz Topic: Steiner Tree and TSP Date: 01/29/15 Scribe: Katie Henry 2.1 Steiner Tree Definition 2.1.1 In the Steiner Tree problem the input
More informationSection Notes 4. Duality, Sensitivity, Dual Simplex, Complementary Slackness. Applied Math 121. Week of February 28, 2011
Section Notes 4 Duality, Sensitivity, Dual Simplex, Complementary Slackness Applied Math 121 Week of February 28, 2011 Goals for the week understand the relationship between primal and dual programs. know
More information! Solve problem to optimality. ! Solve problem in polytime. ! Solve arbitrary instances of the problem. !approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NPhard problem What should I do? A Theory says you're unlikely to find a polytime algorithm Must sacrifice one of
More informationmax cx s.t. Ax c where the matrix A, cost vector c and right hand side b are given and x is a vector of variables. For this example we have x
Linear Programming Linear programming refers to problems stated as maximization or minimization of a linear function subject to constraints that are linear equalities and inequalities. Although the study
More informationLecture 1: Course overview, circuits, and formulas
Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik
More informationAn Approximation Algorithm for Bounded Degree Deletion
An Approximation Algorithm for Bounded Degree Deletion Tomáš Ebenlendr Petr Kolman Jiří Sgall Abstract Bounded Degree Deletion is the following generalization of Vertex Cover. Given an undirected graph
More informationGuessing Game: NPComplete?
Guessing Game: NPComplete? 1. LONGESTPATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES 2. SHORTESTPATH: Given a graph G = (V, E), does there exists a simple
More informationThe UnionFind Problem Kruskal s algorithm for finding an MST presented us with a problem in datastructure design. As we looked at each edge,
The UnionFind Problem Kruskal s algorithm for finding an MST presented us with a problem in datastructure design. As we looked at each edge, cheapest first, we had to determine whether its two endpoints
More informationFull and Complete Binary Trees
Full and Complete Binary Trees Binary Tree Theorems 1 Here are two important types of binary trees. Note that the definitions, while similar, are logically independent. Definition: a binary tree T is full
More informationChapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling
Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NPhard problem. What should I do? A. Theory says you're unlikely to find a polytime algorithm. Must sacrifice one
More informationA simpler and better derandomization of an approximation algorithm for Single Source RentorBuy
A simpler and better derandomization of an approximation algorithm for Single Source RentorBuy David P. Williamson Anke van Zuylen School of Operations Research and Industrial Engineering, Cornell University,
More information9th MaxPlanck Advanced Course on the Foundations of Computer Science (ADFOCS) PrimalDual Algorithms for Online Optimization: Lecture 1
9th MaxPlanck Advanced Course on the Foundations of Computer Science (ADFOCS) PrimalDual Algorithms for Online Optimization: Lecture 1 Seffi Naor Computer Science Dept. Technion Haifa, Israel Introduction
More informationMinimum Makespan Scheduling
Minimum Makespan Scheduling Minimum makespan scheduling: Definition and variants Factor 2 algorithm for identical machines PTAS for identical machines Factor 2 algorithm for unrelated machines Martin Zachariasen,
More informationLecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs
CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like
More informationExample 5.5. The heuristics are demonstrated with this example of n=5 vertices and distances in the following matrix:
39 5.3. HEURISTICS FOR THE TSP Notation: f = length of the tour given by the algortihm a f = length of the optimal tour min Most of the following heuristics are designed to solve symmetric instances of
More informationLevel Sets of Arbitrary Dimension Polynomials with Positive Coefficients and Real Exponents
Level Sets of Arbitrary Dimension Polynomials with Positive Coefficients and Real Exponents Spencer Greenberg April 20, 2006 Abstract In this paper we consider the set of positive points at which a polynomial
More informationMinimize subject to. x S R
Chapter 12 Lagrangian Relaxation This chapter is mostly inspired by Chapter 16 of [1]. In the previous chapters, we have succeeded to find efficient algorithms to solve several important problems such
More information17.3.1 Follow the Perturbed Leader
CS787: Advanced Algorithms Topic: Online Learning Presenters: David He, Chris Hopman 17.3.1 Follow the Perturbed Leader 17.3.1.1 Prediction Problem Recall the prediction problem that we discussed in class.
More informationLecture 22: November 10
CS271 Randomness & Computation Fall 2011 Lecture 22: November 10 Lecturer: Alistair Sinclair Based on scribe notes by Rafael Frongillo Disclaimer: These notes have not been subjected to the usual scrutiny
More informationMinimum cost maximum flow, Minimum cost circulation, Cost/Capacity scaling
6.854 Advanced Algorithms Lecture 16: 10/11/2006 Lecturer: David Karger Scribe: Kermin Fleming and Chris Crutchfield, based on notes by Wendy Chu and Tudor Leu Minimum cost maximum flow, Minimum cost circulation,
More information8.1 Min Degree Spanning Tree
CS880: Approximations Algorithms Scribe: Siddharth Barman Lecturer: Shuchi Chawla Topic: Min Degree Spanning Tree Date: 02/15/07 In this lecture we give a local search based algorithm for the Min Degree
More informationONLINE DEGREEBOUNDED STEINER NETWORK DESIGN. Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015
ONLINE DEGREEBOUNDED STEINER NETWORK DESIGN Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015 ONLINE STEINER FOREST PROBLEM An initially given graph G. s 1 s 2 A sequence of demands (s i, t i ) arriving
More information1 Distances and Metric Spaces
1 Distances and Metric Spaces Given a set X of points, a distance function on X is a map d : X X R + that is symmetric, and satisfies d(i, i) = 0 for all i X. The distance is said to be a metric if the
More informationAN EXTENSION OF THE ELIMINATION METHOD FOR A SPARSE SOS POLYNOMIAL
AN EXTENSION OF THE ELIMINATION METHOD FOR A SPARSE SOS POLYNOMIAL Hayato Waki Masakazu Muramatsu The University of ElectroCommunications (September 20, 2011) Abstract We propose a method to reduce the
More informationA constantfactor approximation algorithm for the kmedian problem
A constantfactor approximation algorithm for the kmedian problem Moses Charikar Sudipto Guha Éva Tardos David B. Shmoys July 23, 2002 Abstract We present the first constantfactor approximation algorithm
More informationIntroduction to Algorithms Review information for Prelim 1 CS 4820, Spring 2010 Distributed Wednesday, February 24
Introduction to Algorithms Review information for Prelim 1 CS 4820, Spring 2010 Distributed Wednesday, February 24 The final exam will cover seven topics. 1. greedy algorithms 2. divideandconquer algorithms
More informationCMPSCI611: Primality Testing Lecture 14
CMPSCI611: Primality Testing Lecture 14 One of the most important domains for randomized algorithms is number theory, particularly with its applications to cryptography. Today we ll look at one of the
More informationMinMax Multiway Cut
MinMax Multiway Cut Zoya Svitkina and Éva Tardos Cornell University, Department of Computer Science, Upson Hall, Ithaca, NY 14853. {zoya, eva}@cs.cornell.edu Abstract. We propose the Minmax multiway
More informationCS174 Lecture 3 John Canny
CS174 Lecture 3 John Canny Randomized Quicksort and BSTs A couple more needed results about permutations: Q1: What s the probability that 4 comes before 5 in a random permutation? Tempting to say 1/2,
More information1. R In this and the next section we are going to study the properties of sequences of real numbers.
+a 1. R In this and the next section we are going to study the properties of sequences of real numbers. Definition 1.1. (Sequence) A sequence is a function with domain N. Example 1.2. A sequence of real
More informationAlgorithm Design and Analysis
Algorithm Design and Analysis LECTURE 27 Approximation Algorithms Load Balancing Weighted Vertex Cover Reminder: Fill out SRTEs online Don t forget to click submit Sofya Raskhodnikova 12/6/2011 S. Raskhodnikova;
More informationSteven Skiena. skiena
Lecture 3: Recurrence Relations (1997) Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.sunysb.edu/ skiena Argue the solution to T(n) =
More informationTheorem (The division theorem) Suppose that a and b are integers with b > 0. There exist unique integers q and r so that. a = bq + r and 0 r < b.
Theorem (The division theorem) Suppose that a and b are integers with b > 0. There exist unique integers q and r so that a = bq + r and 0 r < b. We re dividing a by b: q is the quotient and r is the remainder,
More informationSection Notes 3. The Simplex Algorithm. Applied Math 121. Week of February 14, 2011
Section Notes 3 The Simplex Algorithm Applied Math 121 Week of February 14, 2011 Goals for the week understand how to get from an LP to a simplex tableau. be familiar with reduced costs, optimal solutions,
More informationSample Problems in Discrete Mathematics
Sample Problems in Discrete Mathematics This handout lists some sample problems that you should be able to solve as a prerequisite to Computer Algorithms Try to solve all of them You should also read
More informationApproximating Minimum Bounded Degree Spanning Trees to within One of Optimal
Approximating Minimum Bounded Degree Spanning Trees to within One of Optimal ABSTACT Mohit Singh Tepper School of Business Carnegie Mellon University Pittsburgh, PA USA mohits@andrew.cmu.edu In the MINIMUM
More informationCSL851: Algorithmic Graph Theory Semester I Lecture 4: August 5
CSL851: Algorithmic Graph Theory Semester I 201314 Lecture 4: August 5 Lecturer: Naveen Garg Scribes: Utkarsh Ohm Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More informationLecture 3: Finding integer solutions to systems of linear equations
Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture
More informationDefinition of a Linear Program
Definition of a Linear Program Definition: A function f(x 1, x,..., x n ) of x 1, x,..., x n is a linear function if and only if for some set of constants c 1, c,..., c n, f(x 1, x,..., x n ) = c 1 x 1
More informationInverse Optimization by James Orlin
Inverse Optimization by James Orlin based on research that is joint with Ravi Ahuja Jeopardy 000  the Math Programming Edition The category is linear objective functions The answer: When you maximize
More informationOutline. NPcompleteness. When is a problem easy? When is a problem hard? Today. Euler Circuits
Outline NPcompleteness Examples of Easy vs. Hard problems Euler circuit vs. Hamiltonian circuit Shortest Path vs. Longest Path 2pairs sum vs. general Subset Sum Reducing one problem to another Clique
More informationGood luck, veel succes!
Final exam Advanced Linear Programming, May 7, 13.0016.00 Switch off your mobile phone, PDA and any other mobile device and put it far away. No books or other reading materials are allowed. This exam
More information4.4 The recursiontree method
4.4 The recursiontree method Let us see how a recursion tree would provide a good guess for the recurrence = 3 4 ) Start by nding an upper bound Floors and ceilings usually do not matter when solving
More informationMincost flow problems and network simplex algorithm
Mincost flow problems and network simplex algorithm The particular structure of some LP problems can be sometimes used for the design of solution techniques more efficient than the simplex algorithm.
More informationPartitioning edgecoloured complete graphs into monochromatic cycles and paths
arxiv:1205.5492v1 [math.co] 24 May 2012 Partitioning edgecoloured complete graphs into monochromatic cycles and paths Alexey Pokrovskiy Departement of Mathematics, London School of Economics and Political
More information4 Basics of Trees. Petr Hliněný, FI MU Brno 1 FI: MA010: Trees and Forests
4 Basics of Trees Trees, actually acyclic connected simple graphs, are among the simplest graph classes. Despite their simplicity, they still have rich structure and many useful application, such as in
More informationMultimedia Communications. Huffman Coding
Multimedia Communications Huffman Coding Optimal codes Suppose that a i > w i C + is an encoding scheme for a source alphabet A={a 1,, a N }. Suppose that the source letter a 1,, a N occur with relative
More informationDefinition 11.1. Given a graph G on n vertices, we define the following quantities:
Lecture 11 The Lovász ϑ Function 11.1 Perfect graphs We begin with some background on perfect graphs. graphs. First, we define some quantities on Definition 11.1. Given a graph G on n vertices, we define
More information(67902) Topics in Theory and Complexity Nov 2, 2006. Lecture 7
(67902) Topics in Theory and Complexity Nov 2, 2006 Lecturer: Irit Dinur Lecture 7 Scribe: Rani Lekach 1 Lecture overview This Lecture consists of two parts In the first part we will refresh the definition
More informationConstant Factor Approximation Algorithm for the Knapsack Median Problem
Constant Factor Approximation Algorithm for the Knapsack Median Problem Amit Kumar Abstract We give a constant factor approximation algorithm for the following generalization of the kmedian problem. We
More informationBalanced Binary Search Tree
AVL Trees / Slide 1 Balanced Binary Search Tree Worst case height of binary search tree: N1 Insertion, deletion can be O(N) in the worst case We want a tree with small height Height of a binary tree with
More informationBalanced search trees
Lecture 8 Balanced search trees 8.1 Overview In this lecture we discuss search trees as a method for storing data in a way that supports fast insert, lookup, and delete operations. (Data structures handling
More informationLecture 2 : Basics of Probability Theory
Lecture 2 : Basics of Probability Theory When an experiment is performed, the realization of the experiment is an outcome in the sample space. If the experiment is performed a number of times, different
More informationGraphs and Network Flows IE411 Lecture 1
Graphs and Network Flows IE411 Lecture 1 Dr. Ted Ralphs IE411 Lecture 1 1 References for Today s Lecture Required reading Sections 17.1, 19.1 References AMO Chapter 1 and Section 2.1 and 2.2 IE411 Lecture
More information15750 Graduate Algorithms Spring 2010
15750 Graduate Algorithms Spring 2010 1 Cheap Cut (25 pts.) Suppose we are given a directed network G = (V, E) with a root node r and a set of terminal nodes T V. We would like to remove the fewest number
More informationGRAPH THEORY and APPLICATIONS. Trees
GRAPH THEORY and APPLICATIONS Trees Properties Tree: a connected graph with no cycle (acyclic) Forest: a graph with no cycle Paths are trees. Star: A tree consisting of one vertex adjacent to all the others.
More informationDiscrete (and Continuous) Optimization Solutions of Exercises 1 WI4 131
Discrete (and Continuous) Optimization Solutions of Exercises 1 WI4 131 Kees Roos Technische Universiteit Delft Faculteit Informatietechnologie en Systemen Afdeling Informatie, Systemen en Algoritmiek
More informationRANDOMIZATION IN APPROXIMATION AND ONLINE ALGORITHMS. Manos Thanos Randomized Algorithms NTUA
RANDOMIZATION IN APPROXIMATION AND ONLINE ALGORITHMS 1 Manos Thanos Randomized Algorithms NTUA RANDOMIZED ROUNDING Linear Programming problems are problems for the optimization of a linear objective function,
More informationOn the Union of Arithmetic Progressions
On the Union of Arithmetic Progressions Shoni Gilboa Rom Pinchasi August, 04 Abstract We show that for any integer n and real ɛ > 0, the union of n arithmetic progressions with pairwise distinct differences,
More informationLecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method
Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming
More informationBasic building blocks for a tripledouble intermediate format
Basic building blocks for a tripledouble intermediate format corrected version) Christoph Quirin Lauter February 26, 2009 Abstract The implementation of correctly rounded elementary functions needs high
More informationWhy graph clustering is useful?
Graph Clustering Why graph clustering is useful? Distance matrices are graphs as useful as any other clustering Identification of communities in social networks Webpage clustering for better data management
More information1 The Line vs Point Test
6.875 PCP and Hardness of Approximation MIT, Fall 2010 Lecture 5: Low Degree Testing Lecturer: Dana Moshkovitz Scribe: Gregory Minton and Dana Moshkovitz Having seen a probabilistic verifier for linearity
More informationGENERATING LOWDEGREE 2SPANNERS
SIAM J. COMPUT. c 1998 Society for Industrial and Applied Mathematics Vol. 27, No. 5, pp. 1438 1456, October 1998 013 GENERATING LOWDEGREE 2SPANNERS GUY KORTSARZ AND DAVID PELEG Abstract. A kspanner
More informationCHAPTER 9. Integer Programming
CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral
More informationIntroduction to Stationary Distributions
13 Introduction to Stationary Distributions We first briefly review the classification of states in a Markov chain with a quick example and then begin the discussion of the important notion of stationary
More informationCS420/520 Algorithm Analysis Spring 2010 Lecture 04
CS420/520 Algorithm Analysis Spring 2010 Lecture 04 I. Chapter 4 Recurrences A. Introduction 1. Three steps applied at each level of a recursion a) Divide into a number of subproblems that are smaller
More information2015:07. Tree Approximation for hp Adaptivity. Peter Binev
INTERDISCIPLINARY MATHEMATICS INSTITUTE 205:07 Tree Approximation for hp Adaptivity Peter Binev IMI PREPRINT SERIES COLLEGE OF ARTS AND SCIENCES UNIVERSITY OF SOUTH CAROLINA Tree Approximation for hpadaptivity
More informationAlgorithms for Virtual Private Network Design
Algorithms for Virtual Private Network Design Diploma Thesis of Thomas Rothvoß Advised by Prof. Dr. Friedrich Eisenbrand Second reviewer PD Dr. Piotr Krysta Fachbereich Informatik Universität Dortmund
More informationFE670 Algorithmic Trading Strategies. Stevens Institute of Technology
FE670 Algorithmic Trading Strategies Lecture 6. Portfolio Optimization: Basic Theory and Practice Steve Yang Stevens Institute of Technology 10/03/2013 Outline 1 MeanVariance Analysis: Overview 2 Classical
More informationBindel, Fall 2012 Matrix Computations (CS 6210) Week 8: Friday, Oct 12
Why eigenvalues? Week 8: Friday, Oct 12 I spend a lot of time thinking about eigenvalue problems. In part, this is because I look for problems that can be solved via eigenvalues. But I might have fewer
More informationLecture 6 Online and streaming algorithms for clustering
CSE 291: Unsupervised learning Spring 2008 Lecture 6 Online and streaming algorithms for clustering 6.1 Online kclustering To the extent that clustering takes place in the brain, it happens in an online
More information6.3 Conditional Probability and Independence
222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted
More informationTrees. Tree Definitions: Let T = (V, E, r) be a tree. The size of a tree denotes the number of nodes of the tree.
Trees After lists (including stacks and queues) and hash tables, trees represent one of the most widely used data structures. On one hand trees can be used to represent objects that are recursive in nature
More information