Generalized Binary Search

Size: px
Start display at page:

Download "Generalized Binary Search"

Transcription

1 Generalized Binary Search Robert Nowak, Department of Electrical and Computer Engineering, University of Wisconsin-Madison Abstract This paper studies a generalization of the classic binary search problem of locating a desired value within a sorted list. The classic problem can be viewed as determining the correct one-dimensional, binary-valued threshold function from a finite class of such functions based on queries taking the form of point samples of the function. The classic problem is also equivalent to a simple binary encoding of the threshold location. This paper extends binary search to learning more general binaryvalued functions. Specifically, if the set of target functions and queries satisfy certain geometrical relationships, then an algorithm, based on selecting a query that is imally discriminating at each step, will determine the correct function in a number of steps that is logarithmic in the number of functions under consideration. Examples of classes satisfying the geometrical relationships include linear separators in multiple dimensions. Extensions to handle noise are also discussed. Possible applications include machine learning, channel coding, and sequential experimental design. I. PROBLEM SPECIFICATION Binary search can be viewed as a simple guessing game in which one is given an ordered list and asked to determine an unknown target value by making queries of the form Is the target value greater than x? For example, consider the integer guessing game in which the list is the set of integers from 1 to 100. The optimal strategy, which is familiar to most people, is to first ask if the number is larger than 50, and then ask similar bisecting questions of the intervals that result from this and subsequent queries. At each step of this process, the uncertainty about the location of the unknown target is halved, and thus after j steps the number of remaining possibilities is no larger than (j+1). The binary search problem can also be cast as learning a one-dimensional threshold function from queries in the form of point samples. Consider the threshold function f(x) = 1 {x t} on the interval [0, 1], where t [0, 1) is the threshold location and 1 {x t} is 1 if x t and 0 otherwise. Suppose that t belongs to the set {0, 1 n,..., n 1 n }. The location of t can then determined from O(log n) point samples using a bisection procedure analogous to the process above. In fact, if n = 2 m for some integer m, then each point sample provides one bit in the m-bit binary expansion of t. Very similar strategies can be employed even if the answers to the queries are unreliable [1], [2], [3], the so-called noisy binary search problem. The first result that we are aware of here was due to [1], based on maintaining a probability distribution on the target value (initially uniform), querying/sampling at the median of the distribution at each step, and then adjusting the distribution based on the response/observation according to a quasi-bayes update. The method is based on a binary symmetric channel coding scheme that employs noiseless feedback [4]. Alternative approaches to the noisy binary search problem are essentially based on repeating each query in the classic binary search several times in order to be confident about the correct answer [2], [3]. This paper considers a generalized form of binary search based on the notion of imally discriminative queries. We consider an abstract setting in which queries are selected from a set. The correct response to a query x is either yes (+1) or no (-1), and is revealed by an oracle only after the query is selected. The queries are also put to a finite collection of hypotheses H with cardinality H. Each hypothesis h H is a mapping from to { 1, 1}. We assume that H contains the unknown oracle (i.e., the correct hypothesis) and that no two hypotheses agree on all possible queries (i.e., the hypotheses are unique with respect to ). The goal is to find the correct hypothesis as quickly as possible through a sequence of carefully chosen queries. In particular, we study the following algorithm, which selects the imally discriminating query at each step. Generalized Binary Search (GBS) initialize: i = 1, H 1 = H. while H i > 1 1) Select x i = arg min x i h(x). 2) Query oracle with x i to obtain response y i. 3) Set H i+1 = {h H i : h(x i ) = y i }, i = i + 1.

2 The query selection criterion picks a query that is imally discriminative at each step (e.g., if min x i h(x) = 0 then there exists a query for which half of the hypotheses predict +1 and the other half predict 1). There may be more than one query that achieves the minimum, and in that case any minimizer is acceptable. Since the hypotheses are unique with respect to, it is clear that the algorithm above terminates in at most H queries (since it is always possible to find query that eliminates at least one hypothesis at each step). Note that exhaustive linear search also requires O( H ) queries. However, if it is possible to select queries such that at each step a fixed fraction of the remaining viable hypotheses are eliminated, then the correct hypothesis will be found in O(log H ) steps. The main result of this paper shows that GBS exhibits this property, provided that and H satisfy certain geometrical relationships. Extensions to noisy GBS are also discussed. The emphasis in this paper is on determining the correct hypothesis with the fewest number of queries, and not on the computational complexity of selecting the queries. The motivation for this is that in many applications computational resources might be relatively inexpensive whereas obtaining the correct responses to queries may be very costly. Sequential strategies similar to GBS are quite common in the machine learning literature. For example, [5] considered a very similar problem, and showed that the expected number of queries required by a similar search algorithm is never too much larger than any other strategy. However, general conditions under which such strategies yield exponential speed-ups over exhaustive linear search were not determined. We mention also the work in [6], which draws parallels between binary search and source coding. That work, however, assumes the possibility of making arbitrary queries (rather than queries restricted to a certain set) and so in the present context the problem considered there essentially reduces to encoding each hypothesis with log H bits. Here we are interested in the interplay between a specific query space and the hypothesis space H. Classic binary search is an instance in which and H are matched so that search and source coding are essentially the same problem, as pointed out above. We identify geometrical conditions on the the pair (, H) that guarantee that GBS determines the correct hypothesis in O(log H ) queries. II. COMBINATORIAL CONDITIONS FOR GBS First consider an arbitrary sequential search procedure. Let i = 1, 2,... index the sequential process, x i denote the query at step i, and y i denote the correct response revealed by an oracle after the query is selected. If H i denotes the set of viable hypotheses at step i (i.e., all hypotheses consistent with the queries up to that step), then ideally a query x i is selected such that the resulting viable hypothesis space H i+1 satisfies H i+1 a i H i, for 0 < a i < 1, where H i denotes the cardinality of H i. This condition is met if and only if h(x i ) i c i H i (1) for some 0 c i < 1, in which case a i (1 + c i )/2. Condition (1) quantifies the degree of uncertainty among the hypotheses in H i for the query x i. The smaller the value of i h(x i ) the greater the uncertainty. Assuming that such an uncertainty condition holds for i = 1, 2,..., then after n steps of the algorithm n H n H (1 + c i )/2 i=1 and, in particular, the algorithm will terminate with the correct hypothesis as soon as H n i=1 (1 + c i)/2 1. Note that (1) trivially holds with c i = 1 2 H i 1 (a i = 1 H i 1 ), since there exists a query that eliminates at least one hypothesis at each step (recall that the hypotheses are assumed to be unique with respect to ). Thus, we are interested in cases in which the c i are uniformly bounded from above by a constant 0 c < 1 that does not depend on H. In that case, ( ) 1 + c n H n H 2 and the process terminates after at most log H / log(2/(1 + c)) = O(log H ) steps. Based on the observations above, we can state the following: Theorem 1. Let P(H) denote the power set of H. GBS converges to the correct hypothesis in O(log H ) if there exists a 0 c < 1 that does not depend on H such that inf G P(H) x G 1 h(x) c (2) h G Condition (2) is sufficient, but may not be necessary since certain subsets in P(H) may never result through

3 any sequence of queries. However, note that in the classic binary search setting, the condition does hold with c = 1/3, and thus the number of viable hypotheses is reduced by a factor of at least (1 + c)/2 = 2/3 at each step. The value of c = 1/3 is an upper bound that is achieved only in the worst-case situation, when G consists of three elements; most of the steps in classic binary search reduce the number of viable hypotheses by a factor of roughly 1/2. Unfortunately, verifying condition (2) in general is combinatorial, and so in the next section we seek conditions that are more easily verifiable. III. GEOMETRICAL CONDITIONS FOR GBS Let P denote a probability measure over, assume that every h H is measurable with respect to P, and define the constant 0 c P 1 by c P =. (3) Note that by the triangle inequality, (3) implies c P. (4) G P(H) G 1 h G Inequality (4) is a sort of relaxation of (2), with the minimization over replaced by an average, and its verification requires only the calculation of the first P - moment of each h H. Note that the minimal value of c P is given by c = min P, (5) where the minimization is over probability measures on. It is not hard to see that the minimizer exists because H is finite. Observe that the query space can be partitioned into a finite number of disjoint sets such that every h H is constant for all queries in each such set. Let A = A(, H) denote the collection of these sets, which are at most 2 H in number. The sets in A are equivalence classes in the following sense. For every A A and h H, the value of h(x) is constant (either +1 or 1) for all x A. Note that = A A A. Therefore, the minimization in (5) can be carried out over a space of finite-dimensional probability mass functions over the elements of A. The value of c will play an important role in characterizing the behavior of the GBS, but it does not need to be explicitly determined. Note that for each G P(H) one of two situations can occur: 1) min x G 1 h G h(x) c 2) min x G 1 h G h(x) > c If c is reasonably small, then the first situation guarantees that a good discriminating query exists (i.e., one that will reduce the number of viable hypotheses by a factor of at least (1 + c )/2). In the second situation, a highly discriminating query may not exist. The only guarantee is that there is always a query that eliminates at least one hypothesis, since the hypotheses are assumed to be unique with respect to. Therefore, a condition is required to ensure that such bad situations are not too problematic. Note that if min x G 1 h G h(x) > c, then there exist x, x such that G 1 h G h(x) > c and G 1 h G h(x ) < c. This follows since otherwise (4) cannot be satisfied with c. Under a mild condition discussed next, the existence of such an x and x implies that the cardinality of G must be rather small. The condition is given in terms of the two following definitions. Definition 1. Two sets A, A A are said to be k- neighbors if k or fewer hypotheses predict different values on A and A. For example, A and A are 1-neighbors if all but one element of H satisfy h(x) = h(x ) for all x A and x A. Definition 2. The query and hypothesis space (, H) are said to be k-neighborly if the k-neighborhood graph of A is connected (i.e., for every pair of sets in A there exists a sequence of k-neighbor sets that begins at one of the pair and ends with the other). Theorem 2. If (, H) is k-neighborly, then GBS terminates with the correct hypothesis after at most log H / log(α 1 ) queries, where α = { 1+c 2, k+1 k+2 } and c = min P. Remark 1. Note that GBS requires no knowledge of c. Proof: Let c be any number satisfying c c < 1 and let x i denote the query selected according to GBS at step i. If H i 1 i h(x i ) c, then the query x i reduces the number of viable hypotheses by a factor of at least (1 + c)/2. Otherwise, there exist x, x such that H i 1 i h(x) > c and H i 1 i h(x ) < c, since (4) must be satisfied with c according to the definition of c. Let A, A A denote the subsets containing x and x, respectively. The k-neighborly condition guarantees that there exists a sequence of k-neighbor sets beginning at A and ending at A. Note that H i 1 i h( ) > c on every set and the sign of H i 1 i h( ) must change

4 at some point in the sequence. It follows that there exist points x, x such that H i 1 i h(x) > c and H i 1 i h(x ) < c and furthermore, all but at most k of the hypotheses predict the same value for both x and x. Two inequalities follow from this observation. First, i h(x) i h(x ) > 2c H i. Second, i h(x) i h(x ) 2k. Combining these inequalities yields H i < k/c. Furthermore, there exists a query that eliminates at least one hypothesis due to the uniqueness of the hypotheses with respect to. Thus, at least one hypothesis must respond incorrectly to x i, and so H i+1 H i 1 = H i (1 H i 1 ) < H i (1 c/k). This shows that if H i 1 i h(x i ) > c, then the query x i reduces the number of viable hypotheses by a factor of at least (1 c/k). Also, recall that if H i 1 i h(x i ) c, then the query x i reduces the number of viable hypotheses by a factor of at least (1 + c)/2. It follows that the each GBS query reduces the number of viable hypotheses by a factor of at least min c c { 1 + c 2, 1 c/k } { 1 + c = 2 IV. APPLICATIONS, k + 1 k + 2 }. For a given pair (, H), the effectiveness of GBS hinges on determining (or bounding) c and establishing that (, H) are neighborly. Recall the definition of the bound c from (5). A trivial bound is 1 2 H 1, since this bound simply produces the convergence factor 1 H 1, which is achieved by an exhaustive linear search. Non-trivial moment bounds are those for which c, for a 0 c < 1 that does not depend unfavorably on H. In this section we consider several illustrative applications of GBS, calculating/bounding c and verifying neighborliness of (, H) in each case. A. Classic Binary Search Classic binary search can be viewed as the problem of determining a threshold value t (0, 1). Let H be a set of hypotheses of the form h v (x) = 2 1 {x>v} 1, where v V and V is a finite set of points in (0, 1) and 1 B denotes the indicator of the event B. Each query x [0, 1] receives a correct response y = 2 1 {x>t} 1 from the oracle. Assume that t V (i.e., the oracle is contained in H) and assume that V. First consider c. Assume that contains the points 0 and 1. Then taking P to be two point masses at x = 0 and x = 1 of probability 1/2 each yields = 0 for every h H, since h(0) = 1 and h(1) = 1 for every h H. Thus, c = 0. Now consider the neighborly condition. Recall that A is the partition on induced by H, such that for each set A A every h H has a constant response. In this case, each such set is an interval of the form A i = (v i 1, v i ], i = 1,..., V + 1, where v 1 < v 2 < < v V are the ordered values in V and v 0 = 0 and v V +1 = 1. Note that since V, each set A i contains at least one query. Furthermore, observe that only a single hypothesis, h vi, has different responses to queries from A i and A i+1. Thus, each successive pair of such sets are 1-neighbors. Moreover, the 1-neighborhood graph is connected in this case, and so (, H) are 1-neighborly. We conclude that the generalized binary search algorithm of Theorem 2 determines the optimal hypothesis in O(log H ) steps; i.e., the classic binary search result. B. Interval Classes Let = [0, 1] and consider a finite collection of hypotheses of the form h a,b (x) = 2 1 a x<b 1, with 0 a < b 1. Assume that the hypotheses do not have endpoints in common, and that one produces the correct prediction at all points in [0, 1]. The partition A again consists of intervals, and since there are no common endpoints, the neighborly condition is satisfied with k = 1. To bound c, note that the minimizing P must place some mass within and outside each such interval. If the intervals all have length at least l > 0, then taking P to be the uniform measure on [0, 1] yields that c 2l 1, irrespective of the number of interval hypotheses under consideration. Therefore, in this setting GBS determines the correct hypothesis in O(log H ) steps. However, consider the special case in which the intervals are disjoint. Then it is not hard to see that the best allocation of mass is to place 1/ H mass in each subinterval, resulting in c = 1 2 H 1. And so, GBS is not guaranteed to terminate in fewer than H steps (the number of steps required by exhaustive linear search). In this case, however, note that if queries of a different form were allowed, then much better performance is possible. For example, if queries in the form of dyadic subinterval tests were allowed (e.g., tests that indicate whether or not the correct hypothesis is +1-valued anywhere on a

5 dyadic subinterval of choice), then the correct hypothesis can be identified through O(log H ) queries (essentially a binary encoding of the correct hypothesis). This emphasizes the importance of the geometrical relationship between and H embodied in the neighborly condition and the value of c. Optimizing the query space to the structure of H is somewhat related to the ideas in [6] and to the theory of compressed sensing [7], [8]. C. Linear Separators in [ 1, 1] d Multi-dimensional threshold functions are particularly relevant in machine learning and pattern classification. Learning binary classifiers based on hyperplanes in d > 1 dimensions is thus an important generalization of classic binary search. Let = [ 1, 1] d, d 1, and consider a finite collection of hyperplanes of the form a, x + b = 0, where a R d, b R, and a, x is the inner product between a and x. Assume that every hyperplane in the collection is distinct and intersects the set ( 1, 1) d. Two d-dimensional threshold functions are associated with each hyperplane: h a,b (x) = 21 { a,x +b>0} 1 and h a,b (x). Let H denote the set of threshold functions formed from the finite collection of hyperplanes in this fashion. Assume that the correct label at each point x is given by one function in H. To bound c, let P be point masses of probability 2 d at each of the 2 d vertices of the cube [ 1, 1] d. Then 1 2 d+1 for every h H, since for each h there is at least one vertex on where it predicts +1 and one where it predicts 1. Thus, c 1 2 d+1. To verify the neighborly condition, note that in this case every set in the partition A is a polytope delineated by a subset of the hyperplanes. And, since the hyperplanes are distinct, two sets which share a common face are 2-neighbors (only the two hypotheses associated with the hyperplane that defines that face predict differently on queries from the two sets). Clearly, since the sets in A tesselate, the 2-neighborhood graph is connected and so (, H) is 2-neighborly. We conclude that the GBS determines the optimal hypothesis in O(2 d 1 log H ) steps. This appears to be a new result. A noteworthy case is the collection of hypotheses formed by threshold functions based on hyperplanes of the form a, x = 0, i.e., hyperplanes passing through the origin. In this case, with P as specified above, c = 0, since each hypothesis responds with +1 at half of the vertices and 1 on the other half. Therefore, GBS determines the optimal hypothesis in no more than O(log H ) steps, independent of the dimension. Related results for this special case have been previously reported; see [9] and the references therein. Note that even if the hyperplanes do not pass through the origin (b 0), O(log H ) convergence is still attained so long as b is not too large. This generalizes earlier results. In the case of general linear separators, the dependence on dimension d can also be eliminated with an additional assumption. Suppose that for a certain P on the P -moment of the optimal hypothesis is known to be upper bounded by a constant ρ < 1 that does not depend on H. Then all hypotheses that violate the bound can be eliminated from consideration and GBS applied to the set of remaining hypotheses will determine the correct hypothesis in O(log H ) steps. Situations like this can arise, for example, in binary classification problems with side/prior knowledge that the marginal probabilities of the two classes are somewhat balanced. Then the moment of the correct hypothesis, with respect to the marginal probability distribution of features, is bounded far away from 1 and 1. This provides another generalization of earlier results. D. Discrete Query Spaces In many situations both the hypothesis and query spaces may be discrete. A machine learning application, for example, may have access to a large (but finite) pool of unlabeled examples, any of which may be queried for a label. Because obtaining labels can be costly, active learning algorithms select only those examples that are predicted to be highly informative for labeling. Theorem 2 applies equally well to continuous or discrete query spaces. For example, consider the linear separator case, but instead of the query space [ 1, 1] d suppose that is a finite subset of points in [ 1, 1] d. The hypotheses again induce a partition of into subsets A(, H), but the number of subsets in the partition may be less than the number in A([ 1, 1] d, H). Consequently, the 2- neighborhood graph of A(, H) depends on the specific points that are included in and may or may not be connected. Consider two illustrative examples. Let H be a collection of linear separators as in Section IV-C above and first reconsider the partition A([ 1, 1] d, H). Recall that each set in A([ 1, 1] d, H) is a polytope. Suppose that a discrete set contains at least one point inside each of the polytopes in A([ 1, 1] d, H). Then it follows from the results above that (, H) is 2-neighborly. Second, consider a very simple case in d = 2 dimensions. Suppose consists of just three non-colinear points {x 1, x 2, x 3 } and suppose that H is comprised of six classifers, {h + 1, h 1, h+ 2, h 2, h+ 3, h 3 }, satisfying h+ i (x i) =

6 +1, h + i (x j) = 1, j i, i = 1, 2, 3, and h i = h + i, i = 1, 2, 3. In this case, A(, H) = {{x 1 }, {x 2 }, {x 3 }} and the responses to each pair of queries differ for four of the six hypotheses. Thus, the 4-neighborhood graph of A(, H) is connected, but the 2-neighborhood is not. hypothesis in at most O ( n o log(n o /δ) log log(n o /δ)/ɛ 2) where ɛ = p 1/2. steps, V. ETENSIONS TO NOISY SEARCH We now turn attention to the so-called noisy binary search problem. The situation considered here is that the oracle no longer returns the correct answer to every query, but instead responds correctly with probability at least 1 p and incorrectly with probability at most p, for an unknown 0 < p < 1/2. This is equivalent to the situation in which the oracle (sender) communicates answers to the learner (receiver) over a binary symmetric channel with crossover probability p, but the feedback channel (query channel) is noiseless. The goal remains to identify the correct hypothesis in H, despite the fact that the oracle may respond incorrectly. We will assume that the erroneous responses are determined by a random coin toss. Therefore, since the oracle is probably correct, one can decide the correct response to a given query (with very high confidence) by repeating it several times. This observation is the basis for most noisy binary search procedures, although optimal methods require a fairly delicate and subtle application of this basic intuition. To the best of our knowledge, a version of the classic binary search problem in noise was first considered by Horstein [4] in the context of channel coding with noiseless feedback. The first rigorous analysis, motivated by the work of Horstein, was developed in [1], where the information-theoretic optimality of a multiplicative weighting algorithm was established. A closely related set of results was recently reported in [3], which also includes results similar in spirit to [2]. We also mention the works of [10], [11], which consider adversarial situations in which the total number of erroneous oracle responses is fixed in advance. Based on an approach similar to that used in many of the papers above, we have the following result. Recall that under the assumptions of Theorem 2, GBS terminates after at most n o = O(log H ) queries. Theorem 3. If the oracle error probability is less than or equal to p, for some unknown 0 < p < 1/2, and the assumptions of Theorem 2 hold, then there exists a noisetolerant variant of GBS in the following sense: If GBS terminates in at most n o queries in the noiseless setting (p = 0), then there exists a modified search strategy that, with probability at least 1 δ, terminates with the correct Proof: The modified algorithm is based on the simple idea of repeating each query of the GBS several times, in order to overcome the uncertainty introduced by the noise. Since the value of p is unknown in advance, an adaptive procedure is required. Thus, we first recall Lemma 1 from [2]. Lemma 1. Consider a coin with an unknown probability p of heads. Then for any δ > 0 there exists an adaptive procedure for tossing the coin such that, with probability at least 1 δ, the number of coin tosses is at most m(δ ) = log(2/δ ) 4ɛ 2 ( log(2/δ ) ) log 4ɛ 2 and the procedure reports correctly whether heads or tails is more likely. The proof of the lemma and the procedure itself are based on relatively straightforward, iterated applications of Chernoff s bound; see [2] for further details. For the sake of completeness, we state the procedure here. Adaptive Coin Tossing Procedure initialize: set m o = 1 and toss the coin once. for j = 0, 1,... set 1) p j = frequency of heads (+1) 2) I j = [ p j (j+1) log(2/δ) 2 j, p j + ] (j+1) log(2/δ) 2 j 3) If 1/2 I j, then toss coin m j more times and set m j+1 = 2m j, otherwise break. end If I j [, 1/2], output 1, otherwise output +1. Now consider the n o queries chosen by GBS in the noiseless case. Repeat each query several times, according to the adaptive procedure above. By the union bound, with probability at least 1 n o δ, each query is repeated at most m(δ ) times and the correct responses to all n o queries are determined. Setting δ = δ/n o yields the upper bound on the number of queries. Whether or not the bound in Theorem 3 is optimal in the case of noisy GBS is an open question. For classic binary search with noise, more subtle procedures can be used to obtain slight improvements [1], [3].

7 VI. CONCLUSIONS This paper studied a generalization of the classic binary search problem. In particular, the generalized problem extends binary search techniques to multidimensional threshold functions, which arise in machine learning and pattern classification. If (, H) is neighborly (Definition 2) and if c does not depend explicity on H, then the number of steps required by GBS is O(log H ), exponentially smaller than the number of steps in an exhaustive linear search. The conditions express a geometrical relationship between and H which quantifies how well matched the queries are to the structure of hypotheses. The GBS problem can also be viewed as a source coding problem in which plays the role of a codeset and H plays the role of a source. In certain cases (e.g., classic binary search) GBS and ideal binary encoding are equivalent, but in general they are not. The neighborly condition and the value of c reflect the degree to which matches the source H. Finally, we point out that if the error probability is not bounded away from 1/2 in the noisy setting, then exponential speed-ups over linear search are no longer achievable by any search strategy. However, appropriate noisy binary search strategies can provide polynomial speed-ups over linear search [12], [13]. REFERENCES [1] M. V. Burnashev and K. S. Zigangirov, An interval estimation problem for controlled observations, Problems in Information Transmission, vol. 10, pp , [2] M. Kääriäinen, Active learning in the non-realizable case, in Algorithmic Learning Theory, 2006, pp [3] R. Karp and R. Kleinberg, Noisy binary search and its applications, in Proceedings of the 18th ACM-SIAM Symposium on Discrete Algorithms (SODA 2007), pp [4] M. Horstein, Sequential decoding using noiseless feedback, IEEE Trans. Info. Theory, vol. 9, no. 3, pp , [5] S. Dasgupta, Analysis of a greedy active learning strategy, in Neural Information Processing Systems, [6] S. R. Kulkarni, S. K. Mitter, and J. N. Tsitsiklis, Active learning using arbitrary binary valued queries, Machine Learning, pp , [7] E. J. Candès, J. Romberg, and T. Tao, Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information, IEEE Trans. Inform. Theory, vol. 52, no. 2, pp , Feb [8] D. L. Donoho, Compressed sensing, IEEE Trans. Inform. Theory, vol. 52, no. 4, pp , Apr [9] S. Dasgupta, Coarse sample complexity bounds for active learning, in Neural Information Processing Systems, [10] R. L. Rivest, A. R. Meyer, and D. J. Kleitman, Coping with errors in binary search procedure, J. Comput. System Sci., pp , [11] J. Spencer, Ulam s searching game with a fixed number of lies, in Theoretical Computer Science, 1992, pp. 95: [12] R. Castro and R. Nowak, Upper and lower bounds for active learning, in 44th Annual Allerton Conference on Communication, Control and Computing, [13], Mini bounds for active learning, IEEE Trans. Info. Theory, pp , 2008.

Notes from Week 1: Algorithms for sequential prediction

Notes from Week 1: Algorithms for sequential prediction CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking

More information

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

A Note on Maximum Independent Sets in Rectangle Intersection Graphs

A Note on Maximum Independent Sets in Rectangle Intersection Graphs A Note on Maximum Independent Sets in Rectangle Intersection Graphs Timothy M. Chan School of Computer Science University of Waterloo Waterloo, Ontario N2L 3G1, Canada tmchan@uwaterloo.ca September 12,

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete

More information

Applied Algorithm Design Lecture 5

Applied Algorithm Design Lecture 5 Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design

More information

Lecture 1: Course overview, circuits, and formulas

Lecture 1: Course overview, circuits, and formulas Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik

More information

A Network Flow Approach in Cloud Computing

A Network Flow Approach in Cloud Computing 1 A Network Flow Approach in Cloud Computing Soheil Feizi, Amy Zhang, Muriel Médard RLE at MIT Abstract In this paper, by using network flow principles, we propose algorithms to address various challenges

More information

A Sublinear Bipartiteness Tester for Bounded Degree Graphs

A Sublinear Bipartiteness Tester for Bounded Degree Graphs A Sublinear Bipartiteness Tester for Bounded Degree Graphs Oded Goldreich Dana Ron February 5, 1998 Abstract We present a sublinear-time algorithm for testing whether a bounded degree graph is bipartite

More information

arxiv:1203.1525v1 [math.co] 7 Mar 2012

arxiv:1203.1525v1 [math.co] 7 Mar 2012 Constructing subset partition graphs with strong adjacency and end-point count properties Nicolai Hähnle haehnle@math.tu-berlin.de arxiv:1203.1525v1 [math.co] 7 Mar 2012 March 8, 2012 Abstract Kim defined

More information

Metric Spaces. Chapter 7. 7.1. Metrics

Metric Spaces. Chapter 7. 7.1. Metrics Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some

More information

Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information

Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information Lawrence Ong School of Electrical Engineering and Computer Science, The University of Newcastle, Australia Email: lawrence.ong@cantab.net

More information

The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line

The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line D. Bongiorno, G. Corrao Dipartimento di Ingegneria lettrica, lettronica e delle Telecomunicazioni,

More information

CSC2420 Fall 2012: Algorithm Design, Analysis and Theory

CSC2420 Fall 2012: Algorithm Design, Analysis and Theory CSC2420 Fall 2012: Algorithm Design, Analysis and Theory Allan Borodin November 15, 2012; Lecture 10 1 / 27 Randomized online bipartite matching and the adwords problem. We briefly return to online algorithms

More information

Adaptive Online Gradient Descent

Adaptive Online Gradient Descent Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650

More information

Polarization codes and the rate of polarization

Polarization codes and the rate of polarization Polarization codes and the rate of polarization Erdal Arıkan, Emre Telatar Bilkent U., EPFL Sept 10, 2008 Channel Polarization Given a binary input DMC W, i.i.d. uniformly distributed inputs (X 1,...,

More information

Exponential time algorithms for graph coloring

Exponential time algorithms for graph coloring Exponential time algorithms for graph coloring Uriel Feige Lecture notes, March 14, 2011 1 Introduction Let [n] denote the set {1,..., k}. A k-labeling of vertices of a graph G(V, E) is a function V [k].

More information

A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property

A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property Venkat Chandar March 1, 2008 Abstract In this note, we prove that matrices whose entries are all 0 or 1 cannot achieve

More information

Competitive Analysis of On line Randomized Call Control in Cellular Networks

Competitive Analysis of On line Randomized Call Control in Cellular Networks Competitive Analysis of On line Randomized Call Control in Cellular Networks Ioannis Caragiannis Christos Kaklamanis Evi Papaioannou Abstract In this paper we address an important communication issue arising

More information

Interactive Machine Learning. Maria-Florina Balcan

Interactive Machine Learning. Maria-Florina Balcan Interactive Machine Learning Maria-Florina Balcan Machine Learning Image Classification Document Categorization Speech Recognition Protein Classification Branch Prediction Fraud Detection Spam Detection

More information

Notes on Complexity Theory Last updated: August, 2011. Lecture 1

Notes on Complexity Theory Last updated: August, 2011. Lecture 1 Notes on Complexity Theory Last updated: August, 2011 Jonathan Katz Lecture 1 1 Turing Machines I assume that most students have encountered Turing machines before. (Students who have not may want to look

More information

JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS. Received December May 12, 2003; revised February 5, 2004

JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS. Received December May 12, 2003; revised February 5, 2004 Scientiae Mathematicae Japonicae Online, Vol. 10, (2004), 431 437 431 JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS Ondřej Čepeka and Shao Chin Sung b Received December May 12, 2003; revised February

More information

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem

More information

Influences in low-degree polynomials

Influences in low-degree polynomials Influences in low-degree polynomials Artūrs Bačkurs December 12, 2012 1 Introduction In 3] it is conjectured that every bounded real polynomial has a highly influential variable The conjecture is known

More information

SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH

SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH 31 Kragujevac J. Math. 25 (2003) 31 49. SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH Kinkar Ch. Das Department of Mathematics, Indian Institute of Technology, Kharagpur 721302, W.B.,

More information

Near Optimal Solutions

Near Optimal Solutions Near Optimal Solutions Many important optimization problems are lacking efficient solutions. NP-Complete problems unlikely to have polynomial time solutions. Good heuristics important for such problems.

More information

E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

More information

Offline sorting buffers on Line

Offline sorting buffers on Line Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: rkhandekar@gmail.com 2 IBM India Research Lab, New Delhi. email: pvinayak@in.ibm.com

More information

Adaptive Linear Programming Decoding

Adaptive Linear Programming Decoding Adaptive Linear Programming Decoding Mohammad H. Taghavi and Paul H. Siegel ECE Department, University of California, San Diego Email: (mtaghavi, psiegel)@ucsd.edu ISIT 2006, Seattle, USA, July 9 14, 2006

More information

I. INTRODUCTION. of the biometric measurements is stored in the database

I. INTRODUCTION. of the biometric measurements is stored in the database 122 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL 6, NO 1, MARCH 2011 Privacy Security Trade-Offs in Biometric Security Systems Part I: Single Use Case Lifeng Lai, Member, IEEE, Siu-Wai

More information

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem

More information

5.1 Bipartite Matching

5.1 Bipartite Matching CS787: Advanced Algorithms Lecture 5: Applications of Network Flow In the last lecture, we looked at the problem of finding the maximum flow in a graph, and how it can be efficiently solved using the Ford-Fulkerson

More information

8.1 Min Degree Spanning Tree

8.1 Min Degree Spanning Tree CS880: Approximations Algorithms Scribe: Siddharth Barman Lecturer: Shuchi Chawla Topic: Min Degree Spanning Tree Date: 02/15/07 In this lecture we give a local search based algorithm for the Min Degree

More information

THIS paper deals with a situation where a communication

THIS paper deals with a situation where a communication IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 44, NO. 3, MAY 1998 973 The Compound Channel Capacity of a Class of Finite-State Channels Amos Lapidoth, Member, IEEE, İ. Emre Telatar, Member, IEEE Abstract

More information

Ideal Class Group and Units

Ideal Class Group and Units Chapter 4 Ideal Class Group and Units We are now interested in understanding two aspects of ring of integers of number fields: how principal they are (that is, what is the proportion of principal ideals

More information

A Practical Scheme for Wireless Network Operation

A Practical Scheme for Wireless Network Operation A Practical Scheme for Wireless Network Operation Radhika Gowaikar, Amir F. Dana, Babak Hassibi, Michelle Effros June 21, 2004 Abstract In many problems in wireline networks, it is known that achieving

More information

arxiv:1112.0829v1 [math.pr] 5 Dec 2011

arxiv:1112.0829v1 [math.pr] 5 Dec 2011 How Not to Win a Million Dollars: A Counterexample to a Conjecture of L. Breiman Thomas P. Hayes arxiv:1112.0829v1 [math.pr] 5 Dec 2011 Abstract Consider a gambling game in which we are allowed to repeatedly

More information

How To Find An Optimal Search Protocol For An Oblivious Cell

How To Find An Optimal Search Protocol For An Oblivious Cell The Conference Call Search Problem in Wireless Networks Leah Epstein 1, and Asaf Levin 2 1 Department of Mathematics, University of Haifa, 31905 Haifa, Israel. lea@math.haifa.ac.il 2 Department of Statistics,

More information

When is missing data recoverable?

When is missing data recoverable? When is missing data recoverable? Yin Zhang CAAM Technical Report TR06-15 Department of Computational and Applied Mathematics Rice University, Houston, TX 77005 October, 2006 Abstract Suppose a non-random

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

More information

Lecture 2: Universality

Lecture 2: Universality CS 710: Complexity Theory 1/21/2010 Lecture 2: Universality Instructor: Dieter van Melkebeek Scribe: Tyson Williams In this lecture, we introduce the notion of a universal machine, develop efficient universal

More information

CS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010

CS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010 CS 598CSC: Combinatorial Optimization Lecture date: /4/010 Instructor: Chandra Chekuri Scribe: David Morrison Gomory-Hu Trees (The work in this section closely follows [3]) Let G = (V, E) be an undirected

More information

1 The Line vs Point Test

1 The Line vs Point Test 6.875 PCP and Hardness of Approximation MIT, Fall 2010 Lecture 5: Low Degree Testing Lecturer: Dana Moshkovitz Scribe: Gregory Minton and Dana Moshkovitz Having seen a probabilistic verifier for linearity

More information

Basics of information theory and information complexity

Basics of information theory and information complexity Basics of information theory and information complexity a tutorial Mark Braverman Princeton University June 1, 2013 1 Part I: Information theory Information theory, in its modern format was introduced

More information

Multi-layer Structure of Data Center Based on Steiner Triple System

Multi-layer Structure of Data Center Based on Steiner Triple System Journal of Computational Information Systems 9: 11 (2013) 4371 4378 Available at http://www.jofcis.com Multi-layer Structure of Data Center Based on Steiner Triple System Jianfei ZHANG 1, Zhiyi FANG 1,

More information

Chapter 6: Episode discovery process

Chapter 6: Episode discovery process Chapter 6: Episode discovery process Algorithmic Methods of Data Mining, Fall 2005, Chapter 6: Episode discovery process 1 6. Episode discovery process The knowledge discovery process KDD process of analyzing

More information

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma Please Note: The references at the end are given for extra reading if you are interested in exploring these ideas further. You are

More information

ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN. Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015

ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN. Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015 ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015 ONLINE STEINER FOREST PROBLEM An initially given graph G. s 1 s 2 A sequence of demands (s i, t i ) arriving

More information

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results

More information

Lecture 17 : Equivalence and Order Relations DRAFT

Lecture 17 : Equivalence and Order Relations DRAFT CS/Math 240: Introduction to Discrete Mathematics 3/31/2011 Lecture 17 : Equivalence and Order Relations Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT Last lecture we introduced the notion

More information

ALMOST COMMON PRIORS 1. INTRODUCTION

ALMOST COMMON PRIORS 1. INTRODUCTION ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type

More information

Collinear Points in Permutations

Collinear Points in Permutations Collinear Points in Permutations Joshua N. Cooper Courant Institute of Mathematics New York University, New York, NY József Solymosi Department of Mathematics University of British Columbia, Vancouver,

More information

Fairfield Public Schools

Fairfield Public Schools Mathematics Fairfield Public Schools AP Statistics AP Statistics BOE Approved 04/08/2014 1 AP STATISTICS Critical Areas of Focus AP Statistics is a rigorous course that offers advanced students an opportunity

More information

INCIDENCE-BETWEENNESS GEOMETRY

INCIDENCE-BETWEENNESS GEOMETRY INCIDENCE-BETWEENNESS GEOMETRY MATH 410, CSUSM. SPRING 2008. PROFESSOR AITKEN This document covers the geometry that can be developed with just the axioms related to incidence and betweenness. The full

More information

An Approximation Algorithm for Bounded Degree Deletion

An Approximation Algorithm for Bounded Degree Deletion An Approximation Algorithm for Bounded Degree Deletion Tomáš Ebenlendr Petr Kolman Jiří Sgall Abstract Bounded Degree Deletion is the following generalization of Vertex Cover. Given an undirected graph

More information

An Overview of Knowledge Discovery Database and Data mining Techniques

An Overview of Knowledge Discovery Database and Data mining Techniques An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,

More information

Definition 11.1. Given a graph G on n vertices, we define the following quantities:

Definition 11.1. Given a graph G on n vertices, we define the following quantities: Lecture 11 The Lovász ϑ Function 11.1 Perfect graphs We begin with some background on perfect graphs. graphs. First, we define some quantities on Definition 11.1. Given a graph G on n vertices, we define

More information

A Non-Linear Schema Theorem for Genetic Algorithms

A Non-Linear Schema Theorem for Genetic Algorithms A Non-Linear Schema Theorem for Genetic Algorithms William A Greene Computer Science Department University of New Orleans New Orleans, LA 70148 bill@csunoedu 504-280-6755 Abstract We generalize Holland

More information

Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling

Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one

More information

Efficient Recovery of Secrets

Efficient Recovery of Secrets Efficient Recovery of Secrets Marcel Fernandez Miguel Soriano, IEEE Senior Member Department of Telematics Engineering. Universitat Politècnica de Catalunya. C/ Jordi Girona 1 i 3. Campus Nord, Mod C3,

More information

Introduction to Algorithms. Part 3: P, NP Hard Problems

Introduction to Algorithms. Part 3: P, NP Hard Problems Introduction to Algorithms Part 3: P, NP Hard Problems 1) Polynomial Time: P and NP 2) NP-Completeness 3) Dealing with Hard Problems 4) Lower Bounds 5) Books c Wayne Goddard, Clemson University, 2004 Chapter

More information

Mechanisms for Fair Attribution

Mechanisms for Fair Attribution Mechanisms for Fair Attribution Eric Balkanski Yaron Singer Abstract We propose a new framework for optimization under fairness constraints. The problems we consider model procurement where the goal is

More information

This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination. IEEE/ACM TRANSACTIONS ON NETWORKING 1 A Greedy Link Scheduler for Wireless Networks With Gaussian Multiple-Access and Broadcast Channels Arun Sridharan, Student Member, IEEE, C Emre Koksal, Member, IEEE,

More information

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm. Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of

More information

MA651 Topology. Lecture 6. Separation Axioms.

MA651 Topology. Lecture 6. Separation Axioms. MA651 Topology. Lecture 6. Separation Axioms. This text is based on the following books: Fundamental concepts of topology by Peter O Neil Elements of Mathematics: General Topology by Nicolas Bourbaki Counterexamples

More information

Testing Hereditary Properties of Non-Expanding Bounded-Degree Graphs

Testing Hereditary Properties of Non-Expanding Bounded-Degree Graphs Testing Hereditary Properties of Non-Expanding Bounded-Degree Graphs Artur Czumaj Asaf Shapira Christian Sohler Abstract We study graph properties which are testable for bounded degree graphs in time independent

More information

ABSTRACT. For example, circle orders are the containment orders of circles (actually disks) in the plane (see [8,9]).

ABSTRACT. For example, circle orders are the containment orders of circles (actually disks) in the plane (see [8,9]). Degrees of Freedom Versus Dimension for Containment Orders Noga Alon 1 Department of Mathematics Tel Aviv University Ramat Aviv 69978, Israel Edward R. Scheinerman 2 Department of Mathematical Sciences

More information

Private Approximation of Clustering and Vertex Cover

Private Approximation of Clustering and Vertex Cover Private Approximation of Clustering and Vertex Cover Amos Beimel, Renen Hallak, and Kobbi Nissim Department of Computer Science, Ben-Gurion University of the Negev Abstract. Private approximation of search

More information

On Adaboost and Optimal Betting Strategies

On Adaboost and Optimal Betting Strategies On Adaboost and Optimal Betting Strategies Pasquale Malacaria School of Electronic Engineering and Computer Science Queen Mary, University of London Email: pm@dcs.qmul.ac.uk Fabrizio Smeraldi School of

More information

On the k-path cover problem for cacti

On the k-path cover problem for cacti On the k-path cover problem for cacti Zemin Jin and Xueliang Li Center for Combinatorics and LPMC Nankai University Tianjin 300071, P.R. China zeminjin@eyou.com, x.li@eyou.com Abstract In this paper we

More information

x a x 2 (1 + x 2 ) n.

x a x 2 (1 + x 2 ) n. Limits and continuity Suppose that we have a function f : R R. Let a R. We say that f(x) tends to the limit l as x tends to a; lim f(x) = l ; x a if, given any real number ɛ > 0, there exists a real number

More information

Lattice-Based Threshold-Changeability for Standard Shamir Secret-Sharing Schemes

Lattice-Based Threshold-Changeability for Standard Shamir Secret-Sharing Schemes Lattice-Based Threshold-Changeability for Standard Shamir Secret-Sharing Schemes Ron Steinfeld (Macquarie University, Australia) (email: rons@ics.mq.edu.au) Joint work with: Huaxiong Wang (Macquarie University)

More information

An Empirical Study of Two MIS Algorithms

An Empirical Study of Two MIS Algorithms An Empirical Study of Two MIS Algorithms Email: Tushar Bisht and Kishore Kothapalli International Institute of Information Technology, Hyderabad Hyderabad, Andhra Pradesh, India 32. tushar.bisht@research.iiit.ac.in,

More information

The Advantages and Disadvantages of Online Linear Optimization

The Advantages and Disadvantages of Online Linear Optimization LINEAR PROGRAMMING WITH ONLINE LEARNING TATSIANA LEVINA, YURI LEVIN, JEFF MCGILL, AND MIKHAIL NEDIAK SCHOOL OF BUSINESS, QUEEN S UNIVERSITY, 143 UNION ST., KINGSTON, ON, K7L 3N6, CANADA E-MAIL:{TLEVIN,YLEVIN,JMCGILL,MNEDIAK}@BUSINESS.QUEENSU.CA

More information

Class constrained bin covering

Class constrained bin covering Class constrained bin covering Leah Epstein Csanád Imreh Asaf Levin Abstract We study the following variant of the bin covering problem. We are given a set of unit sized items, where each item has a color

More information

International Journal of Information Technology, Modeling and Computing (IJITMC) Vol.1, No.3,August 2013

International Journal of Information Technology, Modeling and Computing (IJITMC) Vol.1, No.3,August 2013 FACTORING CRYPTOSYSTEM MODULI WHEN THE CO-FACTORS DIFFERENCE IS BOUNDED Omar Akchiche 1 and Omar Khadir 2 1,2 Laboratory of Mathematics, Cryptography and Mechanics, Fstm, University of Hassan II Mohammedia-Casablanca,

More information

6.852: Distributed Algorithms Fall, 2009. Class 2

6.852: Distributed Algorithms Fall, 2009. Class 2 .8: Distributed Algorithms Fall, 009 Class Today s plan Leader election in a synchronous ring: Lower bound for comparison-based algorithms. Basic computation in general synchronous networks: Leader election

More information

THE SCHEDULING OF MAINTENANCE SERVICE

THE SCHEDULING OF MAINTENANCE SERVICE THE SCHEDULING OF MAINTENANCE SERVICE Shoshana Anily Celia A. Glass Refael Hassin Abstract We study a discrete problem of scheduling activities of several types under the constraint that at most a single

More information

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm. Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three

More information

Guessing Game: NP-Complete?

Guessing Game: NP-Complete? Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES 2. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple

More information

THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING

THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING 1. Introduction The Black-Scholes theory, which is the main subject of this course and its sequel, is based on the Efficient Market Hypothesis, that arbitrages

More information

The Goldberg Rao Algorithm for the Maximum Flow Problem

The Goldberg Rao Algorithm for the Maximum Flow Problem The Goldberg Rao Algorithm for the Maximum Flow Problem COS 528 class notes October 18, 2006 Scribe: Dávid Papp Main idea: use of the blocking flow paradigm to achieve essentially O(min{m 2/3, n 1/2 }

More information

2.1 Complexity Classes

2.1 Complexity Classes 15-859(M): Randomized Algorithms Lecturer: Shuchi Chawla Topic: Complexity classes, Identity checking Date: September 15, 2004 Scribe: Andrew Gilpin 2.1 Complexity Classes In this lecture we will look

More information

GENERIC COMPUTABILITY, TURING DEGREES, AND ASYMPTOTIC DENSITY

GENERIC COMPUTABILITY, TURING DEGREES, AND ASYMPTOTIC DENSITY GENERIC COMPUTABILITY, TURING DEGREES, AND ASYMPTOTIC DENSITY CARL G. JOCKUSCH, JR. AND PAUL E. SCHUPP Abstract. Generic decidability has been extensively studied in group theory, and we now study it in

More information

The Online Set Cover Problem

The Online Set Cover Problem The Online Set Cover Problem Noga Alon Baruch Awerbuch Yossi Azar Niv Buchbinder Joseph Seffi Naor ABSTRACT Let X = {, 2,..., n} be a ground set of n elements, and let S be a family of subsets of X, S

More information

Victor Shoup Avi Rubin. fshoup,rubing@bellcore.com. Abstract

Victor Shoup Avi Rubin. fshoup,rubing@bellcore.com. Abstract Session Key Distribution Using Smart Cards Victor Shoup Avi Rubin Bellcore, 445 South St., Morristown, NJ 07960 fshoup,rubing@bellcore.com Abstract In this paper, we investigate a method by which smart

More information

Factoring & Primality

Factoring & Primality Factoring & Primality Lecturer: Dimitris Papadopoulos In this lecture we will discuss the problem of integer factorization and primality testing, two problems that have been the focus of a great amount

More information

L25: Ensemble learning

L25: Ensemble learning L25: Ensemble learning Introduction Methods for constructing ensembles Combination strategies Stacked generalization Mixtures of experts Bagging Boosting CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna

More information

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

More information

Lecture 4 Online and streaming algorithms for clustering

Lecture 4 Online and streaming algorithms for clustering CSE 291: Geometric algorithms Spring 2013 Lecture 4 Online and streaming algorithms for clustering 4.1 On-line k-clustering To the extent that clustering takes place in the brain, it happens in an on-line

More information

LEARNING OBJECTIVES FOR THIS CHAPTER

LEARNING OBJECTIVES FOR THIS CHAPTER CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional

More information

Protein Protein Interaction Networks

Protein Protein Interaction Networks Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks Young-Rae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics

More information

Large induced subgraphs with all degrees odd

Large induced subgraphs with all degrees odd Large induced subgraphs with all degrees odd A.D. Scott Department of Pure Mathematics and Mathematical Statistics, University of Cambridge, England Abstract: We prove that every connected graph of order

More information

ON THE COMPLEXITY OF THE GAME OF SET. {kamalika,pbg,dratajcz,hoeteck}@cs.berkeley.edu

ON THE COMPLEXITY OF THE GAME OF SET. {kamalika,pbg,dratajcz,hoeteck}@cs.berkeley.edu ON THE COMPLEXITY OF THE GAME OF SET KAMALIKA CHAUDHURI, BRIGHTEN GODFREY, DAVID RATAJCZAK, AND HOETECK WEE {kamalika,pbg,dratajcz,hoeteck}@cs.berkeley.edu ABSTRACT. Set R is a card game played with a

More information

Fairness in Routing and Load Balancing

Fairness in Routing and Load Balancing Fairness in Routing and Load Balancing Jon Kleinberg Yuval Rabani Éva Tardos Abstract We consider the issue of network routing subject to explicit fairness conditions. The optimization of fairness criteria

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Markov random fields and Gibbs measures

Markov random fields and Gibbs measures Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).

More information

Approximated Distributed Minimum Vertex Cover Algorithms for Bounded Degree Graphs

Approximated Distributed Minimum Vertex Cover Algorithms for Bounded Degree Graphs Approximated Distributed Minimum Vertex Cover Algorithms for Bounded Degree Graphs Yong Zhang 1.2, Francis Y.L. Chin 2, and Hing-Fung Ting 2 1 College of Mathematics and Computer Science, Hebei University,

More information

Polynomial Degree and Lower Bounds in Quantum Complexity: Collision and Element Distinctness with Small Range

Polynomial Degree and Lower Bounds in Quantum Complexity: Collision and Element Distinctness with Small Range THEORY OF COMPUTING, Volume 1 (2005), pp. 37 46 http://theoryofcomputing.org Polynomial Degree and Lower Bounds in Quantum Complexity: Collision and Element Distinctness with Small Range Andris Ambainis

More information

P versus NP, and More

P versus NP, and More 1 P versus NP, and More Great Ideas in Theoretical Computer Science Saarland University, Summer 2014 If you have tried to solve a crossword puzzle, you know that it is much harder to solve it than to verify

More information