Lifted MAP Inference for Markov Logic Networks
|
|
|
- Georgiana Harmon
- 10 years ago
- Views:
Transcription
1 Somdeb Sarkhel, Deepak Venugopal Parag Singla Vibhav Gogate Department of Computer Science The University of Texas at Dallas Richardson, TX 75080, USA Department of CSE I.I.T. Delhi New Delhi, India Department of Computer Science The University of Texas at Dallas Richardson, TX 75080, USA Abstract In this paper, we present a new approach for lifted MAP inference in Markov Logic Networks (MLNs). Our approach is based on the following key result that we prove in the paper: if an MLN has no shared terms then MAP inference over it can be reduced to MAP inference over a Markov network having the following properties: (i) the number of random variables in the Markov network is equal to the number of first-order atoms in the MLN; and (ii) the domain size of each variable in the Markov network is equal to the number of groundings of the corresponding firstorder atom. We show that inference over this Markov network is exponentially more efficient than ground inference, namely inference over the Markov network obtained by grounding all firstorder atoms in the MLN. We improve this result further by showing that if non-shared MLNs contain no self joins, namely every atom appears at most once in each of its formulas, then all variables in the corresponding Markov network need only be bi-valued. Our approach is quite general and can be easily applied to an arbitrary MLN by simply grounding all of its shared terms. The key feature of our approach is that because we reduce lifted inference to propositional inference, we can use any propositional MAP inference algorithm for performing lifted MAP inference. Within our approach, we experimented with two propositional MAP inference algorithms: Gurobi and MaxWalkSAT. Our experiments on several benchmark MLNs clearly demonstrate our approach is superior to ground MAP inference in terms of scalability and solution quality. Appearing in Proceedings of the 17 th International Conference on Artificial Intelligence and Statistics (AISTATS) 2014, Reykjavik, Iceland. JMLR: W&CP volume 33. Copyright 2014 by the authors. 1 Introduction Statistical relational models [6] such as Markov logic networks [5] bring the power and compactness of first-order logic to probabilistic graphical models. They are routinely used to solve hard problems in a wide variety of real-world application domains including computer vision, natural language processing, robotics and the Web. Recently, there has been a growing interest in exploiting relational structure during inference in statistical relational models. Unlike propositional inference algorithms which operate on individual random variables, these algorithms, which are often called lifted or first-order inference algorithms [17], perform inference over a group of random variables. As a result, they are more scalable, typically exponentially faster when relational structure is present, than propositional inference algorithms. In this paper, we present a general approach for lifted MAP inference in MLNs. Our approach is based on our key result that when the MLN contains no formulas having shared terms (we refer to such MLNs as a non-shared MLN), MAP inference is domain liftable [3], namely it is polynomial in the domain size of the logical variables in the MLN. In particular, we show that in non-shared MLNs, the set of full assignments having the cardinality O(2 n i=1 di ), where n is the number of predicates in the MLN and d i is the number of possible groundings of the i-th atom in the MLN, 1 can be partitioned into O( n i=1 (d i + 1)) subsets such that each element in a subset has the same probability. Thus, instead of performing a search for the MAP solution over O(2 n i=1 di ) assignments as the ground inference algorithms do, we can perform the search over an exponentially smaller O( n i=1 (d i + 1)) space, yielding a lifted MAP inference algorithm. We further extend this result by showing that if the nonshared MLN has no self joins, namely each atom appears at most once in each formula, then one of the MAP solutions is guaranteed to lie at the following extreme points: for each atom, all of its groundings are either all true or 1 Number of possible groundings of an atom is the product of the domain sizes of the logical variables appearing in it.
2 all false. This helps us to further reduce the complexity of inference from O( n i=1 (d i + 1)) to O(2 n ), i.e., the complexity of inference is independent of the domain size. We utilize the aforementioned results by developing a mapping from the lifted search space to propositional search space of the same size. This helps us reformulate the MAP inference task over non-shared MLNs as propositional MAP inference over a Markov network such that: (i) the number of random variables in the Markov network is equal to the number of atoms in the MLN (i.e., equal to n) and (ii) the domain size of each random variable is either d i or 2 depending upon whether the non-shared MLN has self-joins or not. The two key features of this formulation are: (i) we can plug in any known propositional MAP inference algorithm for inference on this Markov network; and (ii) since the propositional algorithm operates in the lifted search space, it has the same performance guarantees as a lifted algorithm. Thus, by plugging in different propositional MAP inference algorithms, our approach yields a family of lifted MAP inference algorithms. Our approach is quite general and can be extended to arbitrary MLNs that have shared terms in a straight-forward way: simply ground all the shared terms of the MLN to obtain an equivalent non-shared MLN. The only caveat is that if all terms in all atoms of the MLN are shared, our approach will have the same complexity as ground inference but never worse (e.g., in the transitive formula x, y, z R(x, y) R(y, z) R(x, z), all terms of R are shared). Note that this is a limitation of not only our approach but lifted inference in general: it is only useful when symmetries are present. We experimentally evaluated our approach on three benchmark MLNs: WebKB and Information Extraction MLNs available from the Alchemy [12] web page and the Student MLN created by us. We used two state-of-the-art MAP inference algorithms within our approach: (i) Gurobi [8], which is an integer linear programming solver and (ii) MaxWalkSAT [11] which is a popular local search solver. Our experiments clearly show that our approach is significantly better in terms of solution quality and scalability than ground inference. In particular, as we increase the number of objects in the MLN, our algorithms are an order of magnitude faster and better in terms of solution quality. The rest of the paper is organized as follows. In section 2, we present preliminaries. In section 3, we present our new approach and in section 4, we extend it with several heuristics and pruning techniques. In section 5, we present experimental results and conclude in section 6. 2 Notation and Preliminaries In this section, we describe notation and preliminaries on propositional logic, first-order logic, Markov logic and MAP inference. For more details, refer to [5, 9, 13]. 2.1 Propositional and First-order Logic The language of propositional logic consists of atomic sentences called propositions or atoms, and logical connectives such as (conjunction), (disjunction), (negation), (implication) and (equivalence). Each proposition takes values from the binary domain {False, True} (or {0, 1}). A propositional formula f is an atom, or any complex formula that can be constructed from atoms using logical connectives. For example, A, B and C are propositional atoms and f = A B C is a propositional formula. A knowledge base (KB) is a set of formulas. A world is a truth assignment to all atoms in the KB. First-order logic (FOL) generalizes propositional logic by allowing atoms to have internal structure; an atom in FOL is a predicate that represents relations between objects. A predicate consists of a predicate symbol, denoted by Monospace fonts (e.g., Friends, Smokes, etc.), followed by a parenthesized list of arguments called terms. A term is a logical variable, denoted by lower case letters (e.g., x, y, etc.), or a constant, denoted by upper case letters (e.g., X, Y, etc.). We assume that each logical variable, say x, is typed and takes values over a finite set (called domain) x. The language of FOL also includes two quantifiers: (universal) and (existential) which express properties of an entire collection of objects. A formula in first order logic is a predicate (atom), or any complex sentence that can be constructed from atoms using logical connectives and quantifiers. For example, the formula x Smokes(x) Asthma(x) states that all persons who smoke have asthma. Where as x Cancer(x) states that there exists a person x who has cancer. In this paper, we use a subset of FOL which has no function symbols, equality constraints or existential quantifiers. We also assume that domains are finite (and therefore function-free) and that there is a one-to-one mapping between constants and objects in the domain (Herbrand interpretations). We assume that each formula f is of the form x f, where x are the set of variables in f and f is a conjunction or disjunction of literals; each literal being an atom or its negation. For brevity, we will drop from all the formulas. Given variables x = {x 1,..., x n } and constants X = {X 1,..., X n } where X i x i, f[x/x] is obtained by substituting every occurrence of variable x i in f with X i. A ground formula is a formula obtained by substituting all of its variables with a constant. A ground KB is a KB containing all possible groundings of all of its formulas. For example, the grounding of a KB containing one formula, Smokes(x) Asthma(x) where x = {Ana, Bob}, is a KB containing two formulas: Smokes(Ana) Asthma(Ana) and Smokes(Bob) Asthma(Bob).
3 Somdeb Sarkhel, Deepak Venugopal, Parag Singla, Vibhav Gogate 2.2 Markov Logic Markov logic [5] extends FOL by softening the hard constraints expressed by the formulas and is arguably the most popular modeling language for SRL. A soft formula or a weighted formula is a pair (f, w) where f is a formula in FOL and w is a real-number. A Markov logic network (MLN), denoted by M, is a set of weighted formulas (f i, w i ). Given a set of constants that represent objects in the domain, a Markov logic network defines a Markov network or a log-linear model. The Markov network is obtained by grounding the weighted first-order knowledge base and represents the following probability distribution. P M (ω) = 1 Z(M) exp ( ) w i N(f i, ω) where ω is a world, N(f i, ω) is the number of groundings of f i that evaluate to True in the world ω and Z(M) is a normalization constant or the partition function. Throughout the paper, we will assume that the input MLN to our algorithm is in normal form [10]. We require this for simplicity of exposition. A normal MLN is an MLN that satisfies the following two properties: (1) There are no constants in any formula, and (2) If two distinct atoms with the same predicate symbol have variables x and y in the same position then x = y. Any MLN can be converted to a normal MLN. Note that in a normal MLN, we assume that the terms in each atom are ordered and therefore we can identify each term by its position in the order. Furthermore, we assume that the MLN is expressed as a set of weighted clauses. 2.3 MAP Inference in MLNs A common optimization inference task over MLNs is finding the most probable state of the world ω, that is finding a complete assignment to all ground atoms which maximizes the probability. This task is known as Maximum a Posteriori (MAP) inference in the Markov network literature, and Most Probable Explanation (MPE) inference in the Bayesian network literature. For Markov logic, this is formally defined as follows: arg max PM(ω) = arg max ω ω = arg max ω i (1) ( ) 1 Z(M) exp w in(f i, ω) i w in(f i, ω) (2) From Eq. (2), we can see that the MAP problem in Markov logic reduces to finding the truth assignment that maximizes the sum of weights of satisfied clauses. Therefore, any weighted satisfiability solver can be used to solve this problem. The problem is NP-hard in general, but effective solvers exist, both exact and approximate. Examples of such solvers are MaxWalkSAT, a weighted variant i R(A) R(B) S(A) S(B) Weight Groups (0,0) w 1 + w 3 (0,1) w 1 + w 3 (0,1) w 1 + 2w 3 (0,2) w 1 + w 2 (1,0) w 1 + w 2 + w 3 (1,1) w 1 + w 2 + w 3 (1,1) w 1 + w 2 + 2w 3 (1,2) w 1 + w 2 (1,0) w 1 + w 2 + w 3 (1,1) w 1 + w 2 + w 3 (1,1) w 1 + 2w 3 + w 2 (1,2) w 1 + 2w 2 (2,0) w 1 + 2w 2 + w 3 (2,1) w 1 + 2w 2 + w 3 (2,1) w 1 + 2w 2 + 2w 3 (2,2) Figure 1: Weights of all assignments to ground atoms and (lifted) groups for the non-shared MLN: [R(x) S(y), w 1 ]; [R(x), w 2 ]; and [S(y), w 3 ] with domains given by x = y = {A, B}. of the WalkSAT local-search satisfiability solver [18] and Clone [16], a branch-and-bound solver. However, all of these algorithms are propositional and therefore they cannot exploit relational structure that is inherent to MLNs. 3 Lifted Formulation of MAP Inference In this section, we show that we can reduce MAP inference in a sub-class of MLNs, which we call non-shared MLNs, to MAP inference over an equivalent propositional MLN (or a propositional Markov network) such that the number of propositional variables in the propositional MLN is equal to the number of first order atoms in the non-shared MLN. This is in contrast to ground MAP inference in which the number of propositional variables is equal to the number of ground atoms. We begin by defining non-shared MLNs. Definition 1. A normal MLN is called a non-shared MLN if each of its formulas is non-shared. A formula f i is nonshared if every logical variable appears at most once in the formula. In other words, in a non-shared MLN, no logical variable is shared between the atoms in a formula. For example, R(x) S(y) is a non-shared formula. However, R(x) S(x) is not because x is shared. 3.1 Domain-Lifted MAP Inference over Non-Shared MLNs We show that MAP inference over non-shared MLNs is domain liftable [3], namely inference over it is polynomial in the domain size of the logical variables in the MLN. The
4 key reason that non-shared MLNs are domain liftable is that they contain several worlds with the same probability. We can group together these equi-probable worlds and perform MAP inference by just iterating over the groups, selecting the group with the maximum probability. The following example illustrates this grouping. Example 1. Consider the non-shared MLN containing three formulas: [R(x) S(y), w 1 ]; [R(x), w 2 ]; and [S(y), w 3 ]. Let x = y = {A, B}. Figure 1 gives a truth table showing all possible assignments to the ground atoms as well as their weights. Figure 1 also shows nine equi-probable groups for these assignments. It turns out that each group can be represented by a pair (i, j) where i and j are the number of true groundings of R and S respectively. Namely, i, j {0, 1, 2}. Thus, to compute the MAP tuple, we only have to iterate over 9 groups while the ground (naive) MAP inference algorithm will iterate over 16 assignments. In general, the number of groups will be equal to ( x + 1)( y + 1) while the number of possible assignments to the ground atoms equals 2 x + y. We can generalize the ideas presented in Example 1 using the following theorem: Theorem 1. Given a non-shared MLN M, let ω 1 and ω 2 be two worlds such that for each atom R in the MLN, the number of true groundings of R in ω 1 is equal to the number of true groundings of R in ω 2. Then, Pr M (ω 1 ) = Pr M (ω 2 ). Proof. We will prove this theorem by leveraging the generalized binomial rule [10]. The generalized Binomial rule states that if an atom R is non-shared (which is a special case of singleton atoms), then the MLNs obtained by conditioning on the following subset of assignments to all groundings of R are equivalent: the number of true groundings of R is the same in all the assignments in the subset. Moreover, according to the rule, the following two conditions hold: if the MLN is non-shared then the new MLN is also non-shared the number of formulas involving R satisfied by each assignment in the subset is the same. Let R 1,..., R n be the atoms in the MLN M. Let d i be the domain size of R i. Let R i = j i,1 and R i = j i,2 where j i,k 2 di, k {1, 2} denote the assignment to all groundings of R i in the worlds ω 1 and ω 2 respectively. Let us condition the atoms along the order R 1,..., R n. By the generalized Binomial rule, the MLN obtained by conditioning on R 1 = j 1,1, denoted by M R 1 = j 1,1 is equivalent to the MLN M R 1 = j 1,2 obtained by conditioning on R 1 = j 1,2 (since the number of true groundings of R i is the same in both the assignments). Let w(m R i = j i,k ), k {1, 2} denote the sum of the weights of clauses satisfied by conditioning on the assignment R i = j i,k. By the generalized Binomial rule, w(m R 1 = j 1,1 ) = w(m R 1 = j 1,2 ). Moreover, since all atoms in M R 1 = j 1,1 and M R 2 = j 2,1 are non-shared, it follows that the MLNs obtained by further conditioning on R 2 = j 2,1 is the same as the one obtained by conditioning on R 2 = j 2,2. By iteratively (inductively) applying this argument, we have: n w(m R 1 = j 1,1,..., R i 1 = j i 1,1 ) i=1 = n w(m R 1 = j 1,2,..., R i 1 = j i 1,2 ) (3) i=1 In other words, the two worlds ω 1 and ω 2 have the same weight. Therefore, Pr M (ω 1 ) = Pr M (ω 2 ). Theorem 1 yields the following lifted inference algorithm. Let {R 1, R 2,..., R n } be the atoms in the non-shared MLN. Let d i denote the domain size of R i (the domain of an atom equals the cartesian product of the domains of its logical variables). By Theorem 1, all the ground assignments of the MLN can be grouped into assignments of the form (R i, a i ) i {1,..., n} where a i {0,..., d i } and the assignment indicates a i groundings of R i are true. We will refer to (R i, a i ) as a counting assignment [14]. The algorithm iterates over all tuples of the form: (R 1, a 1 ),..., (R n, a n ), computes the weight of the tuple, and returns the tuple with the maximum weight as the MAP tuple. This lifted algorithm is clearly more efficient than its propositional counterpart. The search space over which the propositional algorithm operates is bounded by O(2 n i=1 di ) where n is the number of atoms in the MLN. On the other hand, the search space of the lifted algorithm is bounded by O( n i=1 (d i + 1)). Since the search space is bounded polynomially by the domain size of the logical variables, we have: Theorem 2. MAP inference in non-shared MLNs is domain liftable. Although this represents a significant improvement over propositional MAP algorithms, it turns out that we can further reduce the search space for a sub-class of non-shared MLNs, namely non-shared MLNs without self-joins. 2 We will present this result next. 3.2 MAP Inference over non-shared MLNs without Self-Joins We illustrate the main idea in our result using the following example. Example 2. Consider the MLN used in Example 1. Let w 1 = 4, w 2 = 5 and w 3 = 3. Assume that the domain 2 We say that a formula has no self-joins if a predicate symbol appears at most once in the formula.
5 Somdeb Sarkhel, Deepak Venugopal, Parag Singla, Vibhav Gogate Z axis X axis R(x) v S(y), -4; R(x), 5; S(y), Satisfied Clause Weight Y axis Figure 2: The total weight of satisfied clauses as a function of counting assignment of R and S. The plane is for illustration purpose only. of x, y is {A, B, C, D, E}. If we iterate through all possible counting assignments to R and S and plot the total weight of satisfied clauses as a function of counting assignment of R and S we get the plot in Figure 2. Figure 2 shows that the function in the plot has only four extreme points: (0, 0), (0, 5), (5, 0) and (5, 5). These extreme points correspond to all groundings of R and S as either being all true or all false. Since the MAP value can only lie on these extreme points, we only have to evaluate these extreme points for computing the MAP value. It turns out that the MAP tuple is (R, 0), (S, 0). We observe in the previous example that all the ground atoms of the predicate R (and the predicate S) have the same truth value. We will refer to this kind of assignment (i.e., all ground atoms having the same truth value) as a uniform assignment [1]. This observation, that the atoms have a uniform assignment in the MAP state, holds not only for this example but for any non-shared MLN without selfjoins, and we will prove this formally next. Lemma 1. The sum of weights of satisfied clauses for a non-shared MLN without self-join is a multilinear function on the counting assignment of its predicates. Proof. (Sketch) Consider a non-shared MLN M that contains m weighted clauses {(C i ; w i )} m i=1. Let V(C i) represent the set of all the atoms in the clause C i. Let V (+) (C i ) represent the set of atoms which appear as positive literals in C i. Let V ( ) (C i ) represent the set of atoms appearing as negative literals. Given an atom R, let (R, v R ) denote its counting assignment. It can be easily shown that the number of groundings of C i that are unsatisfied by the counting assignment is given by, R V (+) (C i) ( R v R ) R V ( ) (C i) v R where R represents the number of possible groundings of R. Clearly, the total number of possible groundings of C i is equal to R C i ( R ). Therefore, the sum of weights of satisfied clauses for M is given by, w i( ( R) C i R C i ( R v R ) R V (+) (C i) R V ( ) (C i) v R ) (4) Clearly eq. 4 is a multilinear function in v R since v R never appears more than once in the product term (if there are no self-joins in M). Lemma 2. Consider a multilinear function Φ(v) defined over a tuple of variables v = (v 1, v 2,, v n ). Let each v j take values from the set {0, 1, 2, vj }. Then, at least one of the solutions v to the optimization problem arg max Φ(v) is such that each v v j lies at the extremes i.e. vj = 0 or v j = j j. Proof. We will prove the theorem using induction over n, the number of variables over which the multilinear function is defined. Clearly, the theorem holds true for n = 1 since a linear function of one variable has its maxima at the extremes. Assume the theorem holds for any multilinear function defined over n 1 variables. Consider the function Φ(v) over the variables v = (v 1, v 2, v n ). By re-arranging terms, we can write: max v Φ(v) = max(max Φ(v)) v\v n v n Since Φ(v) is a multilinear function, it can be seen as a linear function of v n (holding other variables as constant). Hence, the inner expression on the right side is optimized at an extreme value of v n (v n = 0 or v n = vn ). Let Φ 0 (v \ v n ) and Φ vn (v \ v n ), respectively, be the two possible resulting functions by substituting the values of 0 and vn, for v n in Φ(v). In both the cases, we get a new function which is multilinear over n 1 variables. Using the induction hypothesis, its maxima will lie at the extreme values of v i, v 2, v n 1. Hence, one of the maxima of the original function φ(v) will lie at the extreme values of v 1, v 2, v n. Hence, proved. Note that above lemma states that at least one of the maxima of a multilinear function will lie at its extremes. It is still possible that there are other maxima which do not lie at the extremes (for instance, think of a constant function). As long as we are interested in finding one of them, above lemma can be put to use. Lemma 1 and Lemma 2 allow us to prove our second main result stated below. Theorem 3. For a non-shared MLN without self-joins, in at least one of the MAP solutions, all predicates have uniform assignments.
6 We can use Theorem 3 to formulate the MAP problem as a weighted Max-SAT problem as formalized in the following corollary. Corollary 1. The MAP inference in a non-shared MLN M that contains no self-joins can be converted to an equivalent propositional weighted Max-SAT problem with number of variables equal to the number of first order atoms in M. Proof. Given a non-shared MLN, M with m weighted clauses {(C i ; w i )} m i=1 that contains no self-joins, we first construct a weighted propositional knowledge base S. We create S = {(C i ; w i )}m i=1 with v propositional variables where every v k v corresponds to a distinct atom R vk in M. All atoms in M have a corresponding variable in v and vice versa. The assignment true to variable v k corresponds to the positive uniform assignment to R vk, i.e. - (R vk, Rvk ) and assigning false to variable v k corresponds to the negative uniform assignment to R vk, i.e. - (R vk, 0). C i is constructed by replacing each atom in C i by its corresponding variable in v. The weight of the clause is computed as w i = C i w i, where Ci is the number of possible groundings of C i. For uniform assignment, all groundings of each clause C i is either satisfied or none of them are satisfied. Since whenever C i is satisfied C i is also satisfied and as the weight of C i is C i w i the sum of weights of satisfied clauses for a complete assignment will be same in both M and S. As theorem 3 proves that the MAP solution consist of uniform assignments, it follows from equation 2 that the MAP inference in M is equivalent to solving the weighted Max-SAT problem over S. Hence the corollary follows. The result of corollary 1 allows us to use any weighted Max-SAT solver to compute the MAP solution of a nonshared MLN without self-joins. This observation also means that for such MLNs, the optimal solution is independent of the number of objects in the MLN which makes MAP inference especially efficient in these cases. 4 Extensions In this section, we propose heuristics to make our approach more practical. These heuristics can be considered as pruning techniques, which allow us to greatly reduce the size of the knowledge base. Moreover these heuristics can be applied to any arbitrary MLN. These can give us orders of magnitude of speedup. We propose to use these heuristics as a preprocessing step to simplify the MLN. 4.1 Unit Propagation Repeated use of unit propagation [4] is one of the key component of highly effective propositional satisfiability testing solvers. The idea in unit propagation is to resolve all clauses with unit clauses, and continue to do this until convergence, i.e., no further unit resolutions are possible. Although this heuristic is very effective for SAT solvers, for Max-SAT, it is not sound. However, this rule can be used for hard unit clauses. We can lift this rule in a straight forward manner, by resolving the hard unit clauses with other clauses. This heuristic in conjunction with the pure literal heuristic can greatly reduce the size of the MLN. 4.2 Pure Literal Elimination The pure literal elimination rule for SAT formulas [4] when lifted to MAP inference for MLNs, removes (i) Clauses guaranteed to be satisfied for all groundings; and (ii) Atoms guaranteed to be false for all groundings. The following proposition specifies the pure literal elimination rule for MLNs. Proposition 1. Given an MLN M, if a predicate S appears in k clauses C = {C i ; w i } k i=1, (i) if w i 0, 1 i k and S either always occurs as a positive literal or always occurs as a negative literal in M, every C i C can be removed from M; and (ii) if w i < 0, 1 < i k and S either always occurs as a positive literal or always occurs as a negative literal in M, then every occurrence of S can be removed from M. 5 Experiments For our experiments, we implemented two lifted MAP algorithms, (i) An anytime exact solver based on Integer Linear Programming (); and (ii) An anytime approximate solver based on WalkSAT architecture (). We implemented using a parallelized solver called Gurobi [8] and implemented using MaxWalkSAT [18], a randomized local-search algorithm. We compared both our algorithms with MaxWalkSAT which is the MAP inference algorithm implemented within two state-of-the-art MLN systems, Alchemy () and Tuffy ()[15]. Since both these systems produce approximate results, we implemented an exact MAP inference algorithm using Gurobi (). All three algorithms,, and work on the propositional search space, i.e. they ground the entire MLN before performing MAP inference. We used three MLNs to evaluate our system, (i) A Student MLN having four formulas: Teaches(teacher,course) Takes(student,course) JobOffers(student,company); Teaches(teacher,course); Takes(student,course); and JobOffers(student,company).
7 Somdeb Sarkhel, Deepak Venugopal, Parag Singla, Vibhav Gogate (ii) WebKB MLN [12] from the Alchemy web page, consisting of three predicates and six formulas. (iii) Citation Information-Extraction (IE) MLN [12] from the Alchemy web page, consisting of five predicates and fourteen formulas. In order to compare the performance and scalability of our algorithms, we ran two sets of experiments illustrated in Fig. 3 and Fig. 4. Fig. 3, plots the solution quality (total weight of false clauses) achieved by each algorithm for varying time-bounds. Fig. 4 plots the relative-gap between the optimal solution and the solution output by each algorithm, for varying domain-sizes. We describe the results of both our experiments below. All our experiments were run on a quad-core CentOS machine with 8GB RAM. 5.1 vs Time The results for the Student MLN are shown in Fig. 3 (a)-(c). We see that the lifted algorithms and are the best performing algorithms for all domain-sizes. At higher domain-sizes (100 and 500), the propositional solvers, and ran out of memory. The performance of was similar to for domain-size equal to 30. For domain-sizes of 100 and 500, gradually converges towards the optimal solution, whereas was able to exactly solve the problem in less than 10 seconds. The results for WebKB are shown in Fig. 3 (d)-(f). Again, we can see that the lifted algorithms and outperform the propositional algorithms and are much more scalable. For the larger domain-sizes (100 and 500), and run out of memory. For domain-size 30 and 100, the performance of both the lifted algorithms and is quite similar. The results for IE are shown in Fig. 3 (g)-(i). For domainsize 30, the performance of both the lifted algorithms L- and is quite similar. For the domain-sizes 100 and 500, was able to find the optimal solution while was far from optimal. 5.2 Accuracy vs Domain-Size Fig. 4 illustrates the variation in accuracy for each algorithm as the domain-size increases. Here, we gave each algorithm a fixed time-bound of 500 seconds and measured the relative-gap between the optimal solution (opt) and the best cost given by the algorithm (c) using opt c opt. We see that both and are quite accurate and scale to much larger domain-sizes. On the other hand, there is a noticeable drop in the accuracy of the propositional algorithms,, and as we increase the domainsize. For larger domain-sizes, the propositional algorithms run out of memory. In summary, our experiments show that our two lifted algorithms and are far more scalable and accurate than propositional approaches. Since the two approaches are fundamentally different, is a complete anytime solver while is an approximate solver, as expected they perform differently on the benchmarks, with being the superior approach. However, the main virtue of our approach is that we could use any off-theshelf solver that is purely propositional in nature to perform lifted inference. This allows us to scale to large domainsizes without implementing a new lifted solver. We believe that this abstraction greatly simplifies the development of lifted algorithms by benefitting from the advances made in propositional algorithms. 6 Summary and Future work In this paper, we proposed a general approach for lifting MAP inference in Markov Logic Networks (MLNs). We identified cases in which we can reduce lifted MAP inference to inference over an equivalent propositional theory such that the number of propositional variables is equal to the number of first order atoms in the MLN. We used this observation in a straight-forward manner: convert the MLN to an equivalent propositional theory and then apply any propositional algorithm to solve it. For our experiments, we used two propositional algorithms, a complete, anytime algorithm (Gurobi) based on Integer Linear Programming () and a local-search algorithm called MaxWalksat. Our experiments clearly demonstrate the scalability and promise of our approach. Directions for future work include: combining our approach with other lifted inference rules such as the power rule [7, 10]; identifying cases where our generalized theorem can be applied; applying the results in this paper to lifted MCMC approaches [19]; and using our approach for exploiting symmetries in probabilistic graphical models. Acknowledgements This research was partly funded by ARO MURI grant W911NF , by the AFRL under contract number FA C-0021 and by the DARPA Probabilistic Programming for Advanced Machine Learning Program under AFRL prime contract number FA C The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of DARPA, AFRL, ARO or the US government. References [1] Apsel, U.; and Brafman, R Exploiting Uniform Assignments in First-Order MPE. In Proceedings of the
8 1e+11 1e+10 1e+09 1e e (a) Student-30 (810K clauses) (b) Student-100 (100 million clauses) (c) Student-500 (62 Billion clauses) 1e+11 1e+10 1e+09 1e+08 1e (d) WebKB-30 (100K clauses) (e) WebKB-100 (100 million clauses) (f) WebKB-500 (65 Billion clauses) 1e+08 1e (g) IE-10 (200K clauses) (h) IE-100 (350 million clauses) (i) IE-500 (200 billion clauses) Figure 3: vs Time: of unsatisfied clauses (smaller is better) against time for benchmark MLNs for different domain sizes. Notation used to label each figure: MLN-domainsize(number of ground clauses in the MLN). The standard deviation is plotted as error bars. For (b),(c),(e),(f),(h) and (i), no results could be obtained for propositional algorithms since they ran out of memory Relative-gap Relative-gap Domain-Size (a) Student Domain-Size (b) WebKB Figure 4: Accuracy vs Domain-Size: The relative-gap i.e. optimal cost is plotted for varying domain-sizes (smaller is better). Every optimal algorithm was given 500 seconds of running-time. The propositional algorithms, and run out of memory for larger domain-sizes. Note: We could not run this experiment for IE since the propositional algorithms ran out of memory when the domain-size exceeded 10.
9 Somdeb Sarkhel, Deepak Venugopal, Parag Singla, Vibhav Gogate Twenty-Eighth Conference on Uncertainty in Artificial Intelligence, AUAI Press. [2] Braz, R.,; Amir, E; and Roth, D Lifted firstorder probabilistic inference. In Proceedings of the 19th International Joint Conference on Artificial Intelligence [3] Van den Broeck, G., On the completeness of first-order knowledge compilation for lifted probabilistic inference. In Advances in Neural Information Processing Systems 24, [4] Davis, M., and Putnam, H A computing procedure for quantification theory. In Journal of the ACM (JACM), 7(3) [5] Domingos, P., and Lowd, D Markov Logic: An Interface Layer for Artificial Intelligence. San Rafael, CA: Morgan & Claypool. [6] Getoor, L., and Taskar, B., eds Introduction to Statistical Relational Learning. MIT Press. [7] Gogate, V., and Domingos, P Probabilistic Theorem Proving. In Proceedings of the Twenty- Seventh Conference on Uncertainty in Artificial Intelligence, AUAI Press. [8] Gurobi Optimization, Inc Gurobi Optimizer Reference Manual. [9] Hoos, H. H., and Stützle, T Stochastic local search: Foundations & applications. Morgan Kaufmann. [10] Jha, A.; Gogate, V.; Meliou, A.; and Suciu, D Lifted Inference from the Other Side: The tractable Features. In Proceedings of the 24th Annual Conference on Neural Information Processing Systems (NIPS), [11] Kautz, H.; Selman, B.; and Jiang, Y A General Stochastic Approach to Solving Problems with Hard and Soft Constraints. In Gu, D.; Du, J.; and Pardalos, P., eds., The Satisfiability Problem: Theory and Applications. New York, NY: American Mathematical Society [12] Kok, S.; Sumner, M.; Richardson, M.; Singla, P.; Poon, H.; and Domingos, P The Alchemy System for Statistical Relational AI. Technical report, Department of Computer Science and Engineering, University of Washington, Seattle, WA. [13] Koller, D., and Friedman, N Probabilistic Graphical Models: Principles and Techniques. MIT Press. [14] Milch, B.; Zettlemoyer, L. S.; Kersting, K.; Haimes, M.; and Kaelbling, L. P Lifted Probabilistic Inference with Counting Formulas. In Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence [15] Niu, F.; Re, C.; Doan, A.; and Shavlik, J Tuffy: Scaling up statistical inference in markov logic networks using an rdbms. In Proceedings of the VLDB Endowment 4(6) [16] Pipatsrisawat, K.; and Darwiche, A Clone: Solving Weighted Max-SAT in a Reduced Search Space. In Proceedings of 20th Australian Joint Conference on Artificial Intelligence (AI07) [17] Poole, D First-Order Probabilistic Inference. In Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence, Acapulco, Mexico: Morgan Kaufmann. [18] Selman, B.; Kautz, H.; and Cohen, B Local Search Strategies for Satisfiability Testing. In Johnson, D. S., and Trick, M. A., eds., Cliques, Coloring, and Satisfiability: Second DIMACS Implementation Challenge. Washington, DC: American Mathematical Society [19] Venugopal, D.; and Gogate, V On Lifting the Gibbs Sampling Algorithm. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems (NIPS)
Lifted First-Order Belief Propagation
Lifted First-Order Belief Propagation Parag Singla Pedro Domingos Department of Computer Science and Engineering University of Washington Seattle, WA 98195-2350, U.S.A. {parag, pedrod}@cs.washington.edu
Cell Phone based Activity Detection using Markov Logic Network
Cell Phone based Activity Detection using Markov Logic Network Somdeb Sarkhel [email protected] 1 Introduction Mobile devices are becoming increasingly sophisticated and the latest generation of smart
npsolver A SAT Based Solver for Optimization Problems
npsolver A SAT Based Solver for Optimization Problems Norbert Manthey and Peter Steinke Knowledge Representation and Reasoning Group Technische Universität Dresden, 01062 Dresden, Germany [email protected]
Sudoku as a SAT Problem
Sudoku as a SAT Problem Inês Lynce IST/INESC-ID, Technical University of Lisbon, Portugal [email protected] Joël Ouaknine Oxford University Computing Laboratory, UK [email protected] Abstract Sudoku
Guessing Game: NP-Complete?
Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES 2. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple
Tensor Factorization for Multi-Relational Learning
Tensor Factorization for Multi-Relational Learning Maximilian Nickel 1 and Volker Tresp 2 1 Ludwig Maximilian University, Oettingenstr. 67, Munich, Germany [email protected] 2 Siemens AG, Corporate
Bounded-width QBF is PSPACE-complete
Bounded-width QBF is PSPACE-complete Albert Atserias 1 and Sergi Oliva 2 1 Universitat Politècnica de Catalunya Barcelona, Spain [email protected] 2 Universitat Politècnica de Catalunya Barcelona, Spain
NP-Completeness and Cook s Theorem
NP-Completeness and Cook s Theorem Lecture notes for COM3412 Logic and Computation 15th January 2002 1 NP decision problems The decision problem D L for a formal language L Σ is the computational task:
Offline sorting buffers on Line
Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: [email protected] 2 IBM India Research Lab, New Delhi. email: [email protected]
Generating models of a matched formula with a polynomial delay
Generating models of a matched formula with a polynomial delay Petr Savicky Institute of Computer Science, Academy of Sciences of Czech Republic, Pod Vodárenskou Věží 2, 182 07 Praha 8, Czech Republic
Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows
TECHNISCHE UNIVERSITEIT EINDHOVEN Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows Lloyd A. Fasting May 2014 Supervisors: dr. M. Firat dr.ir. M.A.A. Boon J. van Twist MSc. Contents
(LMCS, p. 317) V.1. First Order Logic. This is the most powerful, most expressive logic that we will examine.
(LMCS, p. 317) V.1 First Order Logic This is the most powerful, most expressive logic that we will examine. Our version of first-order logic will use the following symbols: variables connectives (,,,,
Big Data Science. Prof. Lise Getoor University of Maryland, College Park. http://www.cs.umd.edu/~getoor. October 17, 2013
Big Data Science Prof Lise Getoor University of Maryland, College Park October 17, 2013 http://wwwcsumdedu/~getoor BIG Data is not flat 2004-2013 lonnitaylor Data is multi-modal, multi-relational, spatio-temporal,
Scheduling Shop Scheduling. Tim Nieberg
Scheduling Shop Scheduling Tim Nieberg Shop models: General Introduction Remark: Consider non preemptive problems with regular objectives Notation Shop Problems: m machines, n jobs 1,..., n operations
ON FUNCTIONAL SYMBOL-FREE LOGIC PROGRAMS
PROCEEDINGS OF THE YEREVAN STATE UNIVERSITY Physical and Mathematical Sciences 2012 1 p. 43 48 ON FUNCTIONAL SYMBOL-FREE LOGIC PROGRAMS I nf or m at i cs L. A. HAYKAZYAN * Chair of Programming and Information
Introduction to Logic in Computer Science: Autumn 2006
Introduction to Logic in Computer Science: Autumn 2006 Ulle Endriss Institute for Logic, Language and Computation University of Amsterdam Ulle Endriss 1 Plan for Today Now that we have a basic understanding
µz An Efficient Engine for Fixed points with Constraints
µz An Efficient Engine for Fixed points with Constraints Kryštof Hoder, Nikolaj Bjørner, and Leonardo de Moura Manchester University and Microsoft Research Abstract. The µz tool is a scalable, efficient
Learning is a very general term denoting the way in which agents:
What is learning? Learning is a very general term denoting the way in which agents: Acquire and organize knowledge (by building, modifying and organizing internal representations of some external reality);
Algebra 1 Course Information
Course Information Course Description: Students will study patterns, relations, and functions, and focus on the use of mathematical models to understand and analyze quantitative relationships. Through
2.1 Complexity Classes
15-859(M): Randomized Algorithms Lecturer: Shuchi Chawla Topic: Complexity classes, Identity checking Date: September 15, 2004 Scribe: Andrew Gilpin 2.1 Complexity Classes In this lecture we will look
Integrating Benders decomposition within Constraint Programming
Integrating Benders decomposition within Constraint Programming Hadrien Cambazard, Narendra Jussien email: {hcambaza,jussien}@emn.fr École des Mines de Nantes, LINA CNRS FRE 2729 4 rue Alfred Kastler BP
OHJ-2306 Introduction to Theoretical Computer Science, Fall 2012 8.11.2012
276 The P vs. NP problem is a major unsolved problem in computer science It is one of the seven Millennium Prize Problems selected by the Clay Mathematics Institute to carry a $ 1,000,000 prize for the
InvGen: An Efficient Invariant Generator
InvGen: An Efficient Invariant Generator Ashutosh Gupta and Andrey Rybalchenko Max Planck Institute for Software Systems (MPI-SWS) Abstract. In this paper we present InvGen, an automatic linear arithmetic
Computational Methods for Database Repair by Signed Formulae
Computational Methods for Database Repair by Signed Formulae Ofer Arieli ([email protected]) Department of Computer Science, The Academic College of Tel-Aviv, 4 Antokolski street, Tel-Aviv 61161, Israel.
Extending Semantic Resolution via Automated Model Building: applications
Extending Semantic Resolution via Automated Model Building: applications Ricardo Caferra Nicolas Peltier LIFIA-IMAG L1F1A-IMAG 46, Avenue Felix Viallet 46, Avenue Felix Viallei 38031 Grenoble Cedex FRANCE
Indiana State Core Curriculum Standards updated 2009 Algebra I
Indiana State Core Curriculum Standards updated 2009 Algebra I Strand Description Boardworks High School Algebra presentations Operations With Real Numbers Linear Equations and A1.1 Students simplify and
Efficient Algorithms for Masking and Finding Quasi-Identifiers
Efficient Algorithms for Masking and Finding Quasi-Identifiers Rajeev Motwani Stanford University [email protected] Ying Xu Stanford University [email protected] ABSTRACT A quasi-identifier refers
Mathematics for Computer Science/Software Engineering. Notes for the course MSM1F3 Dr. R. A. Wilson
Mathematics for Computer Science/Software Engineering Notes for the course MSM1F3 Dr. R. A. Wilson October 1996 Chapter 1 Logic Lecture no. 1. We introduce the concept of a proposition, which is a statement
LAKE ELSINORE UNIFIED SCHOOL DISTRICT
LAKE ELSINORE UNIFIED SCHOOL DISTRICT Title: PLATO Algebra 1-Semester 2 Grade Level: 10-12 Department: Mathematics Credit: 5 Prerequisite: Letter grade of F and/or N/C in Algebra 1, Semester 2 Course Description:
CS510 Software Engineering
CS510 Software Engineering Propositional Logic Asst. Prof. Mathias Payer Department of Computer Science Purdue University TA: Scott A. Carr Slides inspired by Xiangyu Zhang http://nebelwelt.net/teaching/15-cs510-se
Nan Kong, Andrew J. Schaefer. Department of Industrial Engineering, Univeristy of Pittsburgh, PA 15261, USA
A Factor 1 2 Approximation Algorithm for Two-Stage Stochastic Matching Problems Nan Kong, Andrew J. Schaefer Department of Industrial Engineering, Univeristy of Pittsburgh, PA 15261, USA Abstract We introduce
CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA
We Can Early Learning Curriculum PreK Grades 8 12 INSIDE ALGEBRA, GRADES 8 12 CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA April 2016 www.voyagersopris.com Mathematical
The Optimality of Naive Bayes
The Optimality of Naive Bayes Harry Zhang Faculty of Computer Science University of New Brunswick Fredericton, New Brunswick, Canada email: hzhang@unbca E3B 5A3 Abstract Naive Bayes is one of the most
Lecture 7: NP-Complete Problems
IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 7: NP-Complete Problems David Mix Barrington and Alexis Maciel July 25, 2000 1. Circuit
Complexity Theory. IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar
Complexity Theory IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar Outline Goals Computation of Problems Concepts and Definitions Complexity Classes and Problems Polynomial Time Reductions Examples
WHAT ARE MATHEMATICAL PROOFS AND WHY THEY ARE IMPORTANT?
WHAT ARE MATHEMATICAL PROOFS AND WHY THEY ARE IMPORTANT? introduction Many students seem to have trouble with the notion of a mathematical proof. People that come to a course like Math 216, who certainly
Course Outline Department of Computing Science Faculty of Science. COMP 3710-3 Applied Artificial Intelligence (3,1,0) Fall 2015
Course Outline Department of Computing Science Faculty of Science COMP 710 - Applied Artificial Intelligence (,1,0) Fall 2015 Instructor: Office: Phone/Voice Mail: E-Mail: Course Description : Students
Integer Factorization using the Quadratic Sieve
Integer Factorization using the Quadratic Sieve Chad Seibert* Division of Science and Mathematics University of Minnesota, Morris Morris, MN 56567 [email protected] March 16, 2011 Abstract We give
Chapter 6: Episode discovery process
Chapter 6: Episode discovery process Algorithmic Methods of Data Mining, Fall 2005, Chapter 6: Episode discovery process 1 6. Episode discovery process The knowledge discovery process KDD process of analyzing
Notes from Week 1: Algorithms for sequential prediction
CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking
Globally Optimal Crowdsourcing Quality Management
Globally Optimal Crowdsourcing Quality Management Akash Das Sarma Stanford University [email protected] Aditya G. Parameswaran University of Illinois (UIUC) [email protected] Jennifer Widom Stanford
Course: Model, Learning, and Inference: Lecture 5
Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 [email protected] Abstract Probability distributions on structured representation.
3. Mathematical Induction
3. MATHEMATICAL INDUCTION 83 3. Mathematical Induction 3.1. First Principle of Mathematical Induction. Let P (n) be a predicate with domain of discourse (over) the natural numbers N = {0, 1,,...}. If (1)
Learning Sum-Product Networks with Direct and Indirect Variable Interactions
Learning Sum-Product Networks with Direct and Indirect Variable Interactions Amirmohammad Rooshenas Daniel Lowd Department of Computer and Information Science, University of Oregon [email protected]
Handout #1: Mathematical Reasoning
Math 101 Rumbos Spring 2010 1 Handout #1: Mathematical Reasoning 1 Propositional Logic A proposition is a mathematical statement that it is either true or false; that is, a statement whose certainty or
Why? A central concept in Computer Science. Algorithms are ubiquitous.
Analysis of Algorithms: A Brief Introduction Why? A central concept in Computer Science. Algorithms are ubiquitous. Using the Internet (sending email, transferring files, use of search engines, online
An Empirical Study of Two MIS Algorithms
An Empirical Study of Two MIS Algorithms Email: Tushar Bisht and Kishore Kothapalli International Institute of Information Technology, Hyderabad Hyderabad, Andhra Pradesh, India 32. [email protected],
On the effect of data set size on bias and variance in classification learning
On the effect of data set size on bias and variance in classification learning Abstract Damien Brain Geoffrey I Webb School of Computing and Mathematics Deakin University Geelong Vic 3217 With the advent
MapReduce and Distributed Data Analysis. Sergei Vassilvitskii Google Research
MapReduce and Distributed Data Analysis Google Research 1 Dealing With Massive Data 2 2 Dealing With Massive Data Polynomial Memory Sublinear RAM Sketches External Memory Property Testing 3 3 Dealing With
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Understanding Basic Calculus
Understanding Basic Calculus S.K. Chung Dedicated to all the people who have helped me in my life. i Preface This book is a revised and expanded version of the lecture notes for Basic Calculus and other
How To Know If A Domain Is Unique In An Octempo (Euclidean) Or Not (Ecl)
Subsets of Euclidean domains possessing a unique division algorithm Andrew D. Lewis 2009/03/16 Abstract Subsets of a Euclidean domain are characterised with the following objectives: (1) ensuring uniqueness
CHAPTER 7 GENERAL PROOF SYSTEMS
CHAPTER 7 GENERAL PROOF SYSTEMS 1 Introduction Proof systems are built to prove statements. They can be thought as an inference machine with special statements, called provable statements, or sometimes
Approximating the Partition Function by Deleting and then Correcting for Model Edges
Approximating the Partition Function by Deleting and then Correcting for Model Edges Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles Los Angeles, CA 995
HYBRID GENETIC ALGORITHMS FOR SCHEDULING ADVERTISEMENTS ON A WEB PAGE
HYBRID GENETIC ALGORITHMS FOR SCHEDULING ADVERTISEMENTS ON A WEB PAGE Subodha Kumar University of Washington [email protected] Varghese S. Jacob University of Texas at Dallas [email protected]
Factoring & Primality
Factoring & Primality Lecturer: Dimitris Papadopoulos In this lecture we will discuss the problem of integer factorization and primality testing, two problems that have been the focus of a great amount
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
Optimizing Description Logic Subsumption
Topics in Knowledge Representation and Reasoning Optimizing Description Logic Subsumption Maryam Fazel-Zarandi Company Department of Computer Science University of Toronto Outline Introduction Optimization
Finding the M Most Probable Configurations Using Loopy Belief Propagation
Finding the M Most Probable Configurations Using Loopy Belief Propagation Chen Yanover and Yair Weiss School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, Israel
Neovision2 Performance Evaluation Protocol
Neovision2 Performance Evaluation Protocol Version 3.0 4/16/2012 Public Release Prepared by Rajmadhan Ekambaram [email protected] Dmitry Goldgof, Ph.D. [email protected] Rangachar Kasturi, Ph.D.
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
ARTICLE IN PRESS. European Journal of Operational Research xxx (2004) xxx xxx. Discrete Optimization. Nan Kong, Andrew J.
A factor 1 European Journal of Operational Research xxx (00) xxx xxx Discrete Optimization approximation algorithm for two-stage stochastic matching problems Nan Kong, Andrew J. Schaefer * Department of
Cloud Computing is NP-Complete
Working Paper, February 2, 20 Joe Weinman Permalink: http://www.joeweinman.com/resources/joe_weinman_cloud_computing_is_np-complete.pdf Abstract Cloud computing is a rapidly emerging paradigm for computing,
Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm.
Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm. We begin by defining the ring of polynomials with coefficients in a ring R. After some preliminary results, we specialize
Chapter 2: Binomial Methods and the Black-Scholes Formula
Chapter 2: Binomial Methods and the Black-Scholes Formula 2.1 Binomial Trees We consider a financial market consisting of a bond B t = B(t), a stock S t = S(t), and a call-option C t = C(t), where the
Message-passing sequential detection of multiple change points in networks
Message-passing sequential detection of multiple change points in networks Long Nguyen, Arash Amini Ram Rajagopal University of Michigan Stanford University ISIT, Boston, July 2012 Nguyen/Amini/Rajagopal
Predicate Logic. Example: All men are mortal. Socrates is a man. Socrates is mortal.
Predicate Logic Example: All men are mortal. Socrates is a man. Socrates is mortal. Note: We need logic laws that work for statements involving quantities like some and all. In English, the predicate is
Monitoring Metric First-order Temporal Properties
Monitoring Metric First-order Temporal Properties DAVID BASIN, FELIX KLAEDTKE, SAMUEL MÜLLER, and EUGEN ZĂLINESCU, ETH Zurich Runtime monitoring is a general approach to verifying system properties at
Fairness in Routing and Load Balancing
Fairness in Routing and Load Balancing Jon Kleinberg Yuval Rabani Éva Tardos Abstract We consider the issue of network routing subject to explicit fairness conditions. The optimization of fairness criteria
Summary Last Lecture. Automated Reasoning. Outline of the Lecture. Definition sequent calculus. Theorem (Normalisation and Strong Normalisation)
Summary Summary Last Lecture sequent calculus Automated Reasoning Georg Moser Institute of Computer Science @ UIBK Winter 013 (Normalisation and Strong Normalisation) let Π be a proof in minimal logic
An example of a computable
An example of a computable absolutely normal number Verónica Becher Santiago Figueira Abstract The first example of an absolutely normal number was given by Sierpinski in 96, twenty years before the concept
Midterm Practice Problems
6.042/8.062J Mathematics for Computer Science October 2, 200 Tom Leighton, Marten van Dijk, and Brooke Cowan Midterm Practice Problems Problem. [0 points] In problem set you showed that the nand operator
Predicate Logic. For example, consider the following argument:
Predicate Logic The analysis of compound statements covers key aspects of human reasoning but does not capture many important, and common, instances of reasoning that are also logically valid. For example,
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
Online and Offline Selling in Limit Order Markets
Online and Offline Selling in Limit Order Markets Kevin L. Chang 1 and Aaron Johnson 2 1 Yahoo Inc. [email protected] 2 Yale University [email protected] Abstract. Completely automated electronic
Global Multiprocessor Real-Time Scheduling as a Constraint Satisfaction Problem
Global Multiprocessor Real-Time Scheduling as a Constraint Satisfaction Problem Liliana Cucu-Grosean & Olivier Buffet INRIA Nancy Grand-Est 615 rue du Jardin Botanique 54600 Villers-lès-Nancy, France [email protected]
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
This asserts two sets are equal iff they have the same elements, that is, a set is determined by its elements.
3. Axioms of Set theory Before presenting the axioms of set theory, we first make a few basic comments about the relevant first order logic. We will give a somewhat more detailed discussion later, but
Performance Test of Local Search Algorithms Using New Types of
Performance Test of Local Search Algorithms Using New Types of Random CNF Formulas Byungki Cha and Kazuo Iwama Department of Computer Science and Communication Engineering Kyushu University, Fukuoka 812,
The Relative Worst Order Ratio for On-Line Algorithms
The Relative Worst Order Ratio for On-Line Algorithms Joan Boyar 1 and Lene M. Favrholdt 2 1 Department of Mathematics and Computer Science, University of Southern Denmark, Odense, Denmark, [email protected]
South Carolina College- and Career-Ready (SCCCR) Algebra 1
South Carolina College- and Career-Ready (SCCCR) Algebra 1 South Carolina College- and Career-Ready Mathematical Process Standards The South Carolina College- and Career-Ready (SCCCR) Mathematical Process
MATH10040 Chapter 2: Prime and relatively prime numbers
MATH10040 Chapter 2: Prime and relatively prime numbers Recall the basic definition: 1. Prime numbers Definition 1.1. Recall that a positive integer is said to be prime if it has precisely two positive
Repair Checking in Inconsistent Databases: Algorithms and Complexity
Repair Checking in Inconsistent Databases: Algorithms and Complexity Foto Afrati 1 Phokion G. Kolaitis 2 1 National Technical University of Athens 2 UC Santa Cruz and IBM Almaden Research Center Oxford,
Why Study NP- hardness. NP Hardness/Completeness Overview. P and NP. Scaling 9/3/13. Ron Parr CPS 570. NP hardness is not an AI topic
Why Study NP- hardness NP Hardness/Completeness Overview Ron Parr CPS 570 NP hardness is not an AI topic It s important for all computer scienhsts Understanding it will deepen your understanding of AI
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
