Graphical Models and Message passing

Size: px
Start display at page:

Download "Graphical Models and Message passing"

Transcription

1 Graphical Models and Message passing Yunshu Liu ASPITRG Research Group

2 References: [1]. Steffen Lauritzen, Graphical Models, Oxford University Press, 1996 [2]. Michael I. Jordan, Graphical models, Statistical Science, Vol.19, No. 1. Feb., 2004 [3]. Christopher M. Bishop, Pattern Recognition and Machine Learning, Springer-Verlag New York, Inc [4]. Daphne Koller and Nir Friedman, Probabilistic Graphical Models - Principles and Techniques, The MIT Press, 2009 [5].Kevin P. Murphy, Machine Learning - A Probabilistic Perspective, The MIT Press, 2012

3 Outline Preliminaries on Graphical Models Directed graphical model Undirected graphical model Directed graph and Undirected graph Message passing and sum-product algorithm Factor graphs Sum-product algorithms Junction tree algorithm

4 Outline Preliminaries on Graphical Models Message passing and sum-product algorithm

5 Preliminaries on Graphical Models Motivation: Curse of dimensionality Matroid enumeration Polyhedron computation Entropic Region Machine Learning: computing likelihoods, marginal probabilities Graphical Models: Help capturing complex dependencies among random variables, building large-scale statistical models and designing efficient algorithms for Inference.

6 Preliminaries on Graphical Models Definition of Graphical Models: A graphical model is a probabilistic model for which a graph denotes the conditional dependence structure between random variables. Example: Suppose MIT and Stanford accepted undergraduate students only based on GPA MIT : Accepted by MIT Stanford : Accepted by Stanford GPA MIT Stanford Given Alice s GPA as GP A Alice, P(MIT Stanford, GP A Alice )=P(MIT GP A Alice ) We say MIT is conditionally independent of Stanford given GP A Alice Sometimes use symbol (MIT Stanford GP A Alice )

7 Bayesian networks: directed graphical model Bayesian networks A Bayesian network consists of a collection of probability distributions P over x = {x 1,, x K } that factorize over a directed acyclic graph(dag) in the following way: p(x) = p(x 1,, x K ) = k K p(x k pa k ) where pa k is the direct parents nodes of x k. Alias of Bayesian networks: probabilistic directed graphical model: via directed acyclic graph(dag) belief networks causal networks: directed arrows represent causal realtions

8 Bayesian networks: directed graphical model Examples of Bayesian networks Consider an arbitrary joint distribution p(x) = p(x 1, x 2, x 3 ) over three variables, we can write: p(x 1, x 2, x 3 ) = p(x 3 x 1, x 2 )p(x 1, x 2 ) = p(x 3 x 1, x 2 )p(x 2 x 1 )p(x 1 ) which can be expressed in the following directed graph: x 1 x 3 x 2

9 Bayesian networks: directed graphical model Examples Similarly, if we change the order of x 1, x 2 and x 3 (same as consider all permutations of them), we can express p(x 1, x 2, x 3 ) in five other different ways, for example: p(x 1, x 2, x 3 ) = p(x 1 x 2, x 3 )p(x 2, x 3 ) = p(x 1 x 2, x 3 )p(x 2 x 3 )p(x 3 ) which corresponding to the following directed graph: x 1 x 3 x 2

10 Bayesian networks: directed graphical model Examples Recall the previous example about how MIT and Stanford accept undergraduate students, if we assign x 1 to GPA, x 2 to accepted by MIT and x 3 to accepted by Stanford, then since p(x 3 x 1, x 2 ) = p(x 3 x 1 ) we have p(x 1, x 2, x 3 ) = p(x 3 x 1, x 2 )p(x 2 x 1 )p(x 1 ) = p(x 3 x 1 )p(x 2 x 1 )p(x 1 ) which corresponding to the following directed graph: x 1 GPA x 3 Stanford x 2 MIT

11 ,x 4 } is not a clique because of the missing link from x 1 to x 4. re Markov define the random factors infields: the decomposition undirected of graphical the joint distribution model the variables In the undirected in the cliques. case, theinprobability fact, we can distribution consider factorizes functions ques, according without loss to functions of generality, definedbecause on the clique otherof cliques the graph. must be l cliques. A clique Thus, is aif subset {x 1,x 2 of,xnodes 3 } is ainmaximal a graph such clique that and there we define exist a on over linkthis between clique, allthen pairsincluding of nodes inanother subset. factor defined over a iablesawould maximal be redundant. clique is a clique such that it is not possible to a clique by C and the set of variables in that clique by x C. Then include any other nodes from the graph in the set without it ceasing to be a clique. ted graph showing a clique (outlined in al clique (outlined in blue). x 1 Example of cliques: x 2 {x 1,x 2 },{x 2,x 3 },{x 3,x 4 },{x 2,x 4 },{x 1,x 3 }, {x 1,x 2,x 3 }, {x 2,x 3,x 4 } Maximal cliques: {x 1,x 2,x 3 }, {x 2,x 3,x 4 } x 3 x 4

12 Markov random fields: undirected graphical model Markov random fields: Definition Denote C as a clique, x C the set of variables in clique C and ψ C (x C ) a nonnegative potential function associated with clique C. Then a Markov random field is a collection of distributions that factorize as a product of potential functions ψ C (x C ) over the maximal cliques of the graph: p(x) = 1 ψ Z C (x C ) where normalization constant Z = x called the partition function. C C ψ C(x C ) sometimes

13 Markov random fields: undirected graphical model Factorization of undirected graphs Question: how to write the joint distribution for this undirected graph? x 1 x 3 x 2 (2 3 1) hold Answer: p(x) = 1 Z ψ 12(x 1, x 2 )ψ 13 (x 1, x 3 ) where ψ 12 (x 1, x 2 ) and ψ 13 (x 1, x 3 ) are the potential functions and Z is the partition function that make sure p(x) satisfy the conditions to be a probability distribution.

14 Markov random fields: undirected graphical model Markov property Given an undirected graph G = (V,E), a set of random variables X = (X a ) a V indexed by V, we have the following Markov properties: Pairwise Markov property: Any two non-adjacent variables are conditionally independent given all other variables: X a X b X V \{a,b} if {a, b} E Local Markov property: A variable is conditionally independent of all other variables given its neighbors: X a X V \{nb(a) a} X nb(a) where nb(a) is the neighbors of node a. Global Markov property: Any two subsets of variables are conditionally independent given a separating subset: X A X B X S, where every path from a node in A to a node in B passes through S(means when we remove all the nodes in S, there are no paths connecting any nodes in A to any nodes in B).

15 Markov random fields: undirected graphical model G. The dotted red node X 8 is independent of all other clude Examples its parents of Markov (blue), properties children (green) and co-parents GM. Pairwise The redmarkov nodeproperty: X ( ), ( ) 8 is independent of the other black Local Markov property: ( ), ( ) Global Markov property: ( ), ( ) (b)

16 Markov random fields: undirected graphical model Relationship between different Markov properties and factorization property ( F ): Factorization property; ( G ): Global Markov property; ( L ): Local Markov property; ( P ):Pairwise Markov property if assuming strictly positive p( ) (F) (G) (L) (P) (P) (F) which give us the Hammersley-Clifford theorem.

17 Markov random fields: undirected graphical model The Hammersley-Clifford theorem(see Koller and Friedman 2009, p131 for proof) Consider graph G, for strictly positive p( ), the following Markov property and Factorization property are equivalent: Markov property: Any two subsets of variables are conditionally independent given a separating subset (X A, X B X S ) where every path from a node in A to a node in B passes through S. Factorization property: The distribution p factorizes according to G if it can be expressed as a product of potential functions over maximal cliques.

18 Markov random fields: undirected graphical model Example: Gaussian Markov random fields A multivariate normal distribution forms a Markov random field w.r.t. a graph G = (V, E) if the missing edges correspond to zeros on the concentration matrix(the inverse convariance matrix) Consider x = {x 1,, x K } N (0, Σ) with Σ regular and K = Σ 1. The concentration matrix of the conditional distribution of (x i, x j ) given x \{i,j} is ( ) kii k K {i,j} = ij k ji k jj Hence x i x j x \ij k ij = 0 {i, j} E

19 Markov random fields: undirected graphical model Example: Gaussian Markov random fields The joint distribution of gaussian markov random fields can be factorizes as: log p(x) = const 1 2 i k ii x 2 i {i,j} E k ij x i x j The zero entries in K are called structural zeros since they represent the absent edges in the undirected graph.

20 X 11 X 12 X 13 X 14 X 15 Directed graph and Undirected graph X 16 X 17 X 18 X 19 X 20 X 16 X 17 X 18 X 19 X 20 (a) X 11 X 12 X 13 X 14 X 15 Moral graph: from directed graph to undirected graph Figure 19.1 Moralization: marrying the parents, so to speak, force the parents of a node to form a completed sub-graph. (a) A 2d lattice represented as a DAG. The dotted red node X 8 is independent of all other nodes (black) given its Markov blanket, which include its parents (blue), children (green) and co-parents (orange). (b) The same model represented as a UGM. The red node X 8 is independent of the other black nodes given its neighbors (blue nodes). (b) (a) (b) Figure 19.2 (a) A DGM. (b) Its moralized version, represented as a UGM. Motivation: Why not simply convert directed graph to undirected graph? Works fine for Markov chain, not work for when there are nodes that have more than one parents(previous example). The set of nodes that renders a node t conditionally independent of all the other nodes in the graph is called t s Markov blanket; we will denote this by mb(t). Formally, the Markov blanket satisfies the following property: t V\cl(t) mb(t) (19.1) where cl(t) mb(t) {t} is the closure of node t. One can show that, in a UGM, a node s

21 Directed graph and Undirected graph Markov blanket In directed graph, a Markov blanket of a given node x i is the set of nodes comprising the parents, the children and the co-parents(other parents of its children) of x i. In undirected graph, a Markov blanket of a given node x i is the set of 662its neighboring nodes. Chapter 19. Undirected graphical models (Markov random fields) X 1 X 2 X 3 X 4 X 5 X 1 X 2 X 3 X 4 X 5 X 6 X 7 X 8 X 9 X 10 X 6 X 7 X 8 X 9 X 10 X 11 X 12 X 13 X 14 X 15 X 11 X 12 X 13 X 14 X 15 X 16 X 17 X 18 X 19 X 20 (a) X 16 X 17 X 18 X 19 X 20 (b) Figure 19.1 (a) A 2d lattice represented as a DAG. The dotted red node X 8 is independent of all other nodes (black) given its Markov blanket, which include its parents (blue), children (green) and co-parents (orange). (b) The same model represented as a UGM. The red node X 8 is independent of the other black nodes given its neighbors (blue nodes).

22 Directed graph and Undirected graph Markov blanket In directed graph Let x = {x 1,, x K } p(x i x \i ) = p(x 1,, x K ) x i p(x 1,, x K ) k = p(x k pa k ) x i k p(x k pa k ) = f (x i, mb(x i )) where pa k is the parents nodes of x k, mb(x i ) is the markov blanket of node x i. Example: p(x 8 X \8 ) = p(x 8 X 3,X 7 ) p(x 9 X 4,X 8 )p(x 13 X 8,X 12 ) X 8 p(x 8 X 3,X 7 )p(x 9 X 4,X 8 )p(x 13 X 8,X 12 )

23 Directed graph and Undirected graph Markov blanket In undirected graph Let x = {x 1,, x K }, p(x i x \i ) = p(x i mb(x i )) where mb(x i ) is the markov blanket of node x i. Example: p(x 8 X \8 ) = p(x 8 X 3, X 7, X 9, X 13 ) Notes The Markov blanket of a directed graph is the Markov blanket(neighbours) of its moral graph.

24 Directed graph and Undirected graph Directed graph VS. Undirected graph All distributions Directed graph Chordal Undirected graph x 1 x 1 x 1 x 3 x 2 x 4 x 3 x2 x 3 x 2 What is Chordal?

25 Directed graph and Undirected graph What is Chordal? Chordal or decomposable: If we collapse together all the variables in each maximal clique to make the mega-variables, the resulting graph will be a tree. Why Chordal graphs are the intersection of directed graph and undirected graph? Hint: Moral graph

26 Directed graph and Undirected graph Directed Graph:Dropping edges x 1 (2 3 1) x 3 x 2 x 1 x 1 (1 3 2) x 3 x 2 x 3 x 2 x 1 (1 2 ) But not (1 2 3) x 3 x 2

27 Directed graph and Undirected graph Undirected Graphs: remove arrows x 1 x 3 x 2 x 1 x 1 x 1 x 3 x 2 x 3 x 2 x 3 x 2 (2 3 1) hold x 1 x 3 x 2

28 Directed graph and Undirected graph Directed graph and Undirected graph Example of 4 atom distribution: x x 4 x 3 x 2 x 1 p(x) When I(x 3 ; x 4 )= a b-a c-a 1+a-b-c αβ α(1 β) (1 α)β (1 α)(1 β) Satisfy following Conditional independence relations: (1, 2 3), (1, 2 4) (1, 2 34), (1 34), (2 34) (1 234), (2 134), (3 124), (4 123) (3, 4 )

29 Outline Preliminaries on Graphical Models Message passing and sum-product algorithm

30 Factor graphs Common thing about directed graphical model and undirected graphical model Both allow global function to be expressed as product of factors over subsets of variables. Factor graphs: Definition A factor graph is a bipartite(bigraph) that expresses the structure of the factorization (1). p(x) = s f s (x s ) (1) where x s denotes a subset of the variables x. In a factor graph, the random variables usually are round nodes and the factors are square nodes. There is an edge between the factor node f s (x s ) and the variable node x k if and only if x i x s.

31 Factor graphs Example of factor graph: Factor graph: connection only between variable nodes and factor nodes. Variable nodes X 1 X 2 X 3 X 4 X 5 Factor nodes F a F b F c F d p(x) =f a (x 1,x 3 )f b (x 3,x 4 )f c (x 2,x 4,x 5 )f d (x 1,x 3 ) Note: The same subset of variables x s can be repeated multiple times, for example: x s = {x 1, x 3 } in the above factor graph.

32 Factor graphs From directed graph to factor graph: x 1 x 2 x 1 x 2 x 1 x 2 f f c x 3 f x a 3 x3 (a) (b) (c) f b (a ): An directed graph with factorization p(x 1 )p(x 2 )p(x 3 x 1, x 2 ). (b ): A factor graph with factor f (x 1, x 2, x 3 ) = p(x 1 )p(x 2 )p(x 3 x 1, x 2 ). (c ): A different factor graph with factors f a (x 1 )f b (x 2 )f c (x 1, x 2, x 3 ) such that f a (x 1 ) = p(x 1 ), f b (x 2 ) = p(x 2 ) and f c (x 1, x 2, x 3 ) = p(x 3 x 1, x 2 ).

33 Factor graphs From undirected graph to factor graph: difference between 1 Z C ψ C(x C ) and x f s(x s ) x 1 x 1 x 2 x 3 x 1 x 2 x 3 x 2 x 3 f f a f b (a) (b) (c) (a ): An undirected graph with a single clique potential ψ(x 1, x 2, x 3 ). (b ): A factor graph with factor f (x 1, x 2, x 3 ) = ψ(x 1, x 2, x 3 ). (c ): A different factor graph with factors f a (x 1, x 2, x 3 )f b (x 1, x 2 ) = ψ(x 1, x 2, x 3 ) Note: When clique is large, further factorize is usually very useful.

34 Factor graphs Why introduce Factor graphs? Both directed graph and undirected graph can be converted to factor graphs. Factor graphs allows us to be more explicit about the details of the factorization. When connect factors to exponential families, the product of factors becomes sum of parameters under single exp. Multiple different factor graphs can represent the same directed or undirected graph, we can find factor graphs that are better for our inference.

35 Sum-product algorithms Compute marginal probability p(x i ) p(x i ) = x \xi p(x) What if p(x) = 1 Z C ψ C(x C ) or x f s(x s )? Can we design more efficient algorithms by exploring the conditional independence structure of the model?

36 Sum-product algorithms Yes, we can. The key idea of Sum-product algorithm xy + xz = x(y + z) + x + y x z y z Expression tree of xy + xz and x(y+z), where the leaf nodes represent variables and internal nodes represent arithmetic operators(addition, multiplication, negation, etc.).

37 Sum-product algorithms Example of computing p(x 1 ) = x \x1 p(x) on an undirected chain p(x) = 1 Z ψ 12(x 1,x 2 )ψ 23 (x 2,x 3 )ψ 34 (x 3,x 4 ) x 1 x 2 x 3 x 4 m 2 (x 1 ) m 3 (x 2 ) m 4 (x 3 ) p(x 1 ) = 1 Z = 1 Z = 1 Z = 1 Z x 2 x 3 ψ 12 (x 1, x 2 )ψ 23 (x 2, x 3 )ψ 34 (x 3, x 4 ) x 4 ψ 12 (x 1, x 2 ) ψ 23 (x 2, x 3 ) ψ 34 (x 3, x 4 ) x 2 x 3 x 4 ψ 12 (x 1, x 2 ) ψ 23 (x 2, x 3 )m 4 (x 3 ) x 2 x 3 ψ 12 (x 1, x 2 )m 3 (x 2 ) = 1 Z m 2(x 1 ) x 2

38 Sum-product algorithms Example of computing p(x 1 ) = x \x1 p(x) Factor graph: connection only between variable nodes and factor nodes. Variable nodes X 1 X 2 X 3 X 4 X 5 Factor nodes F a F b F c F d p(x) =f a (x 1,x 3 )f b (x 3,x 4 )f c (x 2,x 4,x 5 )f d (x 1,x 3 ) p(x 1 ) = x \x1 p(x) = x 2 p(x) x 3 x 4 x 5 = x 2,x 3 f a (x 1, x 3 )f d (x 1, x 3 ) x 4 f b (x 3, x 4 ) x 5 f c (x 2, x 4, x 5 )

39 Sum-product algorithms Sum-product algorithm The sum-product algorithm involves using a message passing scheme to explore the conditional independence and use these properties to change the order of sum and product.

40 Sum-product algorithms Message passing on a factor graph: Variable x to factor f : m x f = m fk x(x) k nb(x)\f f 1 m f1 x x m x f f m fk x f K nb(x) denote the neighboring factor nodes of x If nb(x)\f = then m x f = 1.

41 Sum-product algorithms Message passing on a factor graph: Factor f to variable x: m f x = f (x, x 1,, x M ) x 1, x M m nb(f )\x m xm f (x m ) x 1 m x1 f f m f x x x M m xm f nb(f ) denote the neighboring variable nodes of f If nb(f )\x = then m f x = f (x).

42 Sum-product algorithms Message passing on a tree structured factor graph x 3 f b x 2 f a f c x 1 x 4

43 Sum-product algorithms Message passing on a tree structured factor graph Down - up: m x1 f a (x 1 ) = 1 m fa x 2 (x 2 ) = x 1 f a (x 1, x 2 ) m x4 f c (x 4 ) = 1 m fc x 2 (x 2 ) = x 4 f c (x 2, x 4 ) m x2 f b (x 2 ) = m fa x 2 (x 2 )m fc x 2 (x 2 ) m fb x 3 (x 3 ) = x 2 f b (x 2, x 3 )m x2 f b

44 Sum-product algorithms Message passing on a tree structured factor graph Up - down: m x3 f b (x 3 ) = 1 m fb x 2 (x 2 ) = x 3 f b (x 2, x 3 ) m x2 f a (x 2 ) = m fb x 2 (x 2 )m fc x 2 (x 2 ) m fa x 1 (x 1 ) = x 2 f a (x 1, x 2 )m x2 f a (x 2 ) m x2 f c (x 2 ) = m fa x 2 (x 2 )m fb x 2 (x 2 ) m fc x 4 (x 4 ) = x 2 f c (x 2.x 4 )m x2 f c (x 2 )

45 Sum-product algorithms Message passing on a tree structured factor graph Calculating p(x 2 ) p(x 2 ) = 1 Z m f a x 2 (x 2 )m fb x 2 (x 2 )m fc x2 (x 2 ) = 1 { }{ }{ } f a (x 1, x 2 ) f b (x 2, x 3 ) f c (x 2, x 4 ) Z x 1 x 3 x 4 = 1 f a (x 1, x 2 )f b (x 2, x 3 )f c (x 2, x 4 ) Z x 1 x 3 x 4 = p(x) x 1 x 3 x 4

46 Junction tree algorithm Junction tree algorithm: message passing on general graphs 1. Moralize to a undirected graph if starting with a directed graph 2. Triangulate the graph: to make the graph chordal 3. Build the junction tree 4. Run message passing algorithm on the junction tree

47 Junction tree algorithm 662 Chapter 19. Undirected graphical models (Markov random fields) X1 X2 X3 X4 X5 X1 X2 X3 X4 X5 Junction tree algorithm X6 X7 X8 X9 X10 X6 X7 X8 X9 X10 X11 X12 X13 X14 X15 1. Moralize to a undirected graph if starting with a directed X16 X17 X18 X19 X20 X16 X17 X18 X19 X20 graph (a) X11 X12 X13 X14 X15 Moralization: Figure 19.1 marrying (a) A 2d lattice represented theasparents, a DAG. The dotted red sonode tox8 speak, is independentforce of all otherthe nodes (black) given its Markov blanket, which include its parents (blue), children (green) and co-parents parents of (orange). a (b) node The sameto model form represented aas completed a UGM. The red node X8 sub-graph. is independent of the other black nodes given its neighbors (blue nodes). (b) (a) (b) Figure 19.2 (a) A DGM. (b) Its moralized version, represented as a UGM. The set of nodes that renders a node t conditionally independent of all the other nodes in the graph is called t s Markov blanket; we will denote this by mb(t). Formally, the Markov blanket satisfies the following property: t V\cl(t) mb(t) (19.1)

48 Junction tree algorithm Junction tree algorithm 2. Triangulate the graph: to make the graph chordal Finding chord-less cycles containing four or more nodes and adding extra links to eliminate such chord-less cycles. x 1 x 1 x 5 x 2 x 5 x 2 x 4 x 3 x 4 x 3

49 Junction tree algorithm Junction tree algorithm 3. Build the junction tree The nodes of a junction tree are the maximal cliques of the triangulated graph, The links connect pairs of nodes(maximal cliques) that have variables in common. The junction tree is build so that the summation of the edge-weight in the tree is maximal, where the weight of an edge is the number of variables shared by the two maximal cliques it connected. x 1 x 1,x 2,x 5 x 1,x 2,x 5 x 5 x 2 x 5 x 3,x 4,x 5 x 2,x 5 x 2,x 3,x 5 x 3,x 5 x 3,x 5 x 4 x 3 x 2,x 3,x 5 No x 3,x 4,x 5 Yes

50 Junction tree algorithm Junction tree algorithm 4. Run message passing algorithm on the junction tree The message passing algorithm on the junction tree is similarly to the sum-product algorithm on a tree, the result give the joint distribution of each maximal clique of the triangulated graph. The computational complexity of the junction tree algorithm is determined by the size of the largest clique, which is lower bounded by the treewidth of the graph,where treewidth is one less than the size of the largest clique.

51 Thanks! Questions!

The Basics of Graphical Models

The Basics of Graphical Models The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures

More information

3. The Junction Tree Algorithms

3. The Junction Tree Algorithms A Short Course on Graphical Models 3. The Junction Tree Algorithms Mark Paskin mark@paskin.org 1 Review: conditional independence Two random variables X and Y are independent (written X Y ) iff p X ( )

More information

5 Directed acyclic graphs

5 Directed acyclic graphs 5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate

More information

Course: Model, Learning, and Inference: Lecture 5

Course: Model, Learning, and Inference: Lecture 5 Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 yuille@stat.ucla.edu Abstract Probability distributions on structured representation.

More information

Markov random fields and Gibbs measures

Markov random fields and Gibbs measures Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).

More information

Practice with Proofs

Practice with Proofs Practice with Proofs October 6, 2014 Recall the following Definition 0.1. A function f is increasing if for every x, y in the domain of f, x < y = f(x) < f(y) 1. Prove that h(x) = x 3 is increasing, using

More information

Scheduling Shop Scheduling. Tim Nieberg

Scheduling Shop Scheduling. Tim Nieberg Scheduling Shop Scheduling Tim Nieberg Shop models: General Introduction Remark: Consider non preemptive problems with regular objectives Notation Shop Problems: m machines, n jobs 1,..., n operations

More information

Conditional Random Fields: An Introduction

Conditional Random Fields: An Introduction Conditional Random Fields: An Introduction Hanna M. Wallach February 24, 2004 1 Labeling Sequential Data The task of assigning label sequences to a set of observation sequences arises in many fields, including

More information

Bayesian networks - Time-series models - Apache Spark & Scala

Bayesian networks - Time-series models - Apache Spark & Scala Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly

More information

Lecture 16 : Relations and Functions DRAFT

Lecture 16 : Relations and Functions DRAFT CS/Math 240: Introduction to Discrete Mathematics 3/29/2011 Lecture 16 : Relations and Functions Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT In Lecture 3, we described a correspondence

More information

Approximating the Partition Function by Deleting and then Correcting for Model Edges

Approximating the Partition Function by Deleting and then Correcting for Model Edges Approximating the Partition Function by Deleting and then Correcting for Model Edges Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles Los Angeles, CA 995

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct

More information

Life of A Knowledge Base (KB)

Life of A Knowledge Base (KB) Life of A Knowledge Base (KB) A knowledge base system is a special kind of database management system to for knowledge base management. KB extraction: knowledge extraction using statistical models in NLP/ML

More information

Structured Learning and Prediction in Computer Vision. Contents

Structured Learning and Prediction in Computer Vision. Contents Foundations and Trends R in Computer Graphics and Vision Vol. 6, Nos. 3 4 (2010) 185 365 c 2011 S. Nowozin and C. H. Lampert DOI: 10.1561/0600000033 Structured Learning and Prediction in Computer Vision

More information

SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis

SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October 17, 2015 Outline

More information

Lecture 1: Course overview, circuits, and formulas

Lecture 1: Course overview, circuits, and formulas Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik

More information

Bayesian Networks. Mausam (Slides by UW-AI faculty)

Bayesian Networks. Mausam (Slides by UW-AI faculty) Bayesian Networks Mausam (Slides by UW-AI faculty) Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential space for representation & inference BNs provide

More information

Statistical machine learning, high dimension and big data

Statistical machine learning, high dimension and big data Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

More information

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...

More information

UNCOUPLING THE PERRON EIGENVECTOR PROBLEM

UNCOUPLING THE PERRON EIGENVECTOR PROBLEM UNCOUPLING THE PERRON EIGENVECTOR PROBLEM Carl D Meyer INTRODUCTION Foranonnegative irreducible matrix m m with spectral radius ρ,afundamental problem concerns the determination of the unique normalized

More information

LECTURE 4. Last time: Lecture outline

LECTURE 4. Last time: Lecture outline LECTURE 4 Last time: Types of convergence Weak Law of Large Numbers Strong Law of Large Numbers Asymptotic Equipartition Property Lecture outline Stochastic processes Markov chains Entropy rate Random

More information

1. Prove that the empty set is a subset of every set.

1. Prove that the empty set is a subset of every set. 1. Prove that the empty set is a subset of every set. Basic Topology Written by Men-Gen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since

More information

Lecture 7: NP-Complete Problems

Lecture 7: NP-Complete Problems IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 7: NP-Complete Problems David Mix Barrington and Alexis Maciel July 25, 2000 1. Circuit

More information

Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs

Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Kousha Etessami U. of Edinburgh, UK Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 6) 1 / 13 Overview Graphs and Graph

More information

Part 2: Community Detection

Part 2: Community Detection Chapter 8: Graph Data Part 2: Community Detection Based on Leskovec, Rajaraman, Ullman 2014: Mining of Massive Datasets Big Data Management and Analytics Outline Community Detection - Social networks -

More information

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact

More information

DATA ANALYSIS II. Matrix Algorithms

DATA ANALYSIS II. Matrix Algorithms DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where

More information

2.1 Complexity Classes

2.1 Complexity Classes 15-859(M): Randomized Algorithms Lecturer: Shuchi Chawla Topic: Complexity classes, Identity checking Date: September 15, 2004 Scribe: Andrew Gilpin 2.1 Complexity Classes In this lecture we will look

More information

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm. Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

TOPIC 4: DERIVATIVES

TOPIC 4: DERIVATIVES TOPIC 4: DERIVATIVES 1. The derivative of a function. Differentiation rules 1.1. The slope of a curve. The slope of a curve at a point P is a measure of the steepness of the curve. If Q is a point on the

More information

Formal Languages and Automata Theory - Regular Expressions and Finite Automata -

Formal Languages and Automata Theory - Regular Expressions and Finite Automata - Formal Languages and Automata Theory - Regular Expressions and Finite Automata - Samarjit Chakraborty Computer Engineering and Networks Laboratory Swiss Federal Institute of Technology (ETH) Zürich March

More information

Large Margin DAGs for Multiclass Classification

Large Margin DAGs for Multiclass Classification S.A. Solla, T.K. Leen and K.-R. Müller (eds.), 57 55, MIT Press (000) Large Margin DAGs for Multiclass Classification John C. Platt Microsoft Research Microsoft Way Redmond, WA 9805 jplatt@microsoft.com

More information

Tree-representation of set families and applications to combinatorial decompositions

Tree-representation of set families and applications to combinatorial decompositions Tree-representation of set families and applications to combinatorial decompositions Binh-Minh Bui-Xuan a, Michel Habib b Michaël Rao c a Department of Informatics, University of Bergen, Norway. buixuan@ii.uib.no

More information

CS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010

CS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010 CS 598CSC: Combinatorial Optimization Lecture date: /4/010 Instructor: Chandra Chekuri Scribe: David Morrison Gomory-Hu Trees (The work in this section closely follows [3]) Let G = (V, E) be an undirected

More information

Regular Languages and Finite State Machines

Regular Languages and Finite State Machines Regular Languages and Finite State Machines Plan for the Day: Mathematical preliminaries - some review One application formal definition of finite automata Examples 1 Sets A set is an unordered collection

More information

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm. Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three

More information

OPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION

OPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION OPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION Sérgio Pequito, Stephen Kruzick, Soummya Kar, José M. F. Moura, A. Pedro Aguiar Department of Electrical and Computer Engineering

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

More information

3 Some Integer Functions

3 Some Integer Functions 3 Some Integer Functions A Pair of Fundamental Integer Functions The integer function that is the heart of this section is the modulo function. However, before getting to it, let us look at some very simple

More information

Graphical Models, Exponential Families, and Variational Inference

Graphical Models, Exponential Families, and Variational Inference Foundations and Trends R in Machine Learning Vol. 1, Nos. 1 2 (2008) 1 305 c 2008 M. J. Wainwright and M. I. Jordan DOI: 10.1561/2200000001 Graphical Models, Exponential Families, and Variational Inference

More information

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding

More information

Cost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation:

Cost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation: CSE341T 08/31/2015 Lecture 3 Cost Model: Work, Span and Parallelism In this lecture, we will look at how one analyze a parallel program written using Cilk Plus. When we analyze the cost of an algorithm

More information

Burgers vector, Burgers circuit, and Dislocation Line Direction

Burgers vector, Burgers circuit, and Dislocation Line Direction Burgers vector, Burgers circuit, and Dislocation Line Direction Keonwook Kang and Wei Cai November 21, 2007 The 1st version of this white paper was written after the online discussion between Keonwook

More information

CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.

CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning. Lecture Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott

More information

Midterm Practice Problems

Midterm Practice Problems 6.042/8.062J Mathematics for Computer Science October 2, 200 Tom Leighton, Marten van Dijk, and Brooke Cowan Midterm Practice Problems Problem. [0 points] In problem set you showed that the nand operator

More information

CMPSCI611: Approximating MAX-CUT Lecture 20

CMPSCI611: Approximating MAX-CUT Lecture 20 CMPSCI611: Approximating MAX-CUT Lecture 20 For the next two lectures we ll be seeing examples of approximation algorithms for interesting NP-hard problems. Today we consider MAX-CUT, which we proved to

More information

. 0 1 10 2 100 11 1000 3 20 1 2 3 4 5 6 7 8 9

. 0 1 10 2 100 11 1000 3 20 1 2 3 4 5 6 7 8 9 Introduction The purpose of this note is to find and study a method for determining and counting all the positive integer divisors of a positive integer Let N be a given positive integer We say d is a

More information

Incorporating Evidence in Bayesian networks with the Select Operator

Incorporating Evidence in Bayesian networks with the Select Operator Incorporating Evidence in Bayesian networks with the Select Operator C.J. Butz and F. Fang Department of Computer Science, University of Regina Regina, Saskatchewan, Canada SAS 0A2 {butz, fang11fa}@cs.uregina.ca

More information

Problem Set 7 Solutions

Problem Set 7 Solutions 8 8 Introduction to Algorithms May 7, 2004 Massachusetts Institute of Technology 6.046J/18.410J Professors Erik Demaine and Shafi Goldwasser Handout 25 Problem Set 7 Solutions This problem set is due in

More information

The Open University s repository of research publications and other research outputs

The Open University s repository of research publications and other research outputs Open Research Online The Open University s repository of research publications and other research outputs The degree-diameter problem for circulant graphs of degree 8 and 9 Journal Article How to cite:

More information

Basics of Statistical Machine Learning

Basics of Statistical Machine Learning CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar

More information

E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

More information

Topology-based network security

Topology-based network security Topology-based network security Tiit Pikma Supervised by Vitaly Skachek Research Seminar in Cryptography University of Tartu, Spring 2013 1 Introduction In both wired and wireless networks, there is the

More information

Tagging with Hidden Markov Models

Tagging with Hidden Markov Models Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,

More information

Lecture 17 : Equivalence and Order Relations DRAFT

Lecture 17 : Equivalence and Order Relations DRAFT CS/Math 240: Introduction to Discrete Mathematics 3/31/2011 Lecture 17 : Equivalence and Order Relations Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT Last lecture we introduced the notion

More information

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5.

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5. PUTNAM TRAINING POLYNOMIALS (Last updated: November 17, 2015) Remark. This is a list of exercises on polynomials. Miguel A. Lerma Exercises 1. Find a polynomial with integral coefficients whose zeros include

More information

Large-Sample Learning of Bayesian Networks is NP-Hard

Large-Sample Learning of Bayesian Networks is NP-Hard Journal of Machine Learning Research 5 (2004) 1287 1330 Submitted 3/04; Published 10/04 Large-Sample Learning of Bayesian Networks is NP-Hard David Maxwell Chickering David Heckerman Christopher Meek Microsoft

More information

Mining Social Network Graphs

Mining Social Network Graphs Mining Social Network Graphs Debapriyo Majumdar Data Mining Fall 2014 Indian Statistical Institute Kolkata November 13, 17, 2014 Social Network No introduc+on required Really? We s7ll need to understand

More information

Outline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits

Outline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits Outline NP-completeness Examples of Easy vs. Hard problems Euler circuit vs. Hamiltonian circuit Shortest Path vs. Longest Path 2-pairs sum vs. general Subset Sum Reducing one problem to another Clique

More information

Mining Social-Network Graphs

Mining Social-Network Graphs 342 Chapter 10 Mining Social-Network Graphs There is much information to be gained by analyzing the large-scale data that is derived from social networks. The best-known example of a social network is

More information

Chapter 7: Termination Detection

Chapter 7: Termination Detection Chapter 7: Termination Detection Ajay Kshemkalyani and Mukesh Singhal Distributed Computing: Principles, Algorithms, and Systems Cambridge University Press A. Kshemkalyani and M. Singhal (Distributed Computing)

More information

VERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS

VERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS VERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS MICHAEL DRMOTA, OMER GIMENEZ, AND MARC NOY Abstract. We show that the number of vertices of a given degree k in several kinds of series-parallel labelled

More information

Why? A central concept in Computer Science. Algorithms are ubiquitous.

Why? A central concept in Computer Science. Algorithms are ubiquitous. Analysis of Algorithms: A Brief Introduction Why? A central concept in Computer Science. Algorithms are ubiquitous. Using the Internet (sending email, transferring files, use of search engines, online

More information

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like

More information

Bayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University caizhua@gmail.com

Bayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University caizhua@gmail.com Bayesian Machine Learning (ML): Modeling And Inference in Big Data Zhuhua Cai Google Rice University caizhua@gmail.com 1 Syllabus Bayesian ML Concepts (Today) Bayesian ML on MapReduce (Next morning) Bayesian

More information

Bayesian Network Modeling of Hangul Characters for On-line Handwriting Recognition

Bayesian Network Modeling of Hangul Characters for On-line Handwriting Recognition Bayesian Network Modeling of Hangul haracters for On-line Handwriting Recognition Sung-ung ho and in H. Kim S Div., EES Dept., KAIS, 373- Kusong-dong, Yousong-ku, Daejon, 305-70, KOREA {sjcho, jkim}@ai.kaist.ac.kr

More information

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called

More information

Lecture 4: BK inequality 27th August and 6th September, 2007

Lecture 4: BK inequality 27th August and 6th September, 2007 CSL866: Percolation and Random Graphs IIT Delhi Amitabha Bagchi Scribe: Arindam Pal Lecture 4: BK inequality 27th August and 6th September, 2007 4. Preliminaries The FKG inequality allows us to lower bound

More information

Chapter 6: Graph Theory

Chapter 6: Graph Theory Chapter 6: Graph Theory Graph theory deals with routing and network problems and if it is possible to find a best route, whether that means the least expensive, least amount of time or the least distance.

More information

Bayesian Networks. Read R&N Ch. 14.1-14.2. Next lecture: Read R&N 18.1-18.4

Bayesian Networks. Read R&N Ch. 14.1-14.2. Next lecture: Read R&N 18.1-18.4 Bayesian Networks Read R&N Ch. 14.1-14.2 Next lecture: Read R&N 18.1-18.4 You will be expected to know Basic concepts and vocabulary of Bayesian networks. Nodes represent random variables. Directed arcs

More information

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that

More information

Reliability. 26.1 Reliability Models. Chapter 26 Page 1

Reliability. 26.1 Reliability Models. Chapter 26 Page 1 Chapter 26 Page 1 Reliability Although the technological achievements of the last 50 years can hardly be disputed, there is one weakness in all mankind's devices. That is the possibility of failure. What

More information

Learning Sum-Product Networks with Direct and Indirect Variable Interactions

Learning Sum-Product Networks with Direct and Indirect Variable Interactions Learning Sum-Product Networks with Direct and Indirect Variable Interactions Amirmohammad Rooshenas Daniel Lowd Department of Computer and Information Science, University of Oregon PEDRAM@CS.UOREGON.EDU

More information

HT2015: SC4 Statistical Data Mining and Machine Learning

HT2015: SC4 Statistical Data Mining and Machine Learning HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric

More information

Definition 11.1. Given a graph G on n vertices, we define the following quantities:

Definition 11.1. Given a graph G on n vertices, we define the following quantities: Lecture 11 The Lovász ϑ Function 11.1 Perfect graphs We begin with some background on perfect graphs. graphs. First, we define some quantities on Definition 11.1. Given a graph G on n vertices, we define

More information

Triangle deletion. Ernie Croot. February 3, 2010

Triangle deletion. Ernie Croot. February 3, 2010 Triangle deletion Ernie Croot February 3, 2010 1 Introduction The purpose of this note is to give an intuitive outline of the triangle deletion theorem of Ruzsa and Szemerédi, which says that if G = (V,

More information

Compression algorithm for Bayesian network modeling of binary systems

Compression algorithm for Bayesian network modeling of binary systems Compression algorithm for Bayesian network modeling of binary systems I. Tien & A. Der Kiureghian University of California, Berkeley ABSTRACT: A Bayesian network (BN) is a useful tool for analyzing the

More information

IE 680 Special Topics in Production Systems: Networks, Routing and Logistics*

IE 680 Special Topics in Production Systems: Networks, Routing and Logistics* IE 680 Special Topics in Production Systems: Networks, Routing and Logistics* Rakesh Nagi Department of Industrial Engineering University at Buffalo (SUNY) *Lecture notes from Network Flows by Ahuja, Magnanti

More information

Compact Representations and Approximations for Compuation in Games

Compact Representations and Approximations for Compuation in Games Compact Representations and Approximations for Compuation in Games Kevin Swersky April 23, 2008 Abstract Compact representations have recently been developed as a way of both encoding the strategic interactions

More information

TU e. Advanced Algorithms: experimentation project. The problem: load balancing with bounded look-ahead. Input: integer m 2: number of machines

TU e. Advanced Algorithms: experimentation project. The problem: load balancing with bounded look-ahead. Input: integer m 2: number of machines The problem: load balancing with bounded look-ahead Input: integer m 2: number of machines integer k 0: the look-ahead numbers t 1,..., t n : the job sizes Problem: assign jobs to machines machine to which

More information

6.3 Conditional Probability and Independence

6.3 Conditional Probability and Independence 222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted

More information

Unified Lecture # 4 Vectors

Unified Lecture # 4 Vectors Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,

More information

Extracting correlation structure from large random matrices

Extracting correlation structure from large random matrices Extracting correlation structure from large random matrices Alfred Hero University of Michigan - Ann Arbor Feb. 17, 2012 1 / 46 1 Background 2 Graphical models 3 Screening for hubs in graphical model 4

More information

Finding the M Most Probable Configurations Using Loopy Belief Propagation

Finding the M Most Probable Configurations Using Loopy Belief Propagation Finding the M Most Probable Configurations Using Loopy Belief Propagation Chen Yanover and Yair Weiss School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, Israel

More information

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year. This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra

More information

An Application of Graphical Modeling to the Analysis of Intranet Benefits and Applications

An Application of Graphical Modeling to the Analysis of Intranet Benefits and Applications Journal of Data Science 3(2005), 1-17 An Application of Graphical Modeling to the Analysis of Intranet Benefits and Applications Raffaella Settimi, Linda V. Knight, Theresa A. Steinbach and James D. White

More information

Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University

Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University 1 Chapter 1 Probability 1.1 Basic Concepts In the study of statistics, we consider experiments

More information

Graph Theory Problems and Solutions

Graph Theory Problems and Solutions raph Theory Problems and Solutions Tom Davis tomrdavis@earthlink.net http://www.geometer.org/mathcircles November, 005 Problems. Prove that the sum of the degrees of the vertices of any finite graph is

More information

Chapter 2 Introduction to Sequence Analysis for Human Behavior Understanding

Chapter 2 Introduction to Sequence Analysis for Human Behavior Understanding Chapter 2 Introduction to Sequence Analysis for Human Behavior Understanding Hugues Salamin and Alessandro Vinciarelli 2.1 Introduction Human sciences recognize sequence analysis as a key aspect of any

More information

A Sublinear Bipartiteness Tester for Bounded Degree Graphs

A Sublinear Bipartiteness Tester for Bounded Degree Graphs A Sublinear Bipartiteness Tester for Bounded Degree Graphs Oded Goldreich Dana Ron February 5, 1998 Abstract We present a sublinear-time algorithm for testing whether a bounded degree graph is bipartite

More information

Discuss the size of the instance for the minimum spanning tree problem.

Discuss the size of the instance for the minimum spanning tree problem. 3.1 Algorithm complexity The algorithms A, B are given. The former has complexity O(n 2 ), the latter O(2 n ), where n is the size of the instance. Let n A 0 be the size of the largest instance that can

More information

Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling

Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one

More information

Inner Product Spaces

Inner Product Spaces Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

More information

Solutions to Homework 6

Solutions to Homework 6 Solutions to Homework 6 Debasish Das EECS Department, Northwestern University ddas@northwestern.edu 1 Problem 5.24 We want to find light spanning trees with certain special properties. Given is one example

More information

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

More information

Properties of Real Numbers

Properties of Real Numbers 16 Chapter P Prerequisites P.2 Properties of Real Numbers What you should learn: Identify and use the basic properties of real numbers Develop and use additional properties of real numbers Why you should

More information

Evaluation of Machine Learning Techniques for Green Energy Prediction

Evaluation of Machine Learning Techniques for Green Energy Prediction arxiv:1406.3726v1 [cs.lg] 14 Jun 2014 Evaluation of Machine Learning Techniques for Green Energy Prediction 1 Objective Ankur Sahai University of Mainz, Germany We evaluate Machine Learning techniques

More information