4. Sum-product algorithm

Size: px
Start display at page:

Download "4. Sum-product algorithm"

Transcription

1 Sum-product algorithm Sum-product algorithm Elimination algorithm Sum-product algorithm on a line Sum-product algorithm on a tree

2 Sum-product algorithm 4-2 Inference tasks on graphical models consider an undirected graphical model (a.k.a. Markov random field) µ(x) = 1 ψ c (x c ) Z c C where C is the set of all maximal cliques in G we want to calculate marginals: µ(x A ) = x V \A µ(x) calculating conditional distributions µ(x A x B ) = µ(x A, x B ) µ(x B ) calculation maximum a posteriori estimates: arg maxˆx µ(ˆx) calculating the partition function Z sample from this distribution

3 Elimination algorithm for calculating marginals Elimination algorithm is exact but can require O( X V ) operations µ(x) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) we want to compute µ(x 1 ) brute force marginalization: µ(x 1 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) x 2,x 3,x 4,x 5 X requires O( C X 5 ) operations, where big-o denotes that it is upper bounded by C C X 5 for some constant C Sum-product algorithm 4-3

4 consider an elimination ordering (5, 4, 3, 2) µ(x 1 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) = = = = x 2,x 3,x 4,x 5 X x 2,x 3,x 4 X x 2,x 3,x 4 X x 2,x 3 X x 2,x 3 X = x 2 X ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) x 5 X ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )m 5 (x 2, x 3, x 4 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) x 4 X ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) } {{ } m 5 (x 2,x 3,x 4 ) m 5 (x 2, x 3, x 4 ) } {{ } m 4 (x 2,x 3 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) ψ 12 (x 1, x 2 ) x 3 X ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) } {{ } m 3 (x 1,x 2 ) Sum-product algorithm 4-4

5 Sum-product algorithm 4-5 µ(x 1 ) ψ 12 (x 1, x 2 ) ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) x 2 X = x 2 X m 2 (x 1 ) x 3 X ψ 12 (x 1, x 2 )m 3 (x 1, x 2 ) } {{ } m 3 (x 1,x 2 ) normalize m 2 ( ) to get µ(x 1 ) µ(x 1 ) = m 2(x 1 ) ˆx 1 m 2 (ˆx 1 ) computational complexity depends on the elimination ordering how do we know which ordering is better?

6 Computational complexity of elimination algorithm µ(x 1 ) = x 2,x 3,x 4,x 5 X x 2,x 3,x 4 X = x 2,x 3 X = x 2 X = x 2 X ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) ψ 12 (x 1, x 2 ) x 3 X x 5 X ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) }{{} m 5(S 5), S 5={x 2,x 3,x 4}, Ψ 5={ψ 25,ψ 345} x 4 X m 5 (x 2, x 3, x 4 ) }{{} m 4(S 4), S 4={x 2,x 3},Ψ 4={m 5} ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) }{{} m 3(S 3), S 3={x 1,x 2}, Ψ 3={ψ 13,m 4} ψ 12 (x 1, x 2 )m 3 (x 1, x 2 ) }{{} m 2(S 2), S 2={x 1}, Ψ 2={ψ 12,m 3} = m 2 (x 1 ) Total complexity: i O( Ψ i X 1+ Si ) = O( V max i Ψ i X 1+maxi Si ) Sum-product algorithm 4-6

7 theorem: every maximal clique in G(G, I) corresponds to a domain of a message S i {i} for some i size of the largest clique in G(G, I) is 1 + max i S i different orderings I s give different cliques, resulting in varying complexity Sum-product algorithm 4-7 Induced graph elimination algorithm as transformation of graphs induced graph G(G, I) for a graph G and an elimination ordering I is the union of (the edges of) all the transformed graphs or equivalently, start from G and for each i I connect all pairs in Si

8 Sum-product algorithm 4-8 theorem: finding optimal elimination ordering is NP-hard any suggestions? heuristics give I = (4, 5, 3, 2, 1) for Bayesian networks the same algorithm works with conditional probabilities instead of compatibility functions complexity analysis can be done on moralized undirected graph intermediate messages do not correspond to a conditional distribution

9 Sum-product algorithm 4-9 Elimination algorithm input: {ψ c } c C, alphabet X, subset A V, elimination ordering I output: marginal µ(x A ) 1. initialize active set Ψ to be the set of input compatibility functions 2. for node i in I that is not in A do let S i be the set of nodes, not including i, that share a compatibility function with i end let Ψ i be the set of compatibility functions in Ψ involving x i compute m i (x Si ) = x i ψ Ψ i ψ(x i, x Si ) remove elements of Ψ i from Ψ add m i to Ψ 3. normalize µ(x A ) = ψ Ψ ψ(x A) / x A ψ Ψ ψ(x A)

10 Sum-product algorithm 4-10 Belief Propagation for approximate inference given a pairwise MRF: µ(x) = 1 Z compute marginal: µ(x i ) (i,j) E ψ i,j(x i, x j ) message update: set of 2 E messages on each (directed) edge {ν i j (x i )} (i,j) E, where ν i j : X R + encoding our belief (or approximate P(x i )) decision: ν (t+1) i j (x i) = k i\j O(di X 2 ) computations ν (t) i (x i ) = k i { } xk ν(t) k i (x k)ψ i,k (x i, x k ) = k i { x k { ν (t) i k (x i) } ν (t) k i (x k)ψ i,k (x i, x k ) } 1/(di 1) µ(xi ) = νi(xi) x νi(x )

11 Sum-product algorithm on a line ν 3 4 (x 3 ) ν 5 4 (x 5 ) x 1 x 2 x 3 x 4 x 5 x 6 x 7 Remove node i and recursively compute marginals on sub-graphs µ(x i) = x [n]\{i} µ(x) µ(x) = 1 Z n 1 i=1 ψ i,i+1 (x i, x i+1 ) ψ(x 1, x 2) ψ(x i 2, x i 1) ψ(x i 1, x i)ψ(x i, x i+1) ψ(x i+1, x i+2) ψ(x n 1, x n) }{{}}{{} x [n]\{i} µ i 1 i :joint dist. on a sub-graph µ i+1 i ν i 1 i(x i 1) }{{} x i 1,x i+1 marginal dist. on a subrgaph ψ(x i 1, x i)ψ(x i, x i+1)ν i+1 i(x i+1) ν i 1 i(x i 1) µ i 1 i(x 1,..., x i 1) x 1,...,x i 2 ν i+1 i(x i+1) µ i+1 i(x i+1,..., x n) x i+2,...,x Sum-product algorithm n 4-11

12 Sum-product algorithm 4-12 ν i (x i ) x i 1,x i+1 ν i 1 i (x i 1 )ψ(x i 1, x i )ψ(x i, x i+1 )ν i+1 i (x i+1 ) µ(x i ) = ν i(x i ) x i ν i (x i ) definitions: of joint distribution and marginal on sub-graphs µ i 1 i (x 1,..., x i 1 ) ν i 1 i (x i 1 ) µ i+1 i (x i+1,..., x n ) ν i+1 i (x i+1 ) 1 Z i 1 i k [i 2] ψ(x k, x k+1 ) x 1,...,x i 2 µ i 1 i (x 1,..., x i 1 ) 1 Z i+1 i k {i+2,...,n} ψ(x k 1, x k ) x i+2,...,x n µ i+1 i (x i+1,..., x n )

13 how can we compute the messages ν, recursively? µ i i+1 (x 1,..., x i ) µ i 1 i (x 1,..., x i 1 )ψ(x i 1, x 1 ) ν i i+1 (x i ) = µ i i+1 (x 1,..., x i ) x 1,...,x i 1 µ i 1 i (x 1,..., x i 1 )ψ(x i 1, x 1 ) x 1,...,x i 1 = x i 1 ν i 1 i (x i 1 )ψ(x i 1, x i ) ν 1 2 (x 1 ) = 1/ X ν 2 3 (x 2 ) x 1 1 X ψ(x 1, x 2 ) how many operations are required? O(n X 2 ) operations to compute one marginal µ(x i ) what if we want all the marginals? compute all the messages forward and backward O(n X 2 ) then compute all the marginals in O(n X 2 ) operations Sum-product algorithm 4-13

14 Sum-product algorithm 4-14 Computing partition function is as easy as computing marginals computing the partition function from the messages Z i (i+1) = ψ(x k, x k+1 ) x 1,...,x i k [i 1] = Z (i 1) i ν (i 1) i (x i 1 ) ψ(x i 1, x i ) x i,x i 1 Z 1 = 1 Z n = Z how many operations do we need? O(n X 2 ) operations

15 Sum-product algorithm 4-15 Sum-product algorithm for hidden Markov models hidden Markov model Sequence of r.v. s {(X 1, Y 1 ); (X 2, Y 2 );... ; (X n, Y n )} n 1 hidden state {X i } Markov Chain P{x} = P{x 1 } P{x i+1 x i } {Y i } noisy observations P{y x} = i=1 n P{y i x i } i=i x 1 x 2 x 3 x 4 x 5 x 6 x 7 y 1 y 2 y 3 y 4 y 5 y 6 y 7 (equivalent directed form) x 1 x 2 x 3 x 4 x 5 x 6 x 7 y 1 y 2 y 3 y 4 y 5 y 6 y 7

16 Sum-product algorithm 4-16 time homogeneous hidden Markov models Sequence of r.v. s {(X 1, Y 1 ); (X 2, Y 2 );... ; (X n, Y n )} n 1 {X i } Markov Chain P{x} = q 0 (x 1 ) q(x i, x i+1 ) {Y i } noisy observations P{y x} = i=1 n r(x i, y i ) i=i µ(x, y) = 1 Z n 1 i=1 ψ i (x i, x i+1 ) n ψ i (x i, y i ), i=1 ψ i (x i, x i+1 ) = q(x i, x i+1 ), ψi (x i, y i ) = r(x i, y i ).

17 Sum-product algorithm 4-17 we want to compute marginals of the following graphical model on a line (q 0 uniform) x 1 x 2 x 3 x 4 x 5 x 6 x 7 y 1 y 2 y 3 y 4 y 5 y 6 y 7 µ y (x) = P{x y} Bayes thm = 1 Z(y) n 1 i=1 q(x i, x i+1 ) n r(x i, y i ). i=1 µ y (x) = = n 1 1 q(x i, x i+1 ) Z(y) i=1 n 1 1 ψ i (x i, x i+1 ) Z(y) i=1 n r(x i, y i ) i=1 ψ i (x i, x i+1 ) = q(x i, x i+1 ) r(x i, y i ) (for i < n 1) ψ n 1 (x n 1, x n ) = q(x n 1, x n ) r(x n 1, y n 1 ) r(x n, y n ).

18 Sum-product algorithm 4-18 apply sum-product algorithm to compute marginals in O(n X 2 ) time ν i (i+1) (x i ) q(x i 1, x i ) r(x i 1, y i 1 ) ν (i 1) i (x i 1 ), ν (i+1) i (x i ) x i 1 X x i+1 X q(x i, x i+1 ) r(x i, y i ) ν (i+2) (i+1) (x i+1 ). known as forward-backward algorithm a special case of the sum-product algorithm BCJR algorithm for convolutional codes ([Bahl, Cocke, Jelinek and Raviv 1974]) cannot find the maximum likelihood estimate (cf. Viterbi algorithm) implement sum-product algorithm for HMM [Homework 2.7] consider an extension of inference on HMM [Homework 2.8]

19 Sum-product algorithm 4-19 Homework 2.7 S&P 500 index over a period of time For each week, measure the price movement relative to the previous week: +1 indicates up and 1 indicates down a hidden Markov model in which x t denotes the economic state (good or bad) of week t and y t denotes the price movement (up or down) x t+1 = x t with probability 0.8 P Yt X t (y t = +1 x t = good ) = P Yt X t (y t = 1 x t = bad ) = q

20 Sum-product algorithm 4-20 Example: Neuron firing patterns Hypothesis Assemblies of neurones activate in a coordinate way in correspondence to specific cognitive functions. Performing of the function corresponds sequence of these activity states. Approach Firing process Observed variables Activity states Hidden variables

21 single subject ensemble neural spikes horizontal eye position vertical eye position baseline planning moving automatically detect (as opposed to manually specify) baseline, plan, and perimovement epochs of neural activity detect target movement in advance goal: neural prosthesis to help patients with spinal cord injury or neurodegenerative disease and significantly impaired motor control [C. Kemere, G. Santhanam, B. M. Yu, A. Afshar, S.I. Ryu, T. H. Meng and Sum-product algorithm 4-21

22 Sum-product algorithm 4-22 discrete time 10ms P(xt+1 x t ) = A ij, P(# spikes for measurement k = d x t = i) e λ k,i λ d k,i likelihood: P(x t=s) s P(xt=s )

23 Sum-product algorithm 4-23 Belief propagation for factor graphs b x 5 x 6 x 4 c x 3 d x 2 x 1 factor ψ a (x 1, x 5, x 6 ) a }{{} a µ(x) = 1 Z ψ a (x a ) a F

24 variable nodes i, j, etc.; factor nodes a, b, etc. set of messages {ν i a } (i,a) E and { ν a i } (a,i) E messages from variables to factors ν i a (x i ) = ν b i (x i ) b i\{a} messages from factors to variables ν a i (x i ) = ψ a (x a ) x a\{i} j a\{i} messages from variables to factors ν i (x i ) = ν b i (x i ) b i ν j a (x j ) this includes belief propagation (=sum-product algorithm) on (general) Markov random fields exact on factor trees Sum-product algorithm 4-24

25 Example: decoding LDPC codes LDPC code is defined by a factor graph model variable nodes factor nodes x i {0, 1} x1 a x2 ψ a(x i, x j, x k ) = I(x i x j x k = 0) x3 b block length n = 4 number of factors m = 2 allowed messages = {0000, 0111, 1010, 1101} decoding using belief propagation (for BSC with ɛ = ) µ y (x) = 1 P Z Y X (y i x i ) I( x a = 0) i V a F use (parallel) sum-product algorithm to find µ(x i ) and let x4 ˆx i = arg max µ(x i ) minimizes bit error rate Sum-product algorithm 4-25

26 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] [ ] [ ] [ ] a b [ ] [ ] [ ] [ ] [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26

27 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] 0.7 [ ] 0.7 a b [ ] [ ] [ ] [ ] [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26

28 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] 0.7 [ ] 0.7 a b [ ] 0.42 = 0.58 [ ] [ ] [ ] 0.7 [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26

29 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] a [ ] [ ] b [ ] [ ] [ ] [ ] 0.7 ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26

30 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] a b [ ] [ ] [ ] [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26

31 ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26 Decoding by sum-product algorithm ˆµ(x i ) [ ] [ ] y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] a b [ ] [ ] [ ] [ ]

32 Sum-product algorithm on trees a motivating example: influenza virus complete sequence of the gene of the 1918 influenza virus [A.H. Reid, T.G. Fanning, J.V. Hultin, and J.K. Taubenberger, Proc. Natl. Acad. Sci. 96 (1999) ] Sum-product algorithm 4-27

33 challenges in phylogeny phylogeny reconstruction: given DNA sequences at vertices (only at leaves), infer the underlying tree T = (V, E). phylogeny evaluation: given a tree T = (V, E) evaluate the probability of observed DNA sequences at vertices (only at leaves). Bayesian network model for phylogeny evaluation T = (V, D) directed graph, DNA sequences x = (x i ) i V X V µ T (x) = q o (x o ) q i,j (x i, x j ), (i,j) D q i,j (x i, x j ) = Probability that the descendent is x j if ancestor is x i. Sum-product algorithm 4-28

34 simplified model: X = {+1, 1} q o (x o ) = 1 { 2 1 q if x j = x i q(x i, x j ) = q if x i x j MRF representation: q(x i, x j ) e θx ix j with θ = 1 2 log q 1 q probability of certain tree of mutations x: µ T (x) = 1 Z θ (T ) (i,j) E e θx ix j problem: for given T, compute marginal µ T (x i ) we prove the correctness of sum-product algorithm for this model, but the same proof holds for any general pairwise MRF (and also for general MRF and FG) Sum-product algorithm 4-29

35 Sum-product algorithm 4-30 define graphical model on sub-trees j i T j i T j i = (V j i, E j i) Subtree rooted at j and excluding i µ j i(x Vj i ) 1 Z(T j i) ν j i(x j) x Vj i \{j} (u,v) E j i e θxuxv µ j i(x Vj i )

36 T k1 i = (V 1, E 1) T k2 i k 1 i k 2 k 4 k 3 T k4 i T k3 i the messages from neighbors k 1, k 2, k 3, k 4 are sufficient to compute the marginal µ(x i) µ T (x i) e θxuxv = = x V \{i} (u,v) E x V1,x V2,x V3,x V4 l=1 4 l=1 x kl,x Vl \{k l } 4 { l=1 x kl e θxixkl 4 {e θx ix kl {e θx ix kl x Vl \{k l } (u,v) E l e θxuxv (u,v) E l e θxuxv } µ kl i(x Vl ) } {{ } ν kl i(x kl ) Sum-product algorithm 4-31 } }

37 Sum-product algorithm 4-32 recursion on sub-trees to compute the messages ν j i k l T i j µ i j(x Vi j ) = = 1 Z(T i j) (u,v) E i j e θxuxv 1 { Z(T i j) eθx ix k e θx ix l e θxuxv e θx ix l { (u,v) E k i e θxuxv (u,v) E k i e θxuxv }{ e θx ix k e θx ix l µ k i (x Vk i )µ l i (x Vl i ) }{ (u,v) E l i e θxuxv (u,v) E l i e θxuxv } }

38 ν i j j ν k i k i ν l i l ν i j(x i) = µ i j(x Vi j ) x Vi j \i e θx ix k e θx ix l µ k i (x Vk i )µ l i (x Vl i ) = x Vi j \i { }{ } e θxixk µ k i (x Vk i ) e θxixl µ l i (x Vl i ) x Vk i x Vl i { }{ e θxixk µ k i (x Vk i ) e θxixl x k x Vk i \{k} { x k }{ ) x l } ) e θxixk ν k i (x k e θxixl ν l i (x l x l x Vl i \{l} } µ l i (x Vl i ) with uniform initialization ν i j(x i) = 1 X for all leaves i Sum-product algorithm 4-33

39 sum-product algorithm (for our example) ν i j (x i ) { } e θx ix k ν k i (x k ) ν i (x i ) k i\j x k { } e θx ix k ν k i (x k ) k i x k µ T (x i ) = ν i (x i ) x i ν i (x i ) what if we want all the marginals? choose an arbitrary root φ compute all the messages towards the root ( E messages) then compute all the messages outwards from the root ( E messages) then compute all the marginals (n marginals) how many operations are required? naive implementation requires O( X 2 i d2 i ) if i has degree d i, then computing ν i j requires d i X 2 operations d i messages start at each node i, each require d i X 2 operations total computation for 2 E messages is { } i d i (d i X 2 ) however, we can compute all marginals in O(n X 2 ) operations Sum-product algorithm 4-34

40 Sum-product algorithm 4-35 let D = {(i, j), (j, i) (i, j) E} be the directed version of E (cf. D = 2 E ) (sequential) sum-product algorithm 1. initialize ν i j (x i ) = 1/ X for all leaves i 2. recursively over (i, j) D compute (from leaves) { } ν i j (x i ) = ψ ik (x i, x k )ν k i (x k ) x k k i\j 3. for each i V compute marginal ν i (x i ) = { } ψ ik (x i, x k )ν k i (x k ) x k µ T (x i ) = k i ν i (x i ) x i ν i (x i )

41 (parallel) sum-product algorithm 1. initialize ν (0) i j (x i) = 1/ X for all (i, j) D 2. for t {0, 1,..., t max } for all (i, j) D compute ν (t+1) i j (x { } i) = ψ ik (x i, x k )ν (t) k i (x k) k i k i\j 3. for each i V compute marginal ν i (x i ) = { } ψ ik (x i, x k )ν (tmax+1) k i (x k ) µ T (x i ) = x k ν i (x i ) x i ν i (x i ) also called belief propagation when t max is larger than the diameter of the tree (the length of the longest path), this converges to the correct marginal [Homework 2.5] more operations than the sequential version ( O(n X 2 diam(t )) ) a naive implementation requires O( X 2 diam(t ) d 2 i ) naturally extends to general graphs but no proof of exactness Sum-product algorithm 4-36 x k

42 Sum-product algorithm on general graphs (loopy) belief propagation 1. initialize ν i j (x i ) = 1/ X for all (i, j) D 2. for t {0, 1,..., t max } for all (i, j) D compute ν (t+1) i j (x i) = k i k i\j x k { x k } ψ ik (x i, x k )ν (t) k i (x k) 3. for each i V compute marginal ν i (x i ) = { } ψ ik (x i, x k )ν (tmax+1) k i (x k ) µ T (x i ) = ν i (x i ) x i ν i (x i ) computes approximate marginals in O(n X 2 t max ) operations generally it does not converge; even if it does, it might be incorrect folklore about loopy BP works better when G has few short loops works better when ψij (x i, x j ) = ψ ij,1 (x i )ψ ij,2 (x j ) + small(x i, x j ) nonconvex variational principle Sum-product algorithm 4-37

43 Exercise: partition function on trees using the recursion for messages: { } ν i j (x i ) = ψ ik (x i, x k )ν k i (x k ) k i\j x k ν i (x i ) = { } ψ ik (x i, x k )ν k i (x k ) x k k i it follows that we can easily compute the partition function as Z(T ) = ν i (x i ) x i alternatively, if we had a black box that computes marginals for any tree, then we can use it to compute partition functions efficiently Z(T i j ) = { } ψ ik (x i, x k ) Z(T k i ) µ k i (x k ) Z(T ) = x i X k i\j x i X k i x k X { } ψ ik (x i, x k ) Z(T k i ) µ k i (x k ) x k X this recursive algorithm naturally extends to general graphs Sum-product algorithm 4-38

44 Sum-product algorithm 4-39 Why would one want to compute the partition function? Suppose you observe x = (+1, +1, +1, +1, +1, +1, +1, +1, +1) and you know this comes from either of µ(x) = 1 Z(T ) (e.g. coloring) (i,j) E ψ(x i, x j ) which one has highest likelihood?

45 Exercise: sampling on the tree if we have a black-box for computing marginals on any tree, we can use it to sample from any distribution on a tree Sampling( Tree T = (V, E), ψ = {ψ ij } (ij) E ) 1: Choose a root o V ; 2: Sample X o µ o ( ); 2: Recursively over i V (from root to leaves): 3: Compute µ i π(i) (x i x π(i) ); 4: Sample X i µ i π(i) ( x π(i) ); π(i) is the parent of node i in the rooted tree T o we use the black-box to compute the conditional distribution k i j µ T (x Vi j x j ) ψ ij (x i x j ) µ i j (x Vi j ) µ T (x i x j ) ψ ij (x i x j ) µ i j (x i ) Sum-product algorithm 4-40

46 Sum-product algorithm 4-41 Tree decomposition when we don t have a tree we can create an equivalent tree graph by enlarging the alphabet X X k Treewidth(G) Minimum such k it is NP-hard to determine the treewidth of a graph problem: in general Treewidth(G) = Θ(n)

47 Sum-product algorithm 4-42 Tree decomposition of G = (V, E) 1 1,2 1, , , A tree T = (V T, E T ) and a mapping V : V T SUBSETS(V ) s.t.: For each i V there exists at least one u V T with i V (u). For each (i, j) E there exists at least one u V T with i, j V (u). If i V (u 1 ) and i V (u 2 ), then i V (w) for any w on the path between u 1 and u 2 in T.

Course: Model, Learning, and Inference: Lecture 5

Course: Model, Learning, and Inference: Lecture 5 Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 yuille@stat.ucla.edu Abstract Probability distributions on structured representation.

More information

The Basics of Graphical Models

The Basics of Graphical Models The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures

More information

5 Directed acyclic graphs

5 Directed acyclic graphs 5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate

More information

SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis

SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October 17, 2015 Outline

More information

Finding the M Most Probable Configurations Using Loopy Belief Propagation

Finding the M Most Probable Configurations Using Loopy Belief Propagation Finding the M Most Probable Configurations Using Loopy Belief Propagation Chen Yanover and Yair Weiss School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, Israel

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

More information

GRAPH THEORY LECTURE 4: TREES

GRAPH THEORY LECTURE 4: TREES GRAPH THEORY LECTURE 4: TREES Abstract. 3.1 presents some standard characterizations and properties of trees. 3.2 presents several different types of trees. 3.7 develops a counting method based on a bijection

More information

Graphical Models, Exponential Families, and Variational Inference

Graphical Models, Exponential Families, and Variational Inference Foundations and Trends R in Machine Learning Vol. 1, Nos. 1 2 (2008) 1 305 c 2008 M. J. Wainwright and M. I. Jordan DOI: 10.1561/2200000001 Graphical Models, Exponential Families, and Variational Inference

More information

Hidden Markov Models

Hidden Markov Models 8.47 Introduction to omputational Molecular Biology Lecture 7: November 4, 2004 Scribe: Han-Pang hiu Lecturer: Ross Lippert Editor: Russ ox Hidden Markov Models The G island phenomenon The nucleotide frequencies

More information

3. The Junction Tree Algorithms

3. The Junction Tree Algorithms A Short Course on Graphical Models 3. The Junction Tree Algorithms Mark Paskin mark@paskin.org 1 Review: conditional independence Two random variables X and Y are independent (written X Y ) iff p X ( )

More information

Analysis of Algorithms I: Binary Search Trees

Analysis of Algorithms I: Binary Search Trees Analysis of Algorithms I: Binary Search Trees Xi Chen Columbia University Hash table: A data structure that maintains a subset of keys from a universe set U = {0, 1,..., p 1} and supports all three dictionary

More information

Why? A central concept in Computer Science. Algorithms are ubiquitous.

Why? A central concept in Computer Science. Algorithms are ubiquitous. Analysis of Algorithms: A Brief Introduction Why? A central concept in Computer Science. Algorithms are ubiquitous. Using the Internet (sending email, transferring files, use of search engines, online

More information

Tagging with Hidden Markov Models

Tagging with Hidden Markov Models Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,

More information

Complexity Theory. IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar

Complexity Theory. IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar Complexity Theory IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar Outline Goals Computation of Problems Concepts and Definitions Complexity Classes and Problems Polynomial Time Reductions Examples

More information

Social Media Mining. Graph Essentials

Social Media Mining. Graph Essentials Graph Essentials Graph Basics Measures Graph and Essentials Metrics 2 2 Nodes and Edges A network is a graph nodes, actors, or vertices (plural of vertex) Connections, edges or ties Edge Node Measures

More information

Approximating the Partition Function by Deleting and then Correcting for Model Edges

Approximating the Partition Function by Deleting and then Correcting for Model Edges Approximating the Partition Function by Deleting and then Correcting for Model Edges Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles Los Angeles, CA 995

More information

Cost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation:

Cost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation: CSE341T 08/31/2015 Lecture 3 Cost Model: Work, Span and Parallelism In this lecture, we will look at how one analyze a parallel program written using Cilk Plus. When we analyze the cost of an algorithm

More information

Tree-representation of set families and applications to combinatorial decompositions

Tree-representation of set families and applications to combinatorial decompositions Tree-representation of set families and applications to combinatorial decompositions Binh-Minh Bui-Xuan a, Michel Habib b Michaël Rao c a Department of Informatics, University of Bergen, Norway. buixuan@ii.uib.no

More information

Scheduling Shop Scheduling. Tim Nieberg

Scheduling Shop Scheduling. Tim Nieberg Scheduling Shop Scheduling Tim Nieberg Shop models: General Introduction Remark: Consider non preemptive problems with regular objectives Notation Shop Problems: m machines, n jobs 1,..., n operations

More information

Solutions to Homework 6

Solutions to Homework 6 Solutions to Homework 6 Debasish Das EECS Department, Northwestern University ddas@northwestern.edu 1 Problem 5.24 We want to find light spanning trees with certain special properties. Given is one example

More information

Hidden Markov Models in Bioinformatics. By Máthé Zoltán Kőrösi Zoltán 2006

Hidden Markov Models in Bioinformatics. By Máthé Zoltán Kőrösi Zoltán 2006 Hidden Markov Models in Bioinformatics By Máthé Zoltán Kőrösi Zoltán 2006 Outline Markov Chain HMM (Hidden Markov Model) Hidden Markov Models in Bioinformatics Gene Finding Gene Finding Model Viterbi algorithm

More information

Structured Learning and Prediction in Computer Vision. Contents

Structured Learning and Prediction in Computer Vision. Contents Foundations and Trends R in Computer Graphics and Vision Vol. 6, Nos. 3 4 (2010) 185 365 c 2011 S. Nowozin and C. H. Lampert DOI: 10.1561/0600000033 Structured Learning and Prediction in Computer Vision

More information

LECTURE 4. Last time: Lecture outline

LECTURE 4. Last time: Lecture outline LECTURE 4 Last time: Types of convergence Weak Law of Large Numbers Strong Law of Large Numbers Asymptotic Equipartition Property Lecture outline Stochastic processes Markov chains Entropy rate Random

More information

Analysis of Algorithms, I

Analysis of Algorithms, I Analysis of Algorithms, I CSOR W4231.002 Eleni Drinea Computer Science Department Columbia University Thursday, February 26, 2015 Outline 1 Recap 2 Representing graphs 3 Breadth-first search (BFS) 4 Applications

More information

Graphical degree sequences and realizations

Graphical degree sequences and realizations swap Graphical and realizations Péter L. Erdös Alfréd Rényi Institute of Mathematics Hungarian Academy of Sciences MAPCON 12 MPIPKS - Dresden, May 15, 2012 swap Graphical and realizations Péter L. Erdös

More information

Chapter 7: Termination Detection

Chapter 7: Termination Detection Chapter 7: Termination Detection Ajay Kshemkalyani and Mukesh Singhal Distributed Computing: Principles, Algorithms, and Systems Cambridge University Press A. Kshemkalyani and M. Singhal (Distributed Computing)

More information

SEQUENCES OF MAXIMAL DEGREE VERTICES IN GRAPHS. Nickolay Khadzhiivanov, Nedyalko Nenov

SEQUENCES OF MAXIMAL DEGREE VERTICES IN GRAPHS. Nickolay Khadzhiivanov, Nedyalko Nenov Serdica Math. J. 30 (2004), 95 102 SEQUENCES OF MAXIMAL DEGREE VERTICES IN GRAPHS Nickolay Khadzhiivanov, Nedyalko Nenov Communicated by V. Drensky Abstract. Let Γ(M) where M V (G) be the set of all vertices

More information

Big Data, Machine Learning, Causal Models

Big Data, Machine Learning, Causal Models Big Data, Machine Learning, Causal Models Sargur N. Srihari University at Buffalo, The State University of New York USA Int. Conf. on Signal and Image Processing, Bangalore January 2014 1 Plan of Discussion

More information

Maximum Likelihood Graph Structure Estimation with Degree Distributions

Maximum Likelihood Graph Structure Estimation with Degree Distributions Maximum Likelihood Graph Structure Estimation with Distributions Bert Huang Computer Science Department Columbia University New York, NY 17 bert@cs.columbia.edu Tony Jebara Computer Science Department

More information

Network (Tree) Topology Inference Based on Prüfer Sequence

Network (Tree) Topology Inference Based on Prüfer Sequence Network (Tree) Topology Inference Based on Prüfer Sequence C. Vanniarajan and Kamala Krithivasan Department of Computer Science and Engineering Indian Institute of Technology Madras Chennai 600036 vanniarajanc@hcl.in,

More information

Small Maximal Independent Sets and Faster Exact Graph Coloring

Small Maximal Independent Sets and Faster Exact Graph Coloring Small Maximal Independent Sets and Faster Exact Graph Coloring David Eppstein Univ. of California, Irvine Dept. of Information and Computer Science The Exact Graph Coloring Problem: Given an undirected

More information

2. (a) Explain the strassen s matrix multiplication. (b) Write deletion algorithm, of Binary search tree. [8+8]

2. (a) Explain the strassen s matrix multiplication. (b) Write deletion algorithm, of Binary search tree. [8+8] Code No: R05220502 Set No. 1 1. (a) Describe the performance analysis in detail. (b) Show that f 1 (n)+f 2 (n) = 0(max(g 1 (n), g 2 (n)) where f 1 (n) = 0(g 1 (n)) and f 2 (n) = 0(g 2 (n)). [8+8] 2. (a)

More information

Single machine parallel batch scheduling with unbounded capacity

Single machine parallel batch scheduling with unbounded capacity Workshop on Combinatorics and Graph Theory 21th, April, 2006 Nankai University Single machine parallel batch scheduling with unbounded capacity Yuan Jinjiang Department of mathematics, Zhengzhou University

More information

Bayesian networks - Time-series models - Apache Spark & Scala

Bayesian networks - Time-series models - Apache Spark & Scala Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly

More information

Introduced by Stuart Kauffman (ca. 1986) as a tunable family of fitness landscapes.

Introduced by Stuart Kauffman (ca. 1986) as a tunable family of fitness landscapes. 68 Part II. Combinatorial Models can require a number of spin flips that is exponential in N (A. Haken et al. ca. 1989), and that one can in fact embed arbitrary computations in the dynamics (Orponen 1995).

More information

Lecture 1: Course overview, circuits, and formulas

Lecture 1: Course overview, circuits, and formulas Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik

More information

Monte Carlo-based statistical methods (MASM11/FMS091)

Monte Carlo-based statistical methods (MASM11/FMS091) Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February 5, 2013 J. Olsson Monte Carlo-based

More information

Classification/Decision Trees (II)

Classification/Decision Trees (II) Classification/Decision Trees (II) Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Right Sized Trees Let the expected misclassification rate of a tree T be R (T ).

More information

Factor Graphs and the Sum-Product Algorithm

Factor Graphs and the Sum-Product Algorithm 498 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 47, NO. 2, FEBRUARY 2001 Factor Graphs and the Sum-Product Algorithm Frank R. Kschischang, Senior Member, IEEE, Brendan J. Frey, Member, IEEE, and Hans-Andrea

More information

Introduction to Deep Learning Variational Inference, Mean Field Theory

Introduction to Deep Learning Variational Inference, Mean Field Theory Introduction to Deep Learning Variational Inference, Mean Field Theory 1 Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Center for Visual Computing Ecole Centrale Paris Galen Group INRIA-Saclay Lecture 3: recap

More information

ON INDUCED SUBGRAPHS WITH ALL DEGREES ODD. 1. Introduction

ON INDUCED SUBGRAPHS WITH ALL DEGREES ODD. 1. Introduction ON INDUCED SUBGRAPHS WITH ALL DEGREES ODD A.D. SCOTT Abstract. Gallai proved that the vertex set of any graph can be partitioned into two sets, each inducing a subgraph with all degrees even. We prove

More information

Bayesian probability theory

Bayesian probability theory Bayesian probability theory Bruno A. Olshausen arch 1, 2004 Abstract Bayesian probability theory provides a mathematical framework for peforming inference, or reasoning, using probability. The foundations

More information

CHAPTER 2 Estimating Probabilities

CHAPTER 2 Estimating Probabilities CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a

More information

A Non-Linear Schema Theorem for Genetic Algorithms

A Non-Linear Schema Theorem for Genetic Algorithms A Non-Linear Schema Theorem for Genetic Algorithms William A Greene Computer Science Department University of New Orleans New Orleans, LA 70148 bill@csunoedu 504-280-6755 Abstract We generalize Holland

More information

How To Solve A K Path In Time (K)

How To Solve A K Path In Time (K) What s next? Reductions other than kernelization Dániel Marx Humboldt-Universität zu Berlin (with help from Fedor Fomin, Daniel Lokshtanov and Saket Saurabh) WorKer 2010: Workshop on Kernelization Nov

More information

Gambling and Data Compression

Gambling and Data Compression Gambling and Data Compression Gambling. Horse Race Definition The wealth relative S(X) = b(x)o(x) is the factor by which the gambler s wealth grows if horse X wins the race, where b(x) is the fraction

More information

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

More information

Guessing Game: NP-Complete?

Guessing Game: NP-Complete? Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES 2. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple

More information

On the k-path cover problem for cacti

On the k-path cover problem for cacti On the k-path cover problem for cacti Zemin Jin and Xueliang Li Center for Combinatorics and LPMC Nankai University Tianjin 300071, P.R. China zeminjin@eyou.com, x.li@eyou.com Abstract In this paper we

More information

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm. Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three

More information

Asking Hard Graph Questions. Paul Burkhardt. February 3, 2014

Asking Hard Graph Questions. Paul Burkhardt. February 3, 2014 Beyond Watson: Predictive Analytics and Big Data U.S. National Security Agency Research Directorate - R6 Technical Report February 3, 2014 300 years before Watson there was Euler! The first (Jeopardy!)

More information

6.852: Distributed Algorithms Fall, 2009. Class 2

6.852: Distributed Algorithms Fall, 2009. Class 2 .8: Distributed Algorithms Fall, 009 Class Today s plan Leader election in a synchronous ring: Lower bound for comparison-based algorithms. Basic computation in general synchronous networks: Leader election

More information

Variational Mean Field for Graphical Models

Variational Mean Field for Graphical Models Variational Mean Field for Graphical Models CS/CNS/EE 155 Baback Moghaddam Machine Learning Group baback @ jpl.nasa.gov Approximate Inference Consider general UGs (i.e., not tree-structured) All basic

More information

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014 Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about

More information

Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling

Chapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one

More information

Near Optimal Solutions

Near Optimal Solutions Near Optimal Solutions Many important optimization problems are lacking efficient solutions. NP-Complete problems unlikely to have polynomial time solutions. Good heuristics important for such problems.

More information

IE 680 Special Topics in Production Systems: Networks, Routing and Logistics*

IE 680 Special Topics in Production Systems: Networks, Routing and Logistics* IE 680 Special Topics in Production Systems: Networks, Routing and Logistics* Rakesh Nagi Department of Industrial Engineering University at Buffalo (SUNY) *Lecture notes from Network Flows by Ahuja, Magnanti

More information

Coding and decoding with convolutional codes. The Viterbi Algor

Coding and decoding with convolutional codes. The Viterbi Algor Coding and decoding with convolutional codes. The Viterbi Algorithm. 8 Block codes: main ideas Principles st point of view: infinite length block code nd point of view: convolutions Some examples Repetition

More information

Ex. 2.1 (Davide Basilio Bartolini)

Ex. 2.1 (Davide Basilio Bartolini) ECE 54: Elements of Information Theory, Fall 00 Homework Solutions Ex.. (Davide Basilio Bartolini) Text Coin Flips. A fair coin is flipped until the first head occurs. Let X denote the number of flips

More information

On the independence number of graphs with maximum degree 3

On the independence number of graphs with maximum degree 3 On the independence number of graphs with maximum degree 3 Iyad A. Kanj Fenghui Zhang Abstract Let G be an undirected graph with maximum degree at most 3 such that G does not contain any of the three graphs

More information

Communication Cost in Big Data Processing

Communication Cost in Big Data Processing Communication Cost in Big Data Processing Dan Suciu University of Washington Joint work with Paul Beame and Paris Koutris and the Database Group at the UW 1 Queries on Big Data Big Data processing on distributed

More information

A Sublinear Bipartiteness Tester for Bounded Degree Graphs

A Sublinear Bipartiteness Tester for Bounded Degree Graphs A Sublinear Bipartiteness Tester for Bounded Degree Graphs Oded Goldreich Dana Ron February 5, 1998 Abstract We present a sublinear-time algorithm for testing whether a bounded degree graph is bipartite

More information

Outline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits

Outline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits Outline NP-completeness Examples of Easy vs. Hard problems Euler circuit vs. Hamiltonian circuit Shortest Path vs. Longest Path 2-pairs sum vs. general Subset Sum Reducing one problem to another Clique

More information

Lecture 22: November 10

Lecture 22: November 10 CS271 Randomness & Computation Fall 2011 Lecture 22: November 10 Lecturer: Alistair Sinclair Based on scribe notes by Rafael Frongillo Disclaimer: These notes have not been subjected to the usual scrutiny

More information

Part 2: Community Detection

Part 2: Community Detection Chapter 8: Graph Data Part 2: Community Detection Based on Leskovec, Rajaraman, Ullman 2014: Mining of Massive Datasets Big Data Management and Analytics Outline Community Detection - Social networks -

More information

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma Please Note: The references at the end are given for extra reading if you are interested in exploring these ideas further. You are

More information

Seminar. Path planning using Voronoi diagrams and B-Splines. Stefano Martina stefano.martina@stud.unifi.it

Seminar. Path planning using Voronoi diagrams and B-Splines. Stefano Martina stefano.martina@stud.unifi.it Seminar Path planning using Voronoi diagrams and B-Splines Stefano Martina stefano.martina@stud.unifi.it 23 may 2016 This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International

More information

Finding and counting given length cycles

Finding and counting given length cycles Finding and counting given length cycles Noga Alon Raphael Yuster Uri Zwick Abstract We present an assortment of methods for finding and counting simple cycles of a given length in directed and undirected

More information

Tracking Groups of Pedestrians in Video Sequences

Tracking Groups of Pedestrians in Video Sequences Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal

More information

Distributed Computing over Communication Networks: Maximal Independent Set

Distributed Computing over Communication Networks: Maximal Independent Set Distributed Computing over Communication Networks: Maximal Independent Set What is a MIS? MIS An independent set (IS) of an undirected graph is a subset U of nodes such that no two nodes in U are adjacent.

More information

MapReduce and Distributed Data Analysis. Sergei Vassilvitskii Google Research

MapReduce and Distributed Data Analysis. Sergei Vassilvitskii Google Research MapReduce and Distributed Data Analysis Google Research 1 Dealing With Massive Data 2 2 Dealing With Massive Data Polynomial Memory Sublinear RAM Sketches External Memory Property Testing 3 3 Dealing With

More information

Statistical machine learning, high dimension and big data

Statistical machine learning, high dimension and big data Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,

More information

1 Homework 1. [p 0 q i+j +... + p i 1 q j+1 ] + [p i q j ] + [p i+1 q j 1 +... + p i+j q 0 ]

1 Homework 1. [p 0 q i+j +... + p i 1 q j+1 ] + [p i q j ] + [p i+1 q j 1 +... + p i+j q 0 ] 1 Homework 1 (1) Prove the ideal (3,x) is a maximal ideal in Z[x]. SOLUTION: Suppose we expand this ideal by including another generator polynomial, P / (3, x). Write P = n + x Q with n an integer not

More information

ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN. Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015

ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN. Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015 ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015 ONLINE STEINER FOREST PROBLEM An initially given graph G. s 1 s 2 A sequence of demands (s i, t i ) arriving

More information

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding

More information

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.

! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm. Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of

More information

A Practical Scheme for Wireless Network Operation

A Practical Scheme for Wireless Network Operation A Practical Scheme for Wireless Network Operation Radhika Gowaikar, Amir F. Dana, Babak Hassibi, Michelle Effros June 21, 2004 Abstract In many problems in wireline networks, it is known that achieving

More information

On an anti-ramsey type result

On an anti-ramsey type result On an anti-ramsey type result Noga Alon, Hanno Lefmann and Vojtĕch Rödl Abstract We consider anti-ramsey type results. For a given coloring of the k-element subsets of an n-element set X, where two k-element

More information

Lecture 7: NP-Complete Problems

Lecture 7: NP-Complete Problems IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 7: NP-Complete Problems David Mix Barrington and Alexis Maciel July 25, 2000 1. Circuit

More information

Large induced subgraphs with all degrees odd

Large induced subgraphs with all degrees odd Large induced subgraphs with all degrees odd A.D. Scott Department of Pure Mathematics and Mathematical Statistics, University of Cambridge, England Abstract: We prove that every connected graph of order

More information

DATA ANALYSIS II. Matrix Algorithms

DATA ANALYSIS II. Matrix Algorithms DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where

More information

Coding Theorems for Turbo-Like Codes Abstract. 1. Introduction.

Coding Theorems for Turbo-Like Codes Abstract. 1. Introduction. Coding Theorems for Turbo-Like Codes Dariush Divsalar, Hui Jin, and Robert J. McEliece Jet Propulsion Laboratory and California Institute of Technology Pasadena, California USA E-mail: dariush@shannon.jpl.nasa.gov,

More information

Protein Protein Interaction Networks

Protein Protein Interaction Networks Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks Young-Rae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

Message-passing sequential detection of multiple change points in networks

Message-passing sequential detection of multiple change points in networks Message-passing sequential detection of multiple change points in networks Long Nguyen, Arash Amini Ram Rajagopal University of Michigan Stanford University ISIT, Boston, July 2012 Nguyen/Amini/Rajagopal

More information

International Journal of Advanced Research in Computer Science and Software Engineering

International Journal of Advanced Research in Computer Science and Software Engineering Volume 3, Issue 7, July 23 ISSN: 2277 28X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Greedy Algorithm:

More information

Life of A Knowledge Base (KB)

Life of A Knowledge Base (KB) Life of A Knowledge Base (KB) A knowledge base system is a special kind of database management system to for knowledge base management. KB extraction: knowledge extraction using statistical models in NLP/ML

More information

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091) Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February

More information

Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010. Chapter 7: Digraphs

Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010. Chapter 7: Digraphs MCS-236: Graph Theory Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010 Chapter 7: Digraphs Strong Digraphs Definitions. A digraph is an ordered pair (V, E), where V is the set

More information

Efficient Recovery of Secrets

Efficient Recovery of Secrets Efficient Recovery of Secrets Marcel Fernandez Miguel Soriano, IEEE Senior Member Department of Telematics Engineering. Universitat Politècnica de Catalunya. C/ Jordi Girona 1 i 3. Campus Nord, Mod C3,

More information

Hidden Markov Models Chapter 15

Hidden Markov Models Chapter 15 Hidden Markov Models Chapter 15 Mausam (Slides based on Dan Klein, Luke Zettlemoyer, Alex Simma, Erik Sudderth, David Fernandez-Baca, Drena Dobbs, Serafim Batzoglou, William Cohen, Andrew McCallum, Dan

More information

A 2-factor in which each cycle has long length in claw-free graphs

A 2-factor in which each cycle has long length in claw-free graphs A -factor in which each cycle has long length in claw-free graphs Roman Čada Shuya Chiba Kiyoshi Yoshimoto 3 Department of Mathematics University of West Bohemia and Institute of Theoretical Computer Science

More information

Every tree contains a large induced subgraph with all degrees odd

Every tree contains a large induced subgraph with all degrees odd Every tree contains a large induced subgraph with all degrees odd A.J. Radcliffe Carnegie Mellon University, Pittsburgh, PA A.D. Scott Department of Pure Mathematics and Mathematical Statistics University

More information

Monte Carlo-based statistical methods (MASM11/FMS091)

Monte Carlo-based statistical methods (MASM11/FMS091) Monte Carlo-based statistical methods (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 6 Sequential Monte Carlo methods II February 7, 2014 M. Wiktorsson

More information

Statistical Machine Translation: IBM Models 1 and 2

Statistical Machine Translation: IBM Models 1 and 2 Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation

More information

Dynamic Programming and Graph Algorithms in Computer Vision

Dynamic Programming and Graph Algorithms in Computer Vision Dynamic Programming and Graph Algorithms in Computer Vision Pedro F. Felzenszwalb and Ramin Zabih Abstract Optimization is a powerful paradigm for expressing and solving problems in a wide range of areas,

More information

Bayesian Networks. Mausam (Slides by UW-AI faculty)

Bayesian Networks. Mausam (Slides by UW-AI faculty) Bayesian Networks Mausam (Slides by UW-AI faculty) Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential space for representation & inference BNs provide

More information

Methods for big data in medical genomics

Methods for big data in medical genomics Methods for big data in medical genomics Parallel Hidden Markov Models in Population Genetics Chris Holmes, (Peter Kecskemethy & Chris Gamble) Department of Statistics and, Nuffield Department of Medicine

More information

Polarization codes and the rate of polarization

Polarization codes and the rate of polarization Polarization codes and the rate of polarization Erdal Arıkan, Emre Telatar Bilkent U., EPFL Sept 10, 2008 Channel Polarization Given a binary input DMC W, i.i.d. uniformly distributed inputs (X 1,...,

More information

Exponential time algorithms for graph coloring

Exponential time algorithms for graph coloring Exponential time algorithms for graph coloring Uriel Feige Lecture notes, March 14, 2011 1 Introduction Let [n] denote the set {1,..., k}. A k-labeling of vertices of a graph G(V, E) is a function V [k].

More information