4. Sum-product algorithm
|
|
- Allyson Davis
- 7 years ago
- Views:
Transcription
1 Sum-product algorithm Sum-product algorithm Elimination algorithm Sum-product algorithm on a line Sum-product algorithm on a tree
2 Sum-product algorithm 4-2 Inference tasks on graphical models consider an undirected graphical model (a.k.a. Markov random field) µ(x) = 1 ψ c (x c ) Z c C where C is the set of all maximal cliques in G we want to calculate marginals: µ(x A ) = x V \A µ(x) calculating conditional distributions µ(x A x B ) = µ(x A, x B ) µ(x B ) calculation maximum a posteriori estimates: arg maxˆx µ(ˆx) calculating the partition function Z sample from this distribution
3 Elimination algorithm for calculating marginals Elimination algorithm is exact but can require O( X V ) operations µ(x) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) we want to compute µ(x 1 ) brute force marginalization: µ(x 1 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) x 2,x 3,x 4,x 5 X requires O( C X 5 ) operations, where big-o denotes that it is upper bounded by C C X 5 for some constant C Sum-product algorithm 4-3
4 consider an elimination ordering (5, 4, 3, 2) µ(x 1 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) = = = = x 2,x 3,x 4,x 5 X x 2,x 3,x 4 X x 2,x 3,x 4 X x 2,x 3 X x 2,x 3 X = x 2 X ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) x 5 X ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )m 5 (x 2, x 3, x 4 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) x 4 X ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) } {{ } m 5 (x 2,x 3,x 4 ) m 5 (x 2, x 3, x 4 ) } {{ } m 4 (x 2,x 3 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) ψ 12 (x 1, x 2 ) x 3 X ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) } {{ } m 3 (x 1,x 2 ) Sum-product algorithm 4-4
5 Sum-product algorithm 4-5 µ(x 1 ) ψ 12 (x 1, x 2 ) ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) x 2 X = x 2 X m 2 (x 1 ) x 3 X ψ 12 (x 1, x 2 )m 3 (x 1, x 2 ) } {{ } m 3 (x 1,x 2 ) normalize m 2 ( ) to get µ(x 1 ) µ(x 1 ) = m 2(x 1 ) ˆx 1 m 2 (ˆx 1 ) computational complexity depends on the elimination ordering how do we know which ordering is better?
6 Computational complexity of elimination algorithm µ(x 1 ) = x 2,x 3,x 4,x 5 X x 2,x 3,x 4 X = x 2,x 3 X = x 2 X = x 2 X ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 )ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) ψ 12 (x 1, x 2 )ψ 13 (x 1, x 3 ) ψ 12 (x 1, x 2 ) x 3 X x 5 X ψ 25 (x 2, x 5 )ψ 345 (x 3, x 4, x 5 ) }{{} m 5(S 5), S 5={x 2,x 3,x 4}, Ψ 5={ψ 25,ψ 345} x 4 X m 5 (x 2, x 3, x 4 ) }{{} m 4(S 4), S 4={x 2,x 3},Ψ 4={m 5} ψ 13 (x 1, x 3 )m 4 (x 2, x 3 ) }{{} m 3(S 3), S 3={x 1,x 2}, Ψ 3={ψ 13,m 4} ψ 12 (x 1, x 2 )m 3 (x 1, x 2 ) }{{} m 2(S 2), S 2={x 1}, Ψ 2={ψ 12,m 3} = m 2 (x 1 ) Total complexity: i O( Ψ i X 1+ Si ) = O( V max i Ψ i X 1+maxi Si ) Sum-product algorithm 4-6
7 theorem: every maximal clique in G(G, I) corresponds to a domain of a message S i {i} for some i size of the largest clique in G(G, I) is 1 + max i S i different orderings I s give different cliques, resulting in varying complexity Sum-product algorithm 4-7 Induced graph elimination algorithm as transformation of graphs induced graph G(G, I) for a graph G and an elimination ordering I is the union of (the edges of) all the transformed graphs or equivalently, start from G and for each i I connect all pairs in Si
8 Sum-product algorithm 4-8 theorem: finding optimal elimination ordering is NP-hard any suggestions? heuristics give I = (4, 5, 3, 2, 1) for Bayesian networks the same algorithm works with conditional probabilities instead of compatibility functions complexity analysis can be done on moralized undirected graph intermediate messages do not correspond to a conditional distribution
9 Sum-product algorithm 4-9 Elimination algorithm input: {ψ c } c C, alphabet X, subset A V, elimination ordering I output: marginal µ(x A ) 1. initialize active set Ψ to be the set of input compatibility functions 2. for node i in I that is not in A do let S i be the set of nodes, not including i, that share a compatibility function with i end let Ψ i be the set of compatibility functions in Ψ involving x i compute m i (x Si ) = x i ψ Ψ i ψ(x i, x Si ) remove elements of Ψ i from Ψ add m i to Ψ 3. normalize µ(x A ) = ψ Ψ ψ(x A) / x A ψ Ψ ψ(x A)
10 Sum-product algorithm 4-10 Belief Propagation for approximate inference given a pairwise MRF: µ(x) = 1 Z compute marginal: µ(x i ) (i,j) E ψ i,j(x i, x j ) message update: set of 2 E messages on each (directed) edge {ν i j (x i )} (i,j) E, where ν i j : X R + encoding our belief (or approximate P(x i )) decision: ν (t+1) i j (x i) = k i\j O(di X 2 ) computations ν (t) i (x i ) = k i { } xk ν(t) k i (x k)ψ i,k (x i, x k ) = k i { x k { ν (t) i k (x i) } ν (t) k i (x k)ψ i,k (x i, x k ) } 1/(di 1) µ(xi ) = νi(xi) x νi(x )
11 Sum-product algorithm on a line ν 3 4 (x 3 ) ν 5 4 (x 5 ) x 1 x 2 x 3 x 4 x 5 x 6 x 7 Remove node i and recursively compute marginals on sub-graphs µ(x i) = x [n]\{i} µ(x) µ(x) = 1 Z n 1 i=1 ψ i,i+1 (x i, x i+1 ) ψ(x 1, x 2) ψ(x i 2, x i 1) ψ(x i 1, x i)ψ(x i, x i+1) ψ(x i+1, x i+2) ψ(x n 1, x n) }{{}}{{} x [n]\{i} µ i 1 i :joint dist. on a sub-graph µ i+1 i ν i 1 i(x i 1) }{{} x i 1,x i+1 marginal dist. on a subrgaph ψ(x i 1, x i)ψ(x i, x i+1)ν i+1 i(x i+1) ν i 1 i(x i 1) µ i 1 i(x 1,..., x i 1) x 1,...,x i 2 ν i+1 i(x i+1) µ i+1 i(x i+1,..., x n) x i+2,...,x Sum-product algorithm n 4-11
12 Sum-product algorithm 4-12 ν i (x i ) x i 1,x i+1 ν i 1 i (x i 1 )ψ(x i 1, x i )ψ(x i, x i+1 )ν i+1 i (x i+1 ) µ(x i ) = ν i(x i ) x i ν i (x i ) definitions: of joint distribution and marginal on sub-graphs µ i 1 i (x 1,..., x i 1 ) ν i 1 i (x i 1 ) µ i+1 i (x i+1,..., x n ) ν i+1 i (x i+1 ) 1 Z i 1 i k [i 2] ψ(x k, x k+1 ) x 1,...,x i 2 µ i 1 i (x 1,..., x i 1 ) 1 Z i+1 i k {i+2,...,n} ψ(x k 1, x k ) x i+2,...,x n µ i+1 i (x i+1,..., x n )
13 how can we compute the messages ν, recursively? µ i i+1 (x 1,..., x i ) µ i 1 i (x 1,..., x i 1 )ψ(x i 1, x 1 ) ν i i+1 (x i ) = µ i i+1 (x 1,..., x i ) x 1,...,x i 1 µ i 1 i (x 1,..., x i 1 )ψ(x i 1, x 1 ) x 1,...,x i 1 = x i 1 ν i 1 i (x i 1 )ψ(x i 1, x i ) ν 1 2 (x 1 ) = 1/ X ν 2 3 (x 2 ) x 1 1 X ψ(x 1, x 2 ) how many operations are required? O(n X 2 ) operations to compute one marginal µ(x i ) what if we want all the marginals? compute all the messages forward and backward O(n X 2 ) then compute all the marginals in O(n X 2 ) operations Sum-product algorithm 4-13
14 Sum-product algorithm 4-14 Computing partition function is as easy as computing marginals computing the partition function from the messages Z i (i+1) = ψ(x k, x k+1 ) x 1,...,x i k [i 1] = Z (i 1) i ν (i 1) i (x i 1 ) ψ(x i 1, x i ) x i,x i 1 Z 1 = 1 Z n = Z how many operations do we need? O(n X 2 ) operations
15 Sum-product algorithm 4-15 Sum-product algorithm for hidden Markov models hidden Markov model Sequence of r.v. s {(X 1, Y 1 ); (X 2, Y 2 );... ; (X n, Y n )} n 1 hidden state {X i } Markov Chain P{x} = P{x 1 } P{x i+1 x i } {Y i } noisy observations P{y x} = i=1 n P{y i x i } i=i x 1 x 2 x 3 x 4 x 5 x 6 x 7 y 1 y 2 y 3 y 4 y 5 y 6 y 7 (equivalent directed form) x 1 x 2 x 3 x 4 x 5 x 6 x 7 y 1 y 2 y 3 y 4 y 5 y 6 y 7
16 Sum-product algorithm 4-16 time homogeneous hidden Markov models Sequence of r.v. s {(X 1, Y 1 ); (X 2, Y 2 );... ; (X n, Y n )} n 1 {X i } Markov Chain P{x} = q 0 (x 1 ) q(x i, x i+1 ) {Y i } noisy observations P{y x} = i=1 n r(x i, y i ) i=i µ(x, y) = 1 Z n 1 i=1 ψ i (x i, x i+1 ) n ψ i (x i, y i ), i=1 ψ i (x i, x i+1 ) = q(x i, x i+1 ), ψi (x i, y i ) = r(x i, y i ).
17 Sum-product algorithm 4-17 we want to compute marginals of the following graphical model on a line (q 0 uniform) x 1 x 2 x 3 x 4 x 5 x 6 x 7 y 1 y 2 y 3 y 4 y 5 y 6 y 7 µ y (x) = P{x y} Bayes thm = 1 Z(y) n 1 i=1 q(x i, x i+1 ) n r(x i, y i ). i=1 µ y (x) = = n 1 1 q(x i, x i+1 ) Z(y) i=1 n 1 1 ψ i (x i, x i+1 ) Z(y) i=1 n r(x i, y i ) i=1 ψ i (x i, x i+1 ) = q(x i, x i+1 ) r(x i, y i ) (for i < n 1) ψ n 1 (x n 1, x n ) = q(x n 1, x n ) r(x n 1, y n 1 ) r(x n, y n ).
18 Sum-product algorithm 4-18 apply sum-product algorithm to compute marginals in O(n X 2 ) time ν i (i+1) (x i ) q(x i 1, x i ) r(x i 1, y i 1 ) ν (i 1) i (x i 1 ), ν (i+1) i (x i ) x i 1 X x i+1 X q(x i, x i+1 ) r(x i, y i ) ν (i+2) (i+1) (x i+1 ). known as forward-backward algorithm a special case of the sum-product algorithm BCJR algorithm for convolutional codes ([Bahl, Cocke, Jelinek and Raviv 1974]) cannot find the maximum likelihood estimate (cf. Viterbi algorithm) implement sum-product algorithm for HMM [Homework 2.7] consider an extension of inference on HMM [Homework 2.8]
19 Sum-product algorithm 4-19 Homework 2.7 S&P 500 index over a period of time For each week, measure the price movement relative to the previous week: +1 indicates up and 1 indicates down a hidden Markov model in which x t denotes the economic state (good or bad) of week t and y t denotes the price movement (up or down) x t+1 = x t with probability 0.8 P Yt X t (y t = +1 x t = good ) = P Yt X t (y t = 1 x t = bad ) = q
20 Sum-product algorithm 4-20 Example: Neuron firing patterns Hypothesis Assemblies of neurones activate in a coordinate way in correspondence to specific cognitive functions. Performing of the function corresponds sequence of these activity states. Approach Firing process Observed variables Activity states Hidden variables
21 single subject ensemble neural spikes horizontal eye position vertical eye position baseline planning moving automatically detect (as opposed to manually specify) baseline, plan, and perimovement epochs of neural activity detect target movement in advance goal: neural prosthesis to help patients with spinal cord injury or neurodegenerative disease and significantly impaired motor control [C. Kemere, G. Santhanam, B. M. Yu, A. Afshar, S.I. Ryu, T. H. Meng and Sum-product algorithm 4-21
22 Sum-product algorithm 4-22 discrete time 10ms P(xt+1 x t ) = A ij, P(# spikes for measurement k = d x t = i) e λ k,i λ d k,i likelihood: P(x t=s) s P(xt=s )
23 Sum-product algorithm 4-23 Belief propagation for factor graphs b x 5 x 6 x 4 c x 3 d x 2 x 1 factor ψ a (x 1, x 5, x 6 ) a }{{} a µ(x) = 1 Z ψ a (x a ) a F
24 variable nodes i, j, etc.; factor nodes a, b, etc. set of messages {ν i a } (i,a) E and { ν a i } (a,i) E messages from variables to factors ν i a (x i ) = ν b i (x i ) b i\{a} messages from factors to variables ν a i (x i ) = ψ a (x a ) x a\{i} j a\{i} messages from variables to factors ν i (x i ) = ν b i (x i ) b i ν j a (x j ) this includes belief propagation (=sum-product algorithm) on (general) Markov random fields exact on factor trees Sum-product algorithm 4-24
25 Example: decoding LDPC codes LDPC code is defined by a factor graph model variable nodes factor nodes x i {0, 1} x1 a x2 ψ a(x i, x j, x k ) = I(x i x j x k = 0) x3 b block length n = 4 number of factors m = 2 allowed messages = {0000, 0111, 1010, 1101} decoding using belief propagation (for BSC with ɛ = ) µ y (x) = 1 P Z Y X (y i x i ) I( x a = 0) i V a F use (parallel) sum-product algorithm to find µ(x i ) and let x4 ˆx i = arg max µ(x i ) minimizes bit error rate Sum-product algorithm 4-25
26 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] [ ] [ ] [ ] a b [ ] [ ] [ ] [ ] [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26
27 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] 0.7 [ ] 0.7 a b [ ] [ ] [ ] [ ] [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26
28 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] 0.7 [ ] 0.7 a b [ ] 0.42 = 0.58 [ ] [ ] [ ] 0.7 [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26
29 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] a [ ] [ ] b [ ] [ ] [ ] [ ] 0.7 ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26
30 Decoding by sum-product algorithm y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] a b [ ] [ ] [ ] [ ] ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26
31 ν (t+1) i a (x i) = P(y i x i ) b i\{a} ν(t) b i (x i) ν (t+1) a i (x i) = x a\{i} j a\{i} ν j a(x j )I( x a = 0) Sum-product algorithm 4-26 Decoding by sum-product algorithm ˆµ(x i ) [ ] [ ] y P(y i x i) [ ] 0.7 x1 x2 x3 x4 [ ] [ ] a b [ ] [ ] [ ] [ ]
32 Sum-product algorithm on trees a motivating example: influenza virus complete sequence of the gene of the 1918 influenza virus [A.H. Reid, T.G. Fanning, J.V. Hultin, and J.K. Taubenberger, Proc. Natl. Acad. Sci. 96 (1999) ] Sum-product algorithm 4-27
33 challenges in phylogeny phylogeny reconstruction: given DNA sequences at vertices (only at leaves), infer the underlying tree T = (V, E). phylogeny evaluation: given a tree T = (V, E) evaluate the probability of observed DNA sequences at vertices (only at leaves). Bayesian network model for phylogeny evaluation T = (V, D) directed graph, DNA sequences x = (x i ) i V X V µ T (x) = q o (x o ) q i,j (x i, x j ), (i,j) D q i,j (x i, x j ) = Probability that the descendent is x j if ancestor is x i. Sum-product algorithm 4-28
34 simplified model: X = {+1, 1} q o (x o ) = 1 { 2 1 q if x j = x i q(x i, x j ) = q if x i x j MRF representation: q(x i, x j ) e θx ix j with θ = 1 2 log q 1 q probability of certain tree of mutations x: µ T (x) = 1 Z θ (T ) (i,j) E e θx ix j problem: for given T, compute marginal µ T (x i ) we prove the correctness of sum-product algorithm for this model, but the same proof holds for any general pairwise MRF (and also for general MRF and FG) Sum-product algorithm 4-29
35 Sum-product algorithm 4-30 define graphical model on sub-trees j i T j i T j i = (V j i, E j i) Subtree rooted at j and excluding i µ j i(x Vj i ) 1 Z(T j i) ν j i(x j) x Vj i \{j} (u,v) E j i e θxuxv µ j i(x Vj i )
36 T k1 i = (V 1, E 1) T k2 i k 1 i k 2 k 4 k 3 T k4 i T k3 i the messages from neighbors k 1, k 2, k 3, k 4 are sufficient to compute the marginal µ(x i) µ T (x i) e θxuxv = = x V \{i} (u,v) E x V1,x V2,x V3,x V4 l=1 4 l=1 x kl,x Vl \{k l } 4 { l=1 x kl e θxixkl 4 {e θx ix kl {e θx ix kl x Vl \{k l } (u,v) E l e θxuxv (u,v) E l e θxuxv } µ kl i(x Vl ) } {{ } ν kl i(x kl ) Sum-product algorithm 4-31 } }
37 Sum-product algorithm 4-32 recursion on sub-trees to compute the messages ν j i k l T i j µ i j(x Vi j ) = = 1 Z(T i j) (u,v) E i j e θxuxv 1 { Z(T i j) eθx ix k e θx ix l e θxuxv e θx ix l { (u,v) E k i e θxuxv (u,v) E k i e θxuxv }{ e θx ix k e θx ix l µ k i (x Vk i )µ l i (x Vl i ) }{ (u,v) E l i e θxuxv (u,v) E l i e θxuxv } }
38 ν i j j ν k i k i ν l i l ν i j(x i) = µ i j(x Vi j ) x Vi j \i e θx ix k e θx ix l µ k i (x Vk i )µ l i (x Vl i ) = x Vi j \i { }{ } e θxixk µ k i (x Vk i ) e θxixl µ l i (x Vl i ) x Vk i x Vl i { }{ e θxixk µ k i (x Vk i ) e θxixl x k x Vk i \{k} { x k }{ ) x l } ) e θxixk ν k i (x k e θxixl ν l i (x l x l x Vl i \{l} } µ l i (x Vl i ) with uniform initialization ν i j(x i) = 1 X for all leaves i Sum-product algorithm 4-33
39 sum-product algorithm (for our example) ν i j (x i ) { } e θx ix k ν k i (x k ) ν i (x i ) k i\j x k { } e θx ix k ν k i (x k ) k i x k µ T (x i ) = ν i (x i ) x i ν i (x i ) what if we want all the marginals? choose an arbitrary root φ compute all the messages towards the root ( E messages) then compute all the messages outwards from the root ( E messages) then compute all the marginals (n marginals) how many operations are required? naive implementation requires O( X 2 i d2 i ) if i has degree d i, then computing ν i j requires d i X 2 operations d i messages start at each node i, each require d i X 2 operations total computation for 2 E messages is { } i d i (d i X 2 ) however, we can compute all marginals in O(n X 2 ) operations Sum-product algorithm 4-34
40 Sum-product algorithm 4-35 let D = {(i, j), (j, i) (i, j) E} be the directed version of E (cf. D = 2 E ) (sequential) sum-product algorithm 1. initialize ν i j (x i ) = 1/ X for all leaves i 2. recursively over (i, j) D compute (from leaves) { } ν i j (x i ) = ψ ik (x i, x k )ν k i (x k ) x k k i\j 3. for each i V compute marginal ν i (x i ) = { } ψ ik (x i, x k )ν k i (x k ) x k µ T (x i ) = k i ν i (x i ) x i ν i (x i )
41 (parallel) sum-product algorithm 1. initialize ν (0) i j (x i) = 1/ X for all (i, j) D 2. for t {0, 1,..., t max } for all (i, j) D compute ν (t+1) i j (x { } i) = ψ ik (x i, x k )ν (t) k i (x k) k i k i\j 3. for each i V compute marginal ν i (x i ) = { } ψ ik (x i, x k )ν (tmax+1) k i (x k ) µ T (x i ) = x k ν i (x i ) x i ν i (x i ) also called belief propagation when t max is larger than the diameter of the tree (the length of the longest path), this converges to the correct marginal [Homework 2.5] more operations than the sequential version ( O(n X 2 diam(t )) ) a naive implementation requires O( X 2 diam(t ) d 2 i ) naturally extends to general graphs but no proof of exactness Sum-product algorithm 4-36 x k
42 Sum-product algorithm on general graphs (loopy) belief propagation 1. initialize ν i j (x i ) = 1/ X for all (i, j) D 2. for t {0, 1,..., t max } for all (i, j) D compute ν (t+1) i j (x i) = k i k i\j x k { x k } ψ ik (x i, x k )ν (t) k i (x k) 3. for each i V compute marginal ν i (x i ) = { } ψ ik (x i, x k )ν (tmax+1) k i (x k ) µ T (x i ) = ν i (x i ) x i ν i (x i ) computes approximate marginals in O(n X 2 t max ) operations generally it does not converge; even if it does, it might be incorrect folklore about loopy BP works better when G has few short loops works better when ψij (x i, x j ) = ψ ij,1 (x i )ψ ij,2 (x j ) + small(x i, x j ) nonconvex variational principle Sum-product algorithm 4-37
43 Exercise: partition function on trees using the recursion for messages: { } ν i j (x i ) = ψ ik (x i, x k )ν k i (x k ) k i\j x k ν i (x i ) = { } ψ ik (x i, x k )ν k i (x k ) x k k i it follows that we can easily compute the partition function as Z(T ) = ν i (x i ) x i alternatively, if we had a black box that computes marginals for any tree, then we can use it to compute partition functions efficiently Z(T i j ) = { } ψ ik (x i, x k ) Z(T k i ) µ k i (x k ) Z(T ) = x i X k i\j x i X k i x k X { } ψ ik (x i, x k ) Z(T k i ) µ k i (x k ) x k X this recursive algorithm naturally extends to general graphs Sum-product algorithm 4-38
44 Sum-product algorithm 4-39 Why would one want to compute the partition function? Suppose you observe x = (+1, +1, +1, +1, +1, +1, +1, +1, +1) and you know this comes from either of µ(x) = 1 Z(T ) (e.g. coloring) (i,j) E ψ(x i, x j ) which one has highest likelihood?
45 Exercise: sampling on the tree if we have a black-box for computing marginals on any tree, we can use it to sample from any distribution on a tree Sampling( Tree T = (V, E), ψ = {ψ ij } (ij) E ) 1: Choose a root o V ; 2: Sample X o µ o ( ); 2: Recursively over i V (from root to leaves): 3: Compute µ i π(i) (x i x π(i) ); 4: Sample X i µ i π(i) ( x π(i) ); π(i) is the parent of node i in the rooted tree T o we use the black-box to compute the conditional distribution k i j µ T (x Vi j x j ) ψ ij (x i x j ) µ i j (x Vi j ) µ T (x i x j ) ψ ij (x i x j ) µ i j (x i ) Sum-product algorithm 4-40
46 Sum-product algorithm 4-41 Tree decomposition when we don t have a tree we can create an equivalent tree graph by enlarging the alphabet X X k Treewidth(G) Minimum such k it is NP-hard to determine the treewidth of a graph problem: in general Treewidth(G) = Θ(n)
47 Sum-product algorithm 4-42 Tree decomposition of G = (V, E) 1 1,2 1, , , A tree T = (V T, E T ) and a mapping V : V T SUBSETS(V ) s.t.: For each i V there exists at least one u V T with i V (u). For each (i, j) E there exists at least one u V T with i, j V (u). If i V (u 1 ) and i V (u 2 ), then i V (w) for any w on the path between u 1 and u 2 in T.
Course: Model, Learning, and Inference: Lecture 5
Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 yuille@stat.ucla.edu Abstract Probability distributions on structured representation.
More informationThe Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
More information5 Directed acyclic graphs
5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate
More informationSYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis
SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October 17, 2015 Outline
More informationFinding the M Most Probable Configurations Using Loopy Belief Propagation
Finding the M Most Probable Configurations Using Loopy Belief Propagation Chen Yanover and Yair Weiss School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, Israel
More informationApproximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
More informationGRAPH THEORY LECTURE 4: TREES
GRAPH THEORY LECTURE 4: TREES Abstract. 3.1 presents some standard characterizations and properties of trees. 3.2 presents several different types of trees. 3.7 develops a counting method based on a bijection
More informationGraphical Models, Exponential Families, and Variational Inference
Foundations and Trends R in Machine Learning Vol. 1, Nos. 1 2 (2008) 1 305 c 2008 M. J. Wainwright and M. I. Jordan DOI: 10.1561/2200000001 Graphical Models, Exponential Families, and Variational Inference
More informationHidden Markov Models
8.47 Introduction to omputational Molecular Biology Lecture 7: November 4, 2004 Scribe: Han-Pang hiu Lecturer: Ross Lippert Editor: Russ ox Hidden Markov Models The G island phenomenon The nucleotide frequencies
More information3. The Junction Tree Algorithms
A Short Course on Graphical Models 3. The Junction Tree Algorithms Mark Paskin mark@paskin.org 1 Review: conditional independence Two random variables X and Y are independent (written X Y ) iff p X ( )
More informationAnalysis of Algorithms I: Binary Search Trees
Analysis of Algorithms I: Binary Search Trees Xi Chen Columbia University Hash table: A data structure that maintains a subset of keys from a universe set U = {0, 1,..., p 1} and supports all three dictionary
More informationWhy? A central concept in Computer Science. Algorithms are ubiquitous.
Analysis of Algorithms: A Brief Introduction Why? A central concept in Computer Science. Algorithms are ubiquitous. Using the Internet (sending email, transferring files, use of search engines, online
More informationTagging with Hidden Markov Models
Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,
More informationComplexity Theory. IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar
Complexity Theory IE 661: Scheduling Theory Fall 2003 Satyaki Ghosh Dastidar Outline Goals Computation of Problems Concepts and Definitions Complexity Classes and Problems Polynomial Time Reductions Examples
More informationSocial Media Mining. Graph Essentials
Graph Essentials Graph Basics Measures Graph and Essentials Metrics 2 2 Nodes and Edges A network is a graph nodes, actors, or vertices (plural of vertex) Connections, edges or ties Edge Node Measures
More informationApproximating the Partition Function by Deleting and then Correcting for Model Edges
Approximating the Partition Function by Deleting and then Correcting for Model Edges Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles Los Angeles, CA 995
More informationCost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation:
CSE341T 08/31/2015 Lecture 3 Cost Model: Work, Span and Parallelism In this lecture, we will look at how one analyze a parallel program written using Cilk Plus. When we analyze the cost of an algorithm
More informationTree-representation of set families and applications to combinatorial decompositions
Tree-representation of set families and applications to combinatorial decompositions Binh-Minh Bui-Xuan a, Michel Habib b Michaël Rao c a Department of Informatics, University of Bergen, Norway. buixuan@ii.uib.no
More informationScheduling Shop Scheduling. Tim Nieberg
Scheduling Shop Scheduling Tim Nieberg Shop models: General Introduction Remark: Consider non preemptive problems with regular objectives Notation Shop Problems: m machines, n jobs 1,..., n operations
More informationSolutions to Homework 6
Solutions to Homework 6 Debasish Das EECS Department, Northwestern University ddas@northwestern.edu 1 Problem 5.24 We want to find light spanning trees with certain special properties. Given is one example
More informationHidden Markov Models in Bioinformatics. By Máthé Zoltán Kőrösi Zoltán 2006
Hidden Markov Models in Bioinformatics By Máthé Zoltán Kőrösi Zoltán 2006 Outline Markov Chain HMM (Hidden Markov Model) Hidden Markov Models in Bioinformatics Gene Finding Gene Finding Model Viterbi algorithm
More informationStructured Learning and Prediction in Computer Vision. Contents
Foundations and Trends R in Computer Graphics and Vision Vol. 6, Nos. 3 4 (2010) 185 365 c 2011 S. Nowozin and C. H. Lampert DOI: 10.1561/0600000033 Structured Learning and Prediction in Computer Vision
More informationLECTURE 4. Last time: Lecture outline
LECTURE 4 Last time: Types of convergence Weak Law of Large Numbers Strong Law of Large Numbers Asymptotic Equipartition Property Lecture outline Stochastic processes Markov chains Entropy rate Random
More informationAnalysis of Algorithms, I
Analysis of Algorithms, I CSOR W4231.002 Eleni Drinea Computer Science Department Columbia University Thursday, February 26, 2015 Outline 1 Recap 2 Representing graphs 3 Breadth-first search (BFS) 4 Applications
More informationGraphical degree sequences and realizations
swap Graphical and realizations Péter L. Erdös Alfréd Rényi Institute of Mathematics Hungarian Academy of Sciences MAPCON 12 MPIPKS - Dresden, May 15, 2012 swap Graphical and realizations Péter L. Erdös
More informationChapter 7: Termination Detection
Chapter 7: Termination Detection Ajay Kshemkalyani and Mukesh Singhal Distributed Computing: Principles, Algorithms, and Systems Cambridge University Press A. Kshemkalyani and M. Singhal (Distributed Computing)
More informationSEQUENCES OF MAXIMAL DEGREE VERTICES IN GRAPHS. Nickolay Khadzhiivanov, Nedyalko Nenov
Serdica Math. J. 30 (2004), 95 102 SEQUENCES OF MAXIMAL DEGREE VERTICES IN GRAPHS Nickolay Khadzhiivanov, Nedyalko Nenov Communicated by V. Drensky Abstract. Let Γ(M) where M V (G) be the set of all vertices
More informationBig Data, Machine Learning, Causal Models
Big Data, Machine Learning, Causal Models Sargur N. Srihari University at Buffalo, The State University of New York USA Int. Conf. on Signal and Image Processing, Bangalore January 2014 1 Plan of Discussion
More informationMaximum Likelihood Graph Structure Estimation with Degree Distributions
Maximum Likelihood Graph Structure Estimation with Distributions Bert Huang Computer Science Department Columbia University New York, NY 17 bert@cs.columbia.edu Tony Jebara Computer Science Department
More informationNetwork (Tree) Topology Inference Based on Prüfer Sequence
Network (Tree) Topology Inference Based on Prüfer Sequence C. Vanniarajan and Kamala Krithivasan Department of Computer Science and Engineering Indian Institute of Technology Madras Chennai 600036 vanniarajanc@hcl.in,
More informationSmall Maximal Independent Sets and Faster Exact Graph Coloring
Small Maximal Independent Sets and Faster Exact Graph Coloring David Eppstein Univ. of California, Irvine Dept. of Information and Computer Science The Exact Graph Coloring Problem: Given an undirected
More information2. (a) Explain the strassen s matrix multiplication. (b) Write deletion algorithm, of Binary search tree. [8+8]
Code No: R05220502 Set No. 1 1. (a) Describe the performance analysis in detail. (b) Show that f 1 (n)+f 2 (n) = 0(max(g 1 (n), g 2 (n)) where f 1 (n) = 0(g 1 (n)) and f 2 (n) = 0(g 2 (n)). [8+8] 2. (a)
More informationSingle machine parallel batch scheduling with unbounded capacity
Workshop on Combinatorics and Graph Theory 21th, April, 2006 Nankai University Single machine parallel batch scheduling with unbounded capacity Yuan Jinjiang Department of mathematics, Zhengzhou University
More informationBayesian networks - Time-series models - Apache Spark & Scala
Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
More informationIntroduced by Stuart Kauffman (ca. 1986) as a tunable family of fitness landscapes.
68 Part II. Combinatorial Models can require a number of spin flips that is exponential in N (A. Haken et al. ca. 1989), and that one can in fact embed arbitrary computations in the dynamics (Orponen 1995).
More informationLecture 1: Course overview, circuits, and formulas
Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik
More informationMonte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February 5, 2013 J. Olsson Monte Carlo-based
More informationClassification/Decision Trees (II)
Classification/Decision Trees (II) Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Right Sized Trees Let the expected misclassification rate of a tree T be R (T ).
More informationFactor Graphs and the Sum-Product Algorithm
498 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 47, NO. 2, FEBRUARY 2001 Factor Graphs and the Sum-Product Algorithm Frank R. Kschischang, Senior Member, IEEE, Brendan J. Frey, Member, IEEE, and Hans-Andrea
More informationIntroduction to Deep Learning Variational Inference, Mean Field Theory
Introduction to Deep Learning Variational Inference, Mean Field Theory 1 Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Center for Visual Computing Ecole Centrale Paris Galen Group INRIA-Saclay Lecture 3: recap
More informationON INDUCED SUBGRAPHS WITH ALL DEGREES ODD. 1. Introduction
ON INDUCED SUBGRAPHS WITH ALL DEGREES ODD A.D. SCOTT Abstract. Gallai proved that the vertex set of any graph can be partitioned into two sets, each inducing a subgraph with all degrees even. We prove
More informationBayesian probability theory
Bayesian probability theory Bruno A. Olshausen arch 1, 2004 Abstract Bayesian probability theory provides a mathematical framework for peforming inference, or reasoning, using probability. The foundations
More informationCHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
More informationA Non-Linear Schema Theorem for Genetic Algorithms
A Non-Linear Schema Theorem for Genetic Algorithms William A Greene Computer Science Department University of New Orleans New Orleans, LA 70148 bill@csunoedu 504-280-6755 Abstract We generalize Holland
More informationHow To Solve A K Path In Time (K)
What s next? Reductions other than kernelization Dániel Marx Humboldt-Universität zu Berlin (with help from Fedor Fomin, Daniel Lokshtanov and Saket Saurabh) WorKer 2010: Workshop on Kernelization Nov
More informationGambling and Data Compression
Gambling and Data Compression Gambling. Horse Race Definition The wealth relative S(X) = b(x)o(x) is the factor by which the gambler s wealth grows if horse X wins the race, where b(x) is the fraction
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationGuessing Game: NP-Complete?
Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES 2. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple
More informationOn the k-path cover problem for cacti
On the k-path cover problem for cacti Zemin Jin and Xueliang Li Center for Combinatorics and LPMC Nankai University Tianjin 300071, P.R. China zeminjin@eyou.com, x.li@eyou.com Abstract In this paper we
More information! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three
More informationAsking Hard Graph Questions. Paul Burkhardt. February 3, 2014
Beyond Watson: Predictive Analytics and Big Data U.S. National Security Agency Research Directorate - R6 Technical Report February 3, 2014 300 years before Watson there was Euler! The first (Jeopardy!)
More information6.852: Distributed Algorithms Fall, 2009. Class 2
.8: Distributed Algorithms Fall, 009 Class Today s plan Leader election in a synchronous ring: Lower bound for comparison-based algorithms. Basic computation in general synchronous networks: Leader election
More informationVariational Mean Field for Graphical Models
Variational Mean Field for Graphical Models CS/CNS/EE 155 Baback Moghaddam Machine Learning Group baback @ jpl.nasa.gov Approximate Inference Consider general UGs (i.e., not tree-structured) All basic
More informationProbabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
More informationChapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling
Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one
More informationNear Optimal Solutions
Near Optimal Solutions Many important optimization problems are lacking efficient solutions. NP-Complete problems unlikely to have polynomial time solutions. Good heuristics important for such problems.
More informationIE 680 Special Topics in Production Systems: Networks, Routing and Logistics*
IE 680 Special Topics in Production Systems: Networks, Routing and Logistics* Rakesh Nagi Department of Industrial Engineering University at Buffalo (SUNY) *Lecture notes from Network Flows by Ahuja, Magnanti
More informationCoding and decoding with convolutional codes. The Viterbi Algor
Coding and decoding with convolutional codes. The Viterbi Algorithm. 8 Block codes: main ideas Principles st point of view: infinite length block code nd point of view: convolutions Some examples Repetition
More informationEx. 2.1 (Davide Basilio Bartolini)
ECE 54: Elements of Information Theory, Fall 00 Homework Solutions Ex.. (Davide Basilio Bartolini) Text Coin Flips. A fair coin is flipped until the first head occurs. Let X denote the number of flips
More informationOn the independence number of graphs with maximum degree 3
On the independence number of graphs with maximum degree 3 Iyad A. Kanj Fenghui Zhang Abstract Let G be an undirected graph with maximum degree at most 3 such that G does not contain any of the three graphs
More informationCommunication Cost in Big Data Processing
Communication Cost in Big Data Processing Dan Suciu University of Washington Joint work with Paul Beame and Paris Koutris and the Database Group at the UW 1 Queries on Big Data Big Data processing on distributed
More informationA Sublinear Bipartiteness Tester for Bounded Degree Graphs
A Sublinear Bipartiteness Tester for Bounded Degree Graphs Oded Goldreich Dana Ron February 5, 1998 Abstract We present a sublinear-time algorithm for testing whether a bounded degree graph is bipartite
More informationOutline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits
Outline NP-completeness Examples of Easy vs. Hard problems Euler circuit vs. Hamiltonian circuit Shortest Path vs. Longest Path 2-pairs sum vs. general Subset Sum Reducing one problem to another Clique
More informationLecture 22: November 10
CS271 Randomness & Computation Fall 2011 Lecture 22: November 10 Lecturer: Alistair Sinclair Based on scribe notes by Rafael Frongillo Disclaimer: These notes have not been subjected to the usual scrutiny
More informationPart 2: Community Detection
Chapter 8: Graph Data Part 2: Community Detection Based on Leskovec, Rajaraman, Ullman 2014: Mining of Massive Datasets Big Data Management and Analytics Outline Community Detection - Social networks -
More informationCMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma
CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma Please Note: The references at the end are given for extra reading if you are interested in exploring these ideas further. You are
More informationSeminar. Path planning using Voronoi diagrams and B-Splines. Stefano Martina stefano.martina@stud.unifi.it
Seminar Path planning using Voronoi diagrams and B-Splines Stefano Martina stefano.martina@stud.unifi.it 23 may 2016 This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International
More informationFinding and counting given length cycles
Finding and counting given length cycles Noga Alon Raphael Yuster Uri Zwick Abstract We present an assortment of methods for finding and counting simple cycles of a given length in directed and undirected
More informationTracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
More informationDistributed Computing over Communication Networks: Maximal Independent Set
Distributed Computing over Communication Networks: Maximal Independent Set What is a MIS? MIS An independent set (IS) of an undirected graph is a subset U of nodes such that no two nodes in U are adjacent.
More informationMapReduce and Distributed Data Analysis. Sergei Vassilvitskii Google Research
MapReduce and Distributed Data Analysis Google Research 1 Dealing With Massive Data 2 2 Dealing With Massive Data Polynomial Memory Sublinear RAM Sketches External Memory Property Testing 3 3 Dealing With
More informationStatistical machine learning, high dimension and big data
Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,
More information1 Homework 1. [p 0 q i+j +... + p i 1 q j+1 ] + [p i q j ] + [p i+1 q j 1 +... + p i+j q 0 ]
1 Homework 1 (1) Prove the ideal (3,x) is a maximal ideal in Z[x]. SOLUTION: Suppose we expand this ideal by including another generator polynomial, P / (3, x). Write P = n + x Q with n an integer not
More informationONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN. Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015
ONLINE DEGREE-BOUNDED STEINER NETWORK DESIGN Sina Dehghani Saeed Seddighin Ali Shafahi Fall 2015 ONLINE STEINER FOREST PROBLEM An initially given graph G. s 1 s 2 A sequence of demands (s i, t i ) arriving
More informationInformation Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
More information! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
More informationA Practical Scheme for Wireless Network Operation
A Practical Scheme for Wireless Network Operation Radhika Gowaikar, Amir F. Dana, Babak Hassibi, Michelle Effros June 21, 2004 Abstract In many problems in wireline networks, it is known that achieving
More informationOn an anti-ramsey type result
On an anti-ramsey type result Noga Alon, Hanno Lefmann and Vojtĕch Rödl Abstract We consider anti-ramsey type results. For a given coloring of the k-element subsets of an n-element set X, where two k-element
More informationLecture 7: NP-Complete Problems
IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 7: NP-Complete Problems David Mix Barrington and Alexis Maciel July 25, 2000 1. Circuit
More informationLarge induced subgraphs with all degrees odd
Large induced subgraphs with all degrees odd A.D. Scott Department of Pure Mathematics and Mathematical Statistics, University of Cambridge, England Abstract: We prove that every connected graph of order
More informationDATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
More informationCoding Theorems for Turbo-Like Codes Abstract. 1. Introduction.
Coding Theorems for Turbo-Like Codes Dariush Divsalar, Hui Jin, and Robert J. McEliece Jet Propulsion Laboratory and California Institute of Technology Pasadena, California USA E-mail: dariush@shannon.jpl.nasa.gov,
More informationProtein Protein Interaction Networks
Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks Young-Rae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationMessage-passing sequential detection of multiple change points in networks
Message-passing sequential detection of multiple change points in networks Long Nguyen, Arash Amini Ram Rajagopal University of Michigan Stanford University ISIT, Boston, July 2012 Nguyen/Amini/Rajagopal
More informationInternational Journal of Advanced Research in Computer Science and Software Engineering
Volume 3, Issue 7, July 23 ISSN: 2277 28X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Greedy Algorithm:
More informationLife of A Knowledge Base (KB)
Life of A Knowledge Base (KB) A knowledge base system is a special kind of database management system to for knowledge base management. KB extraction: knowledge extraction using statistical models in NLP/ML
More informationMonte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)
Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February
More informationHandout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010. Chapter 7: Digraphs
MCS-236: Graph Theory Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010 Chapter 7: Digraphs Strong Digraphs Definitions. A digraph is an ordered pair (V, E), where V is the set
More informationEfficient Recovery of Secrets
Efficient Recovery of Secrets Marcel Fernandez Miguel Soriano, IEEE Senior Member Department of Telematics Engineering. Universitat Politècnica de Catalunya. C/ Jordi Girona 1 i 3. Campus Nord, Mod C3,
More informationHidden Markov Models Chapter 15
Hidden Markov Models Chapter 15 Mausam (Slides based on Dan Klein, Luke Zettlemoyer, Alex Simma, Erik Sudderth, David Fernandez-Baca, Drena Dobbs, Serafim Batzoglou, William Cohen, Andrew McCallum, Dan
More informationA 2-factor in which each cycle has long length in claw-free graphs
A -factor in which each cycle has long length in claw-free graphs Roman Čada Shuya Chiba Kiyoshi Yoshimoto 3 Department of Mathematics University of West Bohemia and Institute of Theoretical Computer Science
More informationEvery tree contains a large induced subgraph with all degrees odd
Every tree contains a large induced subgraph with all degrees odd A.J. Radcliffe Carnegie Mellon University, Pittsburgh, PA A.D. Scott Department of Pure Mathematics and Mathematical Statistics University
More informationMonte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 6 Sequential Monte Carlo methods II February 7, 2014 M. Wiktorsson
More informationStatistical Machine Translation: IBM Models 1 and 2
Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation
More informationDynamic Programming and Graph Algorithms in Computer Vision
Dynamic Programming and Graph Algorithms in Computer Vision Pedro F. Felzenszwalb and Ramin Zabih Abstract Optimization is a powerful paradigm for expressing and solving problems in a wide range of areas,
More informationBayesian Networks. Mausam (Slides by UW-AI faculty)
Bayesian Networks Mausam (Slides by UW-AI faculty) Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential space for representation & inference BNs provide
More informationMethods for big data in medical genomics
Methods for big data in medical genomics Parallel Hidden Markov Models in Population Genetics Chris Holmes, (Peter Kecskemethy & Chris Gamble) Department of Statistics and, Nuffield Department of Medicine
More informationPolarization codes and the rate of polarization
Polarization codes and the rate of polarization Erdal Arıkan, Emre Telatar Bilkent U., EPFL Sept 10, 2008 Channel Polarization Given a binary input DMC W, i.i.d. uniformly distributed inputs (X 1,...,
More informationExponential time algorithms for graph coloring
Exponential time algorithms for graph coloring Uriel Feige Lecture notes, March 14, 2011 1 Introduction Let [n] denote the set {1,..., k}. A k-labeling of vertices of a graph G(V, E) is a function V [k].
More information