Graphical Models and Message passing
|
|
- Theresa Perkins
- 7 years ago
- Views:
Transcription
1 Graphical Models and Message passing Yunshu Liu ASPITRG Research Group
2 References: [1]. Steffen Lauritzen, Graphical Models, Oxford University Press, 1996 [2]. Michael I. Jordan, Graphical models, Statistical Science, Vol.19, No. 1. Feb., 2004 [3]. Christopher M. Bishop, Pattern Recognition and Machine Learning, Springer-Verlag New York, Inc [4]. Daphne Koller and Nir Friedman, Probabilistic Graphical Models - Principles and Techniques, The MIT Press, 2009 [5].Kevin P. Murphy, Machine Learning - A Probabilistic Perspective, The MIT Press, 2012
3 Outline Preliminaries on Graphical Models Directed graphical model Undirected graphical model Directed graph and Undirected graph Message passing and sum-product algorithm Factor graphs Sum-product algorithms Junction tree algorithm
4 Outline Preliminaries on Graphical Models Message passing and sum-product algorithm
5 Preliminaries on Graphical Models Motivation: Curse of dimensionality Matroid enumeration Polyhedron computation Entropic Region Machine Learning: computing likelihoods, marginal probabilities Graphical Models: Help capturing complex dependencies among random variables, building large-scale statistical models and designing efficient algorithms for Inference.
6 Preliminaries on Graphical Models Definition of Graphical Models: A graphical model is a probabilistic model for which a graph denotes the conditional dependence structure between random variables. Example: Suppose MIT and Stanford accepted undergraduate students only based on GPA MIT : Accepted by MIT Stanford : Accepted by Stanford GPA MIT Stanford Given Alice s GPA as GP A Alice, P(MIT Stanford, GP A Alice )=P(MIT GP A Alice ) We say MIT is conditionally independent of Stanford given GP A Alice Sometimes use symbol (MIT Stanford GP A Alice )
7 Bayesian networks: directed graphical model Bayesian networks A Bayesian network consists of a collection of probability distributions P over x = {x 1,, x K } that factorize over a directed acyclic graph(dag) in the following way: p(x) = p(x 1,, x K ) = k K p(x k pa k ) where pa k is the direct parents nodes of x k. Alias of Bayesian networks: probabilistic directed graphical model: via directed acyclic graph(dag) belief networks causal networks: directed arrows represent causal realtions
8 Bayesian networks: directed graphical model Examples of Bayesian networks Consider an arbitrary joint distribution p(x) = p(x 1, x 2, x 3 ) over three variables, we can write: p(x 1, x 2, x 3 ) = p(x 3 x 1, x 2 )p(x 1, x 2 ) = p(x 3 x 1, x 2 )p(x 2 x 1 )p(x 1 ) which can be expressed in the following directed graph: x 1 x 3 x 2
9 Bayesian networks: directed graphical model Examples Similarly, if we change the order of x 1, x 2 and x 3 (same as consider all permutations of them), we can express p(x 1, x 2, x 3 ) in five other different ways, for example: p(x 1, x 2, x 3 ) = p(x 1 x 2, x 3 )p(x 2, x 3 ) = p(x 1 x 2, x 3 )p(x 2 x 3 )p(x 3 ) which corresponding to the following directed graph: x 1 x 3 x 2
10 Bayesian networks: directed graphical model Examples Recall the previous example about how MIT and Stanford accept undergraduate students, if we assign x 1 to GPA, x 2 to accepted by MIT and x 3 to accepted by Stanford, then since p(x 3 x 1, x 2 ) = p(x 3 x 1 ) we have p(x 1, x 2, x 3 ) = p(x 3 x 1, x 2 )p(x 2 x 1 )p(x 1 ) = p(x 3 x 1 )p(x 2 x 1 )p(x 1 ) which corresponding to the following directed graph: x 1 GPA x 3 Stanford x 2 MIT
11 ,x 4 } is not a clique because of the missing link from x 1 to x 4. re Markov define the random factors infields: the decomposition undirected of graphical the joint distribution model the variables In the undirected in the cliques. case, theinprobability fact, we can distribution consider factorizes functions ques, according without loss to functions of generality, definedbecause on the clique otherof cliques the graph. must be l cliques. A clique Thus, is aif subset {x 1,x 2 of,xnodes 3 } is ainmaximal a graph such clique that and there we define exist a on over linkthis between clique, allthen pairsincluding of nodes inanother subset. factor defined over a iablesawould maximal be redundant. clique is a clique such that it is not possible to a clique by C and the set of variables in that clique by x C. Then include any other nodes from the graph in the set without it ceasing to be a clique. ted graph showing a clique (outlined in al clique (outlined in blue). x 1 Example of cliques: x 2 {x 1,x 2 },{x 2,x 3 },{x 3,x 4 },{x 2,x 4 },{x 1,x 3 }, {x 1,x 2,x 3 }, {x 2,x 3,x 4 } Maximal cliques: {x 1,x 2,x 3 }, {x 2,x 3,x 4 } x 3 x 4
12 Markov random fields: undirected graphical model Markov random fields: Definition Denote C as a clique, x C the set of variables in clique C and ψ C (x C ) a nonnegative potential function associated with clique C. Then a Markov random field is a collection of distributions that factorize as a product of potential functions ψ C (x C ) over the maximal cliques of the graph: p(x) = 1 ψ Z C (x C ) where normalization constant Z = x called the partition function. C C ψ C(x C ) sometimes
13 Markov random fields: undirected graphical model Factorization of undirected graphs Question: how to write the joint distribution for this undirected graph? x 1 x 3 x 2 (2 3 1) hold Answer: p(x) = 1 Z ψ 12(x 1, x 2 )ψ 13 (x 1, x 3 ) where ψ 12 (x 1, x 2 ) and ψ 13 (x 1, x 3 ) are the potential functions and Z is the partition function that make sure p(x) satisfy the conditions to be a probability distribution.
14 Markov random fields: undirected graphical model Markov property Given an undirected graph G = (V,E), a set of random variables X = (X a ) a V indexed by V, we have the following Markov properties: Pairwise Markov property: Any two non-adjacent variables are conditionally independent given all other variables: X a X b X V \{a,b} if {a, b} E Local Markov property: A variable is conditionally independent of all other variables given its neighbors: X a X V \{nb(a) a} X nb(a) where nb(a) is the neighbors of node a. Global Markov property: Any two subsets of variables are conditionally independent given a separating subset: X A X B X S, where every path from a node in A to a node in B passes through S(means when we remove all the nodes in S, there are no paths connecting any nodes in A to any nodes in B).
15 Markov random fields: undirected graphical model G. The dotted red node X 8 is independent of all other clude Examples its parents of Markov (blue), properties children (green) and co-parents GM. Pairwise The redmarkov nodeproperty: X ( ), ( ) 8 is independent of the other black Local Markov property: ( ), ( ) Global Markov property: ( ), ( ) (b)
16 Markov random fields: undirected graphical model Relationship between different Markov properties and factorization property ( F ): Factorization property; ( G ): Global Markov property; ( L ): Local Markov property; ( P ):Pairwise Markov property if assuming strictly positive p( ) (F) (G) (L) (P) (P) (F) which give us the Hammersley-Clifford theorem.
17 Markov random fields: undirected graphical model The Hammersley-Clifford theorem(see Koller and Friedman 2009, p131 for proof) Consider graph G, for strictly positive p( ), the following Markov property and Factorization property are equivalent: Markov property: Any two subsets of variables are conditionally independent given a separating subset (X A, X B X S ) where every path from a node in A to a node in B passes through S. Factorization property: The distribution p factorizes according to G if it can be expressed as a product of potential functions over maximal cliques.
18 Markov random fields: undirected graphical model Example: Gaussian Markov random fields A multivariate normal distribution forms a Markov random field w.r.t. a graph G = (V, E) if the missing edges correspond to zeros on the concentration matrix(the inverse convariance matrix) Consider x = {x 1,, x K } N (0, Σ) with Σ regular and K = Σ 1. The concentration matrix of the conditional distribution of (x i, x j ) given x \{i,j} is ( ) kii k K {i,j} = ij k ji k jj Hence x i x j x \ij k ij = 0 {i, j} E
19 Markov random fields: undirected graphical model Example: Gaussian Markov random fields The joint distribution of gaussian markov random fields can be factorizes as: log p(x) = const 1 2 i k ii x 2 i {i,j} E k ij x i x j The zero entries in K are called structural zeros since they represent the absent edges in the undirected graph.
20 X 11 X 12 X 13 X 14 X 15 Directed graph and Undirected graph X 16 X 17 X 18 X 19 X 20 X 16 X 17 X 18 X 19 X 20 (a) X 11 X 12 X 13 X 14 X 15 Moral graph: from directed graph to undirected graph Figure 19.1 Moralization: marrying the parents, so to speak, force the parents of a node to form a completed sub-graph. (a) A 2d lattice represented as a DAG. The dotted red node X 8 is independent of all other nodes (black) given its Markov blanket, which include its parents (blue), children (green) and co-parents (orange). (b) The same model represented as a UGM. The red node X 8 is independent of the other black nodes given its neighbors (blue nodes). (b) (a) (b) Figure 19.2 (a) A DGM. (b) Its moralized version, represented as a UGM. Motivation: Why not simply convert directed graph to undirected graph? Works fine for Markov chain, not work for when there are nodes that have more than one parents(previous example). The set of nodes that renders a node t conditionally independent of all the other nodes in the graph is called t s Markov blanket; we will denote this by mb(t). Formally, the Markov blanket satisfies the following property: t V\cl(t) mb(t) (19.1) where cl(t) mb(t) {t} is the closure of node t. One can show that, in a UGM, a node s
21 Directed graph and Undirected graph Markov blanket In directed graph, a Markov blanket of a given node x i is the set of nodes comprising the parents, the children and the co-parents(other parents of its children) of x i. In undirected graph, a Markov blanket of a given node x i is the set of 662its neighboring nodes. Chapter 19. Undirected graphical models (Markov random fields) X 1 X 2 X 3 X 4 X 5 X 1 X 2 X 3 X 4 X 5 X 6 X 7 X 8 X 9 X 10 X 6 X 7 X 8 X 9 X 10 X 11 X 12 X 13 X 14 X 15 X 11 X 12 X 13 X 14 X 15 X 16 X 17 X 18 X 19 X 20 (a) X 16 X 17 X 18 X 19 X 20 (b) Figure 19.1 (a) A 2d lattice represented as a DAG. The dotted red node X 8 is independent of all other nodes (black) given its Markov blanket, which include its parents (blue), children (green) and co-parents (orange). (b) The same model represented as a UGM. The red node X 8 is independent of the other black nodes given its neighbors (blue nodes).
22 Directed graph and Undirected graph Markov blanket In directed graph Let x = {x 1,, x K } p(x i x \i ) = p(x 1,, x K ) x i p(x 1,, x K ) k = p(x k pa k ) x i k p(x k pa k ) = f (x i, mb(x i )) where pa k is the parents nodes of x k, mb(x i ) is the markov blanket of node x i. Example: p(x 8 X \8 ) = p(x 8 X 3,X 7 ) p(x 9 X 4,X 8 )p(x 13 X 8,X 12 ) X 8 p(x 8 X 3,X 7 )p(x 9 X 4,X 8 )p(x 13 X 8,X 12 )
23 Directed graph and Undirected graph Markov blanket In undirected graph Let x = {x 1,, x K }, p(x i x \i ) = p(x i mb(x i )) where mb(x i ) is the markov blanket of node x i. Example: p(x 8 X \8 ) = p(x 8 X 3, X 7, X 9, X 13 ) Notes The Markov blanket of a directed graph is the Markov blanket(neighbours) of its moral graph.
24 Directed graph and Undirected graph Directed graph VS. Undirected graph All distributions Directed graph Chordal Undirected graph x 1 x 1 x 1 x 3 x 2 x 4 x 3 x2 x 3 x 2 What is Chordal?
25 Directed graph and Undirected graph What is Chordal? Chordal or decomposable: If we collapse together all the variables in each maximal clique to make the mega-variables, the resulting graph will be a tree. Why Chordal graphs are the intersection of directed graph and undirected graph? Hint: Moral graph
26 Directed graph and Undirected graph Directed Graph:Dropping edges x 1 (2 3 1) x 3 x 2 x 1 x 1 (1 3 2) x 3 x 2 x 3 x 2 x 1 (1 2 ) But not (1 2 3) x 3 x 2
27 Directed graph and Undirected graph Undirected Graphs: remove arrows x 1 x 3 x 2 x 1 x 1 x 1 x 3 x 2 x 3 x 2 x 3 x 2 (2 3 1) hold x 1 x 3 x 2
28 Directed graph and Undirected graph Directed graph and Undirected graph Example of 4 atom distribution: x x 4 x 3 x 2 x 1 p(x) When I(x 3 ; x 4 )= a b-a c-a 1+a-b-c αβ α(1 β) (1 α)β (1 α)(1 β) Satisfy following Conditional independence relations: (1, 2 3), (1, 2 4) (1, 2 34), (1 34), (2 34) (1 234), (2 134), (3 124), (4 123) (3, 4 )
29 Outline Preliminaries on Graphical Models Message passing and sum-product algorithm
30 Factor graphs Common thing about directed graphical model and undirected graphical model Both allow global function to be expressed as product of factors over subsets of variables. Factor graphs: Definition A factor graph is a bipartite(bigraph) that expresses the structure of the factorization (1). p(x) = s f s (x s ) (1) where x s denotes a subset of the variables x. In a factor graph, the random variables usually are round nodes and the factors are square nodes. There is an edge between the factor node f s (x s ) and the variable node x k if and only if x i x s.
31 Factor graphs Example of factor graph: Factor graph: connection only between variable nodes and factor nodes. Variable nodes X 1 X 2 X 3 X 4 X 5 Factor nodes F a F b F c F d p(x) =f a (x 1,x 3 )f b (x 3,x 4 )f c (x 2,x 4,x 5 )f d (x 1,x 3 ) Note: The same subset of variables x s can be repeated multiple times, for example: x s = {x 1, x 3 } in the above factor graph.
32 Factor graphs From directed graph to factor graph: x 1 x 2 x 1 x 2 x 1 x 2 f f c x 3 f x a 3 x3 (a) (b) (c) f b (a ): An directed graph with factorization p(x 1 )p(x 2 )p(x 3 x 1, x 2 ). (b ): A factor graph with factor f (x 1, x 2, x 3 ) = p(x 1 )p(x 2 )p(x 3 x 1, x 2 ). (c ): A different factor graph with factors f a (x 1 )f b (x 2 )f c (x 1, x 2, x 3 ) such that f a (x 1 ) = p(x 1 ), f b (x 2 ) = p(x 2 ) and f c (x 1, x 2, x 3 ) = p(x 3 x 1, x 2 ).
33 Factor graphs From undirected graph to factor graph: difference between 1 Z C ψ C(x C ) and x f s(x s ) x 1 x 1 x 2 x 3 x 1 x 2 x 3 x 2 x 3 f f a f b (a) (b) (c) (a ): An undirected graph with a single clique potential ψ(x 1, x 2, x 3 ). (b ): A factor graph with factor f (x 1, x 2, x 3 ) = ψ(x 1, x 2, x 3 ). (c ): A different factor graph with factors f a (x 1, x 2, x 3 )f b (x 1, x 2 ) = ψ(x 1, x 2, x 3 ) Note: When clique is large, further factorize is usually very useful.
34 Factor graphs Why introduce Factor graphs? Both directed graph and undirected graph can be converted to factor graphs. Factor graphs allows us to be more explicit about the details of the factorization. When connect factors to exponential families, the product of factors becomes sum of parameters under single exp. Multiple different factor graphs can represent the same directed or undirected graph, we can find factor graphs that are better for our inference.
35 Sum-product algorithms Compute marginal probability p(x i ) p(x i ) = x \xi p(x) What if p(x) = 1 Z C ψ C(x C ) or x f s(x s )? Can we design more efficient algorithms by exploring the conditional independence structure of the model?
36 Sum-product algorithms Yes, we can. The key idea of Sum-product algorithm xy + xz = x(y + z) + x + y x z y z Expression tree of xy + xz and x(y+z), where the leaf nodes represent variables and internal nodes represent arithmetic operators(addition, multiplication, negation, etc.).
37 Sum-product algorithms Example of computing p(x 1 ) = x \x1 p(x) on an undirected chain p(x) = 1 Z ψ 12(x 1,x 2 )ψ 23 (x 2,x 3 )ψ 34 (x 3,x 4 ) x 1 x 2 x 3 x 4 m 2 (x 1 ) m 3 (x 2 ) m 4 (x 3 ) p(x 1 ) = 1 Z = 1 Z = 1 Z = 1 Z x 2 x 3 ψ 12 (x 1, x 2 )ψ 23 (x 2, x 3 )ψ 34 (x 3, x 4 ) x 4 ψ 12 (x 1, x 2 ) ψ 23 (x 2, x 3 ) ψ 34 (x 3, x 4 ) x 2 x 3 x 4 ψ 12 (x 1, x 2 ) ψ 23 (x 2, x 3 )m 4 (x 3 ) x 2 x 3 ψ 12 (x 1, x 2 )m 3 (x 2 ) = 1 Z m 2(x 1 ) x 2
38 Sum-product algorithms Example of computing p(x 1 ) = x \x1 p(x) Factor graph: connection only between variable nodes and factor nodes. Variable nodes X 1 X 2 X 3 X 4 X 5 Factor nodes F a F b F c F d p(x) =f a (x 1,x 3 )f b (x 3,x 4 )f c (x 2,x 4,x 5 )f d (x 1,x 3 ) p(x 1 ) = x \x1 p(x) = x 2 p(x) x 3 x 4 x 5 = x 2,x 3 f a (x 1, x 3 )f d (x 1, x 3 ) x 4 f b (x 3, x 4 ) x 5 f c (x 2, x 4, x 5 )
39 Sum-product algorithms Sum-product algorithm The sum-product algorithm involves using a message passing scheme to explore the conditional independence and use these properties to change the order of sum and product.
40 Sum-product algorithms Message passing on a factor graph: Variable x to factor f : m x f = m fk x(x) k nb(x)\f f 1 m f1 x x m x f f m fk x f K nb(x) denote the neighboring factor nodes of x If nb(x)\f = then m x f = 1.
41 Sum-product algorithms Message passing on a factor graph: Factor f to variable x: m f x = f (x, x 1,, x M ) x 1, x M m nb(f )\x m xm f (x m ) x 1 m x1 f f m f x x x M m xm f nb(f ) denote the neighboring variable nodes of f If nb(f )\x = then m f x = f (x).
42 Sum-product algorithms Message passing on a tree structured factor graph x 3 f b x 2 f a f c x 1 x 4
43 Sum-product algorithms Message passing on a tree structured factor graph Down - up: m x1 f a (x 1 ) = 1 m fa x 2 (x 2 ) = x 1 f a (x 1, x 2 ) m x4 f c (x 4 ) = 1 m fc x 2 (x 2 ) = x 4 f c (x 2, x 4 ) m x2 f b (x 2 ) = m fa x 2 (x 2 )m fc x 2 (x 2 ) m fb x 3 (x 3 ) = x 2 f b (x 2, x 3 )m x2 f b
44 Sum-product algorithms Message passing on a tree structured factor graph Up - down: m x3 f b (x 3 ) = 1 m fb x 2 (x 2 ) = x 3 f b (x 2, x 3 ) m x2 f a (x 2 ) = m fb x 2 (x 2 )m fc x 2 (x 2 ) m fa x 1 (x 1 ) = x 2 f a (x 1, x 2 )m x2 f a (x 2 ) m x2 f c (x 2 ) = m fa x 2 (x 2 )m fb x 2 (x 2 ) m fc x 4 (x 4 ) = x 2 f c (x 2.x 4 )m x2 f c (x 2 )
45 Sum-product algorithms Message passing on a tree structured factor graph Calculating p(x 2 ) p(x 2 ) = 1 Z m f a x 2 (x 2 )m fb x 2 (x 2 )m fc x2 (x 2 ) = 1 { }{ }{ } f a (x 1, x 2 ) f b (x 2, x 3 ) f c (x 2, x 4 ) Z x 1 x 3 x 4 = 1 f a (x 1, x 2 )f b (x 2, x 3 )f c (x 2, x 4 ) Z x 1 x 3 x 4 = p(x) x 1 x 3 x 4
46 Junction tree algorithm Junction tree algorithm: message passing on general graphs 1. Moralize to a undirected graph if starting with a directed graph 2. Triangulate the graph: to make the graph chordal 3. Build the junction tree 4. Run message passing algorithm on the junction tree
47 Junction tree algorithm 662 Chapter 19. Undirected graphical models (Markov random fields) X1 X2 X3 X4 X5 X1 X2 X3 X4 X5 Junction tree algorithm X6 X7 X8 X9 X10 X6 X7 X8 X9 X10 X11 X12 X13 X14 X15 1. Moralize to a undirected graph if starting with a directed X16 X17 X18 X19 X20 X16 X17 X18 X19 X20 graph (a) X11 X12 X13 X14 X15 Moralization: Figure 19.1 marrying (a) A 2d lattice represented theasparents, a DAG. The dotted red sonode tox8 speak, is independentforce of all otherthe nodes (black) given its Markov blanket, which include its parents (blue), children (green) and co-parents parents of (orange). a (b) node The sameto model form represented aas completed a UGM. The red node X8 sub-graph. is independent of the other black nodes given its neighbors (blue nodes). (b) (a) (b) Figure 19.2 (a) A DGM. (b) Its moralized version, represented as a UGM. The set of nodes that renders a node t conditionally independent of all the other nodes in the graph is called t s Markov blanket; we will denote this by mb(t). Formally, the Markov blanket satisfies the following property: t V\cl(t) mb(t) (19.1)
48 Junction tree algorithm Junction tree algorithm 2. Triangulate the graph: to make the graph chordal Finding chord-less cycles containing four or more nodes and adding extra links to eliminate such chord-less cycles. x 1 x 1 x 5 x 2 x 5 x 2 x 4 x 3 x 4 x 3
49 Junction tree algorithm Junction tree algorithm 3. Build the junction tree The nodes of a junction tree are the maximal cliques of the triangulated graph, The links connect pairs of nodes(maximal cliques) that have variables in common. The junction tree is build so that the summation of the edge-weight in the tree is maximal, where the weight of an edge is the number of variables shared by the two maximal cliques it connected. x 1 x 1,x 2,x 5 x 1,x 2,x 5 x 5 x 2 x 5 x 3,x 4,x 5 x 2,x 5 x 2,x 3,x 5 x 3,x 5 x 3,x 5 x 4 x 3 x 2,x 3,x 5 No x 3,x 4,x 5 Yes
50 Junction tree algorithm Junction tree algorithm 4. Run message passing algorithm on the junction tree The message passing algorithm on the junction tree is similarly to the sum-product algorithm on a tree, the result give the joint distribution of each maximal clique of the triangulated graph. The computational complexity of the junction tree algorithm is determined by the size of the largest clique, which is lower bounded by the treewidth of the graph,where treewidth is one less than the size of the largest clique.
51 Thanks! Questions!
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
More information3. The Junction Tree Algorithms
A Short Course on Graphical Models 3. The Junction Tree Algorithms Mark Paskin mark@paskin.org 1 Review: conditional independence Two random variables X and Y are independent (written X Y ) iff p X ( )
More information5 Directed acyclic graphs
5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate
More informationCourse: Model, Learning, and Inference: Lecture 5
Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 yuille@stat.ucla.edu Abstract Probability distributions on structured representation.
More informationMarkov random fields and Gibbs measures
Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).
More informationPractice with Proofs
Practice with Proofs October 6, 2014 Recall the following Definition 0.1. A function f is increasing if for every x, y in the domain of f, x < y = f(x) < f(y) 1. Prove that h(x) = x 3 is increasing, using
More informationScheduling Shop Scheduling. Tim Nieberg
Scheduling Shop Scheduling Tim Nieberg Shop models: General Introduction Remark: Consider non preemptive problems with regular objectives Notation Shop Problems: m machines, n jobs 1,..., n operations
More informationConditional Random Fields: An Introduction
Conditional Random Fields: An Introduction Hanna M. Wallach February 24, 2004 1 Labeling Sequential Data The task of assigning label sequences to a set of observation sequences arises in many fields, including
More informationBayesian networks - Time-series models - Apache Spark & Scala
Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
More informationLecture 16 : Relations and Functions DRAFT
CS/Math 240: Introduction to Discrete Mathematics 3/29/2011 Lecture 16 : Relations and Functions Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT In Lecture 3, we described a correspondence
More informationApproximating the Partition Function by Deleting and then Correcting for Model Edges
Approximating the Partition Function by Deleting and then Correcting for Model Edges Arthur Choi and Adnan Darwiche Computer Science Department University of California, Los Angeles Los Angeles, CA 995
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
More informationLife of A Knowledge Base (KB)
Life of A Knowledge Base (KB) A knowledge base system is a special kind of database management system to for knowledge base management. KB extraction: knowledge extraction using statistical models in NLP/ML
More informationStructured Learning and Prediction in Computer Vision. Contents
Foundations and Trends R in Computer Graphics and Vision Vol. 6, Nos. 3 4 (2010) 185 365 c 2011 S. Nowozin and C. H. Lampert DOI: 10.1561/0600000033 Structured Learning and Prediction in Computer Vision
More informationSYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis
SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October 17, 2015 Outline
More informationLecture 1: Course overview, circuits, and formulas
Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik
More informationBayesian Networks. Mausam (Slides by UW-AI faculty)
Bayesian Networks Mausam (Slides by UW-AI faculty) Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential space for representation & inference BNs provide
More informationStatistical machine learning, high dimension and big data
Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationa 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More informationA linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form
Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...
More informationUNCOUPLING THE PERRON EIGENVECTOR PROBLEM
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM Carl D Meyer INTRODUCTION Foranonnegative irreducible matrix m m with spectral radius ρ,afundamental problem concerns the determination of the unique normalized
More informationLECTURE 4. Last time: Lecture outline
LECTURE 4 Last time: Types of convergence Weak Law of Large Numbers Strong Law of Large Numbers Asymptotic Equipartition Property Lecture outline Stochastic processes Markov chains Entropy rate Random
More information1. Prove that the empty set is a subset of every set.
1. Prove that the empty set is a subset of every set. Basic Topology Written by Men-Gen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since
More informationLecture 7: NP-Complete Problems
IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 7: NP-Complete Problems David Mix Barrington and Alexis Maciel July 25, 2000 1. Circuit
More informationDiscrete Mathematics & Mathematical Reasoning Chapter 10: Graphs
Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Kousha Etessami U. of Edinburgh, UK Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 6) 1 / 13 Overview Graphs and Graph
More informationPart 2: Community Detection
Chapter 8: Graph Data Part 2: Community Detection Based on Leskovec, Rajaraman, Ullman 2014: Mining of Massive Datasets Big Data Management and Analytics Outline Community Detection - Social networks -
More informationCOMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact
More informationDATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
More information2.1 Complexity Classes
15-859(M): Randomized Algorithms Lecturer: Shuchi Chawla Topic: Complexity classes, Identity checking Date: September 15, 2004 Scribe: Andrew Gilpin 2.1 Complexity Classes In this lecture we will look
More information! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationTOPIC 4: DERIVATIVES
TOPIC 4: DERIVATIVES 1. The derivative of a function. Differentiation rules 1.1. The slope of a curve. The slope of a curve at a point P is a measure of the steepness of the curve. If Q is a point on the
More informationFormal Languages and Automata Theory - Regular Expressions and Finite Automata -
Formal Languages and Automata Theory - Regular Expressions and Finite Automata - Samarjit Chakraborty Computer Engineering and Networks Laboratory Swiss Federal Institute of Technology (ETH) Zürich March
More informationLarge Margin DAGs for Multiclass Classification
S.A. Solla, T.K. Leen and K.-R. Müller (eds.), 57 55, MIT Press (000) Large Margin DAGs for Multiclass Classification John C. Platt Microsoft Research Microsoft Way Redmond, WA 9805 jplatt@microsoft.com
More informationTree-representation of set families and applications to combinatorial decompositions
Tree-representation of set families and applications to combinatorial decompositions Binh-Minh Bui-Xuan a, Michel Habib b Michaël Rao c a Department of Informatics, University of Bergen, Norway. buixuan@ii.uib.no
More informationCS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010
CS 598CSC: Combinatorial Optimization Lecture date: /4/010 Instructor: Chandra Chekuri Scribe: David Morrison Gomory-Hu Trees (The work in this section closely follows [3]) Let G = (V, E) be an undirected
More informationRegular Languages and Finite State Machines
Regular Languages and Finite State Machines Plan for the Day: Mathematical preliminaries - some review One application formal definition of finite automata Examples 1 Sets A set is an unordered collection
More information! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three
More informationOPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION
OPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION Sérgio Pequito, Stephen Kruzick, Soummya Kar, José M. F. Moura, A. Pedro Aguiar Department of Electrical and Computer Engineering
More informationApproximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
More information3 Some Integer Functions
3 Some Integer Functions A Pair of Fundamental Integer Functions The integer function that is the heart of this section is the modulo function. However, before getting to it, let us look at some very simple
More informationGraphical Models, Exponential Families, and Variational Inference
Foundations and Trends R in Machine Learning Vol. 1, Nos. 1 2 (2008) 1 305 c 2008 M. J. Wainwright and M. I. Jordan DOI: 10.1561/2200000001 Graphical Models, Exponential Families, and Variational Inference
More informationInformation Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
More informationCost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation:
CSE341T 08/31/2015 Lecture 3 Cost Model: Work, Span and Parallelism In this lecture, we will look at how one analyze a parallel program written using Cilk Plus. When we analyze the cost of an algorithm
More informationBurgers vector, Burgers circuit, and Dislocation Line Direction
Burgers vector, Burgers circuit, and Dislocation Line Direction Keonwook Kang and Wei Cai November 21, 2007 The 1st version of this white paper was written after the online discussion between Keonwook
More informationCS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott
More informationMidterm Practice Problems
6.042/8.062J Mathematics for Computer Science October 2, 200 Tom Leighton, Marten van Dijk, and Brooke Cowan Midterm Practice Problems Problem. [0 points] In problem set you showed that the nand operator
More informationCMPSCI611: Approximating MAX-CUT Lecture 20
CMPSCI611: Approximating MAX-CUT Lecture 20 For the next two lectures we ll be seeing examples of approximation algorithms for interesting NP-hard problems. Today we consider MAX-CUT, which we proved to
More information. 0 1 10 2 100 11 1000 3 20 1 2 3 4 5 6 7 8 9
Introduction The purpose of this note is to find and study a method for determining and counting all the positive integer divisors of a positive integer Let N be a given positive integer We say d is a
More informationIncorporating Evidence in Bayesian networks with the Select Operator
Incorporating Evidence in Bayesian networks with the Select Operator C.J. Butz and F. Fang Department of Computer Science, University of Regina Regina, Saskatchewan, Canada SAS 0A2 {butz, fang11fa}@cs.uregina.ca
More informationProblem Set 7 Solutions
8 8 Introduction to Algorithms May 7, 2004 Massachusetts Institute of Technology 6.046J/18.410J Professors Erik Demaine and Shafi Goldwasser Handout 25 Problem Set 7 Solutions This problem set is due in
More informationThe Open University s repository of research publications and other research outputs
Open Research Online The Open University s repository of research publications and other research outputs The degree-diameter problem for circulant graphs of degree 8 and 9 Journal Article How to cite:
More informationBasics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar
More informationE3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
More informationTopology-based network security
Topology-based network security Tiit Pikma Supervised by Vitaly Skachek Research Seminar in Cryptography University of Tartu, Spring 2013 1 Introduction In both wired and wireless networks, there is the
More informationTagging with Hidden Markov Models
Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,
More informationLecture 17 : Equivalence and Order Relations DRAFT
CS/Math 240: Introduction to Discrete Mathematics 3/31/2011 Lecture 17 : Equivalence and Order Relations Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT Last lecture we introduced the notion
More informationPUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5.
PUTNAM TRAINING POLYNOMIALS (Last updated: November 17, 2015) Remark. This is a list of exercises on polynomials. Miguel A. Lerma Exercises 1. Find a polynomial with integral coefficients whose zeros include
More informationLarge-Sample Learning of Bayesian Networks is NP-Hard
Journal of Machine Learning Research 5 (2004) 1287 1330 Submitted 3/04; Published 10/04 Large-Sample Learning of Bayesian Networks is NP-Hard David Maxwell Chickering David Heckerman Christopher Meek Microsoft
More informationMining Social Network Graphs
Mining Social Network Graphs Debapriyo Majumdar Data Mining Fall 2014 Indian Statistical Institute Kolkata November 13, 17, 2014 Social Network No introduc+on required Really? We s7ll need to understand
More informationOutline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits
Outline NP-completeness Examples of Easy vs. Hard problems Euler circuit vs. Hamiltonian circuit Shortest Path vs. Longest Path 2-pairs sum vs. general Subset Sum Reducing one problem to another Clique
More informationMining Social-Network Graphs
342 Chapter 10 Mining Social-Network Graphs There is much information to be gained by analyzing the large-scale data that is derived from social networks. The best-known example of a social network is
More informationChapter 7: Termination Detection
Chapter 7: Termination Detection Ajay Kshemkalyani and Mukesh Singhal Distributed Computing: Principles, Algorithms, and Systems Cambridge University Press A. Kshemkalyani and M. Singhal (Distributed Computing)
More informationVERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS
VERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS MICHAEL DRMOTA, OMER GIMENEZ, AND MARC NOY Abstract. We show that the number of vertices of a given degree k in several kinds of series-parallel labelled
More informationWhy? A central concept in Computer Science. Algorithms are ubiquitous.
Analysis of Algorithms: A Brief Introduction Why? A central concept in Computer Science. Algorithms are ubiquitous. Using the Internet (sending email, transferring files, use of search engines, online
More informationLecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs
CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like
More informationBayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University caizhua@gmail.com
Bayesian Machine Learning (ML): Modeling And Inference in Big Data Zhuhua Cai Google Rice University caizhua@gmail.com 1 Syllabus Bayesian ML Concepts (Today) Bayesian ML on MapReduce (Next morning) Bayesian
More informationBayesian Network Modeling of Hangul Characters for On-line Handwriting Recognition
Bayesian Network Modeling of Hangul haracters for On-line Handwriting Recognition Sung-ung ho and in H. Kim S Div., EES Dept., KAIS, 373- Kusong-dong, Yousong-ku, Daejon, 305-70, KOREA {sjcho, jkim}@ai.kaist.ac.kr
More informationI. GROUPS: BASIC DEFINITIONS AND EXAMPLES
I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called
More informationLecture 4: BK inequality 27th August and 6th September, 2007
CSL866: Percolation and Random Graphs IIT Delhi Amitabha Bagchi Scribe: Arindam Pal Lecture 4: BK inequality 27th August and 6th September, 2007 4. Preliminaries The FKG inequality allows us to lower bound
More informationChapter 6: Graph Theory
Chapter 6: Graph Theory Graph theory deals with routing and network problems and if it is possible to find a best route, whether that means the least expensive, least amount of time or the least distance.
More informationBayesian Networks. Read R&N Ch. 14.1-14.2. Next lecture: Read R&N 18.1-18.4
Bayesian Networks Read R&N Ch. 14.1-14.2 Next lecture: Read R&N 18.1-18.4 You will be expected to know Basic concepts and vocabulary of Bayesian networks. Nodes represent random variables. Directed arcs
More informationMatrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.
Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that
More informationReliability. 26.1 Reliability Models. Chapter 26 Page 1
Chapter 26 Page 1 Reliability Although the technological achievements of the last 50 years can hardly be disputed, there is one weakness in all mankind's devices. That is the possibility of failure. What
More informationLearning Sum-Product Networks with Direct and Indirect Variable Interactions
Learning Sum-Product Networks with Direct and Indirect Variable Interactions Amirmohammad Rooshenas Daniel Lowd Department of Computer and Information Science, University of Oregon PEDRAM@CS.UOREGON.EDU
More informationHT2015: SC4 Statistical Data Mining and Machine Learning
HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric
More informationDefinition 11.1. Given a graph G on n vertices, we define the following quantities:
Lecture 11 The Lovász ϑ Function 11.1 Perfect graphs We begin with some background on perfect graphs. graphs. First, we define some quantities on Definition 11.1. Given a graph G on n vertices, we define
More informationTriangle deletion. Ernie Croot. February 3, 2010
Triangle deletion Ernie Croot February 3, 2010 1 Introduction The purpose of this note is to give an intuitive outline of the triangle deletion theorem of Ruzsa and Szemerédi, which says that if G = (V,
More informationCompression algorithm for Bayesian network modeling of binary systems
Compression algorithm for Bayesian network modeling of binary systems I. Tien & A. Der Kiureghian University of California, Berkeley ABSTRACT: A Bayesian network (BN) is a useful tool for analyzing the
More informationIE 680 Special Topics in Production Systems: Networks, Routing and Logistics*
IE 680 Special Topics in Production Systems: Networks, Routing and Logistics* Rakesh Nagi Department of Industrial Engineering University at Buffalo (SUNY) *Lecture notes from Network Flows by Ahuja, Magnanti
More informationCompact Representations and Approximations for Compuation in Games
Compact Representations and Approximations for Compuation in Games Kevin Swersky April 23, 2008 Abstract Compact representations have recently been developed as a way of both encoding the strategic interactions
More informationTU e. Advanced Algorithms: experimentation project. The problem: load balancing with bounded look-ahead. Input: integer m 2: number of machines
The problem: load balancing with bounded look-ahead Input: integer m 2: number of machines integer k 0: the look-ahead numbers t 1,..., t n : the job sizes Problem: assign jobs to machines machine to which
More information6.3 Conditional Probability and Independence
222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted
More informationUnified Lecture # 4 Vectors
Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,
More informationExtracting correlation structure from large random matrices
Extracting correlation structure from large random matrices Alfred Hero University of Michigan - Ann Arbor Feb. 17, 2012 1 / 46 1 Background 2 Graphical models 3 Screening for hubs in graphical model 4
More informationFinding the M Most Probable Configurations Using Loopy Belief Propagation
Finding the M Most Probable Configurations Using Loopy Belief Propagation Chen Yanover and Yair Weiss School of Computer Science and Engineering The Hebrew University of Jerusalem 91904 Jerusalem, Israel
More informationAlgebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.
This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra
More informationAn Application of Graphical Modeling to the Analysis of Intranet Benefits and Applications
Journal of Data Science 3(2005), 1-17 An Application of Graphical Modeling to the Analysis of Intranet Benefits and Applications Raffaella Settimi, Linda V. Knight, Theresa A. Steinbach and James D. White
More informationLecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University
Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University 1 Chapter 1 Probability 1.1 Basic Concepts In the study of statistics, we consider experiments
More informationGraph Theory Problems and Solutions
raph Theory Problems and Solutions Tom Davis tomrdavis@earthlink.net http://www.geometer.org/mathcircles November, 005 Problems. Prove that the sum of the degrees of the vertices of any finite graph is
More informationChapter 2 Introduction to Sequence Analysis for Human Behavior Understanding
Chapter 2 Introduction to Sequence Analysis for Human Behavior Understanding Hugues Salamin and Alessandro Vinciarelli 2.1 Introduction Human sciences recognize sequence analysis as a key aspect of any
More informationA Sublinear Bipartiteness Tester for Bounded Degree Graphs
A Sublinear Bipartiteness Tester for Bounded Degree Graphs Oded Goldreich Dana Ron February 5, 1998 Abstract We present a sublinear-time algorithm for testing whether a bounded degree graph is bipartite
More informationDiscuss the size of the instance for the minimum spanning tree problem.
3.1 Algorithm complexity The algorithms A, B are given. The former has complexity O(n 2 ), the latter O(2 n ), where n is the size of the instance. Let n A 0 be the size of the largest instance that can
More informationChapter 11. 11.1 Load Balancing. Approximation Algorithms. Load Balancing. Load Balancing on 2 Machines. Load Balancing: Greedy Scheduling
Approximation Algorithms Chapter Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you're unlikely to find a poly-time algorithm. Must sacrifice one
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationSolutions to Homework 6
Solutions to Homework 6 Debasish Das EECS Department, Northwestern University ddas@northwestern.edu 1 Problem 5.24 We want to find light spanning trees with certain special properties. Given is one example
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationProperties of Real Numbers
16 Chapter P Prerequisites P.2 Properties of Real Numbers What you should learn: Identify and use the basic properties of real numbers Develop and use additional properties of real numbers Why you should
More informationEvaluation of Machine Learning Techniques for Green Energy Prediction
arxiv:1406.3726v1 [cs.lg] 14 Jun 2014 Evaluation of Machine Learning Techniques for Green Energy Prediction 1 Objective Ankur Sahai University of Mainz, Germany We evaluate Machine Learning techniques
More information