Introduction to Flocking {Stochastic Matrices}

Size: px
Start display at page:

Download "Introduction to Flocking {Stochastic Matrices}"

Transcription

1 Supelec EECI Graduate School in Control Introduction to Flocking {Stochastic Matrices} A. S. Morse Yale University Gif sur - Yvette May 21, 2012

2

3 CRAIG REYNOLDS BOIDS The Lion King

4 CRAIG REYNOLDS BOIDS neighborhood Flocking Rules separation alignment cohesion

5 Demetri Terzopoulos Flocking Rules separation alignment cohesion

6 Motivated by simulation results reported in

7 V i c s e k e t a l. s i m u l a t e d a o c k o f n a g e n t s f p a r t i c l e s g a l l m o v i n g i n t h e p l a n e a t t h e s a m e s p e e d s, b u t w i t h d i e r e n t h e a d i n g s µ 1 ; µ 2 ; : : : ; µ n. s µ i s = s p e e d µ i = h e a d i n g Each agent s heading is updated at the same time as the rest using a local rule based on the average of its own current heading plus the headings of its neighbors. Vicsek s simulations demonstrated that these nearest neighbor rules can cause all agents to eventually move in the same direction despite 1. the absence of a leader and/or centralized coordination 2. the fact that each agent s set of neighbors changes with time.

8 Vicsek Model

9 r i neighbors of agent i Each agent is a neighbor of itself agent i Each agent has its own sensing radius r i So neighbor relations are not symmetric

10 HEADING UPDATE EQUATIONS s µ i s = s p e e d µ i = h e a d i n g N i (t) = set of indices of agent i s 0 neighbors at time t n i (t) = number of indices in N i (t) Average at time t of headings of neighbors of agent i. Another rule: Vicsek Flocking Problem: Under what conditions do all n headings converge to a common value? Convex combination {Requires collaboration!}

11 Neighbor Graph N of Index Sets N 1, N 2,., N n G = all directed graphs with vertex set V = {1,2,,n} N = graph in G with an arc from j to i whenever j 2 N i, i 2 {1,2,,n} j i j i s a n e i g h b o r o f i A self-arced graph = any graph G with self-arcs at all vertices 3 1 (1,2) 2 N e i g h b o r g r a p h s = s e l f - a r c e d g r a p h s

12 State Space Model Adjacency Matrix A G of a graph G 2 G: An n n matrix of 0 s and 1 s with a ij = 1 whenever there is an arc in G from i to j. _ (1,2) 2 In-degree = 4, out-degree = In-degree of vertex i = number of arcs entering vertex i Out-degree of vertex i = number of arcs leaving vertex i

13 State Space Model Adjacency Matrix A G of a graph G 2 G: An n n matrix of 0 s and 1 s with a ij = 1 whenever there is an arc in G from i to j. Flocking Matrix F N of a neighbor graph N 2 G: bijection where D N = d i a g o n a l f d 1 ; d 2 ; : : : ; d n g a n d n i = d i = i n - d e g r e e o f v e r t e x i = n X j = 1 a j i Update Eqns: State Model:

14 Vicsek flocking problem: Under what conditions do all n headings converge to a common value? A switched linear system No common quadratic Lyapunov function exists B u t t h e n o n - n e g a t i v e f u n c t i o n V ( µ ) = m a x i f µ g i m i n f µ g i i Verify this! is at least non-increasing along trajectories But it takes much more to conclude that V! 0 µ ( t + 1 ) = F N ( t ) µ ( t )

15 Vicsek flocking problem: Under what conditions do all n headings converge to a common value? µ ( t + 1 ) = F N ( t ) µ ( t ) Problem reduces to determining conditions on the sequence N(0), N(1),... under which where For if this is so, then where and so

16 S n n = stochastic if {Right} Stochastic Matrices 1. i t h a s o n l y n o n - n e g a t i v e e n t r i e s 2. its row sums all equal 1 Stochastic matrices closed under multiplication flocking matrices are not Flocking matrices are stochastic Therefore it is sufficient to determine conditions on an infinite sequence of n n stochastic matrices S 1, S 2,... so that This is a well studied problem in the theory of non-homogeneous Markov chains If S is a compact set of n n stochastic matrices whose members each have at least one positive column, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. Why is this true?

17 Induced Norms and Semi-Norms For M 2 R n n and p > 0, let M p denote the induced matrix p norm on R n n. We will be interested primarily in the cases p = 1, 2, 1 : For any such p, define 1. Nonnegative: M p 0 2. Homogeneous: rm p = r M p 3. Triangle inequality: M 1 + M 2 p M 1 p + M 2 p verify! These three properties mean that p is a semi-norm {If M p = 0 were to imply M = 0, then p would be a norm.} M p = 0 ; M = 0

18 Additional Properties of 1. M p 1 if M p 1 Because M p M p M is semi - contractive in the p semi-norm if M p < 1 2. Sub-multiplicative: Suppose M is a subset of R n n such that M1 = 1 for all M 2 M. Then Proof: Let c 0,c 1 and c 2 denote values of c which minimize M 2 M 1-1c p, M 1-1c p, and M 2-1c p respectively. 1 = M 2 1

19 Suppose M is a subset of R n n such that M1 = 1 for all M 2 M. Let p be fixed and let C be a compact set of semi - contractive matrices in M. Let Then for each infinite sequence of matrices M 1, M 2,... in C, the matrix product converges as i! 1 as fast as i converges to zero, to a rank one matrix of the form 1c. Proof: See board We want to use this fact to prove that: If S is a compact set of n n stochastic matrices whose members each have at least one positive column, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. To do this, it is enough to show that: stochastic matrix S is semi-contractive in the semi-norm 1 if S has a positive column.

20 Any stochastic matrix S can be written as S = 1c + T where c is the largest row vector for which S - 1c is nonnegative and T = S 1c so T1 = S1 1c1 = (1 - c1)1 all row sums of T = (1 - c1) 0 because T 0 Moreover c 0 if and only if S has a positive column. verify! Therefore (1 c1) < 1 if and only if S has a positive column jsj 1 = m in d jjs 1d 0 jj 1 jjs 1cjj 1 = jjt jj 1 = ( 1 c1) stochastic matrix S is semi-contractive in the semi-norm 1 if S has a positive column.

21 Transitioning from Matrices to Graphs For a nonnegative matrix M n n, (M) is that graph whose adjacency matrix is the transpose of the matrix which results when each non-zero entries in M is replaced by a 1. In other words, for a nonnegative matrix M, (M) is that graph which has an arc (i, j) from i to j whenever m j,i 0.

22 Transitioning from Matrices to Graphs For a nonnegative matrix M n n, (M) is that graph whose adjacency matrix is the transpose of the matrix which results when each non-zero entries in M is replaced by a 1. ( F N ) = ( A 0 N ) = N A graph is strongly rooted if at least one vertex is adjacent to every vertex in the graph strongly rooted graph Motivation for strongly rooted: For any nonnegative matrix M, (M) has an arc (i, j) whenever m j,i 0. (M) is strongly rooted, M has a positive column

23 Transitioning from Matrices to Graphs If S is a compact set of n n stochastic matrices whose members each have at least one positive column, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. If S is a compact set of n n stochastic matrices whose members each have a strongly rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast.

24 Transitioning from Matrices to Graphs When does... T T q T 2 T 1 If then Thus establishing convergence to 1c of an infinite product of stochastic matrices boils down to determining when the graph of a product of stochastic matrices is strongly rooted.

25 Transitioning from Matrices to Graphs As before G = set of all directed graphs with vertex set {1,2,...,n}. By the composition of graph G 2 2 G with graph G 1 2 G, written G 2 ± G 1, is that directed graph in G which has an arc (i, j) from i to j whenever there is an integer k such that (i, k) is an arc in G 1 and (k, j) is an arc in G 2. What motivates this definition? If A and B are nonnegative n n matrices and C = BA, then Thus c ji 0 if and only if for some k, b jk 0 and a ki 0. Therefore (i, j) is an arc in (C) if and only if for some k, (i, k) is an arc in (A) and (k, j) is an arc in (B). (BA) = (B) ± (A)

26 Transitioning from Matrices to Graphs Graph composition is defined so that for any two n n stochastic matrices S 1 and S 2 (S 2 S 1 ) = (S 2 ) ± (S 1 ) Thus deciding when a finite product of stochastic matrices has a strongly rooted graph is the same problem as deciding when a finite composition of graphs is strongly rooted. So... When is the composition of a finite number of graphs strongly rooted?

27 A rooted graph is any graph in G which has has at least one vertex v which, for each vertex i 2 V there is a directed path from v to i. 3 roots When is the composition of a finite number of graphs strongly rooted? rooted graph Every composition of (n 1) 2 or more self-arced, rooted graphs in G is strongly rooted. Proof: See notes. The set of self-arced, rooted graphs in G is the largest set of set of self-arced graphs in G for which every sufficiently long composition is strongly rooted. Proof: See notes.

28 Every composition of (n 1) 2 or more self-arced, rooted graphs in G is strongly rooted.

29 Every composition of (n 1) 2 or more self-arced, rooted graphs in G is strongly rooted. Define A (G ) = set of arcs in G For given graphs G 1, G 2 2 G, G 2 ± G 1, is that graph in G for which (i, j) 2 A (G 2 ± G 1 ) whenever there is an integer k such that (i, k) 2 A (G 1 ) and (k, j) 2 A (G 2 ). If G 1 has a self-arc at i, then (i, i) 2 A (G 1 ). If G 1 has a self-arc at i and (i, j) 2 A (G 2 ) for some j, then (i, j) 2 A (G 2 ± G 1 ) If G 2 has a self-arc at j, then (j, j) 2 A (G 2 ). If G 2 has a self-arc at j and (i, j) 2 A (G 1 ) for some i, then (i, j) 2 A (G 2 ± G 1 ) If G 1 and G 2 both have self-arcs at all vertices, then A (G 2 )[A(G 1 ) ½ A (G 2 ±G 1 ). In general for A (G 2 )[A(G 1 ) A (G 2 ±G 1 ) even if both graphs are self-arced. However for self-arced graphs, if there is a directed path between i and j in G 2 ±G 1 then there is a directed bath between i and j in G 2 [ G 1

30 If S is a compact set of n n stochastic matrices whose members each have a strongly rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast.... rooted strongly rooted strongly rooted strongly rooted If S is a compact set of n n stochastic matrices whose members each have a self-arced, rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. We can generalize further still...

31 Repeatedly Jointly Rooted Sequences An finite sequence of graphs G 1, G 2,..., G p in G is jointly rooted if the composed graph G p ± G p-1 ± ± G 1 is rooted. An infinite sequence of graphs G 1, G 2,... in G is repeatedly jointly rooted if there is a finite positive integer m for which each of the sequences G m(k -1)+1,... G k -1, k 1, is jointly rooted.... repeatedly jointly rooted rooted rooted rooted If S is a compact set of n n stochastic matrices whose members each have a self-arced, rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast.

32 Suppose S is a compact set of n n stochastic matrices whose members each have a self-arced graph. Suppose that S 1, S 2,... is an infinite sequence of matrices from S whose corresponding sequence of graphs (S 1 ), (S 2 ),... is repeatedly jointly rooted by sub-sequences of length m. Suppose in addition that the set of all products of m matrices from S with rooted graphs, written C(m), is closed. Then and this limit is approached exponentially fast. Compactness of S does not in general imply compactness of C(m). Construct an example for 2 2 matrices with m = 2. Exception: If S is finite and thus compact {as in flocking applications} so is C(m) Exception: If S is the set of stochastic matrices modeling the convex combo flocking rule, then S and C(m) are both compact. verify this! Flocking Theorem: For each trajectory of the Vicsek flocking system µ(t +1) = F N(t) µ(t) along which the sequence of neighbor graphs N(0), N(1),... is repeatedly jointly rooted, there is a constant steady state heading µ ss which µ(t) approaches exponentially fast, as t! 1.

33 Collectively Rooted Sequences The flocking theorem relies on the notion of jointly rooted sequences: An finite sequence of graphs G 1, G 2,..., G p in G is jointly rooted if the composed graph G p ± G p-1 ± ± G 1 is rooted. By the union of G 1, G 2 2 G is meant that graph G 1 [ G 2 in G with arc set A (G 1 ) [ A (G 2 ). An finite sequence of graphs G 1, G 2,..., G p in G is collectively rooted if the union graph G p [ G p-1 [ [ G 1 is rooted. In general, for self-arced graphs A ( G p [ G p-1 [ [ G 1 ) is a strictly proper subset of A (G p ± G p-1 ± ± G 1 ) However, for each arc (i, j) 2 A (G p ± G p-1 ± ± G 1 ) there must be a directed path between (i, j) in G p [ G p-1 [ [ G 1 Therefore for self-arced graphs, the sequence G 1, G 2,..., G p in G is jointly rooted if and only if it is collectively rooted.

34 Leader Following Suppose that one of the agents in the group, namely agent k, ignores Vicsek s update rule and decides instead to move with some arbitrary but fixed heading θ 0. Suppose that the remaining agents are unaware of this non-conformist s decision and continue to follow Vicsek s rule just as before. Note that under these conditions, agent k must have no neighbors to follow which means that vertex k of any neighbor graph N for the group cannot have any incident arcs. Because of this, the only possible way such a graph N could be rooted or strongly rooted would be if vertex k were the root of N and the only root of N. All of the preceding results are applicable to this case without change. Thus for example, all agents in the group will eventually move in the same direction as agent k if the sequence of neighbor graphs is repeatedly jointly rooted. However more can be said in this special case

35 Leader Following For example, suppose that the neighbor graphs N(1), N(2),... are all rooted. Then each N(t) must be rooted at k. It was noted before that the composition of any (n -1) 2 self-arced rooted graphs in G must be strongly rooted. However in the special case of self-arced, rooted graphs in G which all have a root at the same vertex v, it takes the composition of only (n -1) of them to produce a strongly rooted graph. See notes for a proof Because of this, one would expect faster convergence than in the leaderless case, all other things being equal.

36 FOLLOWING RED LEADER

37 FOLLOWING RED LEADER Leader s Neighbors Yellow

38 FOLLOWING RED LEADER Rectangle Pattern Leader s Neighbors Yellow

39 Symmetric Neighbor Relations The original version of the flocking problem considered the case when all neighbor relations were symmetric that is if agent i is a neighbor of agent j then agent j is a neighbor of agent i. Mathematically, a symmetric neighbor relation means that i 2 N j, j 2 N i The corresponding neighbor graph N would thus be symmetric as well. A directed graph G 2 G is symmetric if (i, j) 2 A (G), (j, i) 2 A (G ) A rooted symmetric graph is the same thing as a strongly connected symmetric graph A graph G 2 G is strongly connected if there is a directed path between any two distinct vertices i and j.

40 Another Way to Write the Vicsek Flocking System L(t) is a symmetric matrix if N(t) is symmetric.

41 Simplified Rule for Symmetric Neighbor Relations Simplified flocking matrix: 1. Symmetric 2. Nonnegative if g > max d i 3. F s 1 = 1 L1 = D1 A1 = 0 F s is stochastic if g > max d i

42 Comparing Flocking Matrices Let N be a given self-arcd directed, symmetric, neighbor graph. A = A 0 (F s ) = (F) = N Therefore all convergence results hold without change for the simplified flocking rule assuming symmetric neighbor relations. Can extend the symmetric case to continuous time.

43 Convergence Rates First we will consider this matter in relation to the semi-norm 1

44 Let C be a compact set of n n stochastic matrices which are semi - contractive in the infinity norm. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. So what we d like is a uniform upper bound on S 1 over C. = a convergence rate bound Any stochastic matrix S can be written as S = 1c + T where c is the largest row vector for which S - 1c is nonnegative and T = S 1c jsj 1 = m in d jjs 1d 0 jj 1 jjs 1cjj 1 = jjt jj 1 = ( 1 c1) Note that the ith entry c i in c must be the smallest entry in the ith column of S. Since (1 c1) 1 - c i for any i, jsj 1 ( 1 c i ) 8i

45 Convergence Rate Bound for Flocking Matrices with Strongly Rooted Graphs If c i is the smallest element in the ith column of S, then S 1 1 c i Suppose that F is a flocking matrix whose graph is strongly rooted at vertex k Then (F) must have an arc from vertex k to each other vertex in the graph which means that the kth row of adjacency matrix A of (F) must be [ ] Since F = D -1 A 0 where D = diagonal {n 1,n 2,..., n n }, the kth column of F must be The smallest entry in this column is bounded below by Therefore for any flocking matrix F with a strongly rooted graph = a convergence rate bound

46 An Explicit Formula for the Infinity Semi-Norm For any nonnegative n n matrix M, See notes of a proof of this fact. If M is a stochastic matrix, the quantity on the right is known as the coefficient of ergodicity. For any real numbers x and y So for a stochastic matrix S

47 * 1. S 1 1 Because S 1 S 1 = 1 2. S 1 = 0 if and only if all rows are equal = iff S = 1c For fixed i and j, the kth term in the sum in will be positive iff s ik > 0 and s jk > 0 * Therefore the sum in * will be positive iff s ik > 0 and s jk > 0 for at least one value of k Therefore S 1 < 1 iff for each distinct i and j, s ik > 0 and s jk > 0 for at least one value of k. A stochastic matrix with this property is called a scrambling matrix

48 Summary A stochastic matrix S is a scrambling matrix for each distinct i and j, s ik > 0 and s jk > 0 for at least one value of k. Equivalently, a stochastic matrix is a scrambling matrix if no two rows are orthogonal. A stochastic matrix is a semi-contraction in the infinity norm iff it is a scrambling matrix. An explicit formula for the infinity semi-norm of any stochastic matrix S is

49 The Graph of a Scrambling Matrix. A stochastic matrix S is a scrambling matrix for each distinct i and j, s ik > 0 and s jk > 0 for at least one value of k. A graph G 2 G is neighbor shared if each two distinct vertices i and j have a common neighbor k A stochastic matrix S is a scrambling matrix if and only if its graph is neighbor shared. In a strongly rooted graph there must be a root which is the neighbor of each vertex in the graph. So... Every strongly rooted graph is neighbor shared. The converse is clearly false.

50 Neighbor-Shared Directed Graph A neighbor shared graph is a directed graph in which each pair of distinct vertices share a common neighbor and 2 share 2 1 and 3 share 2 1 and 4 share 4 2 and 3 share 2 2 and 4 share 4 3 and 4 share 1

51 Suppose G 2 G is neighbor shared. Then any pair of vertices (i, j) must be reachable from a {common neighbor} vertex k. Suppose for some integer p 2 {2, 3,..., n -1}, each subset of p vertices is reachable from a single vertex. Let {v 1, v 2,..., v p } be any any such set and let v be a vertex from which all of the v i can be reached. Let w be any vertex not in the set {v 1, v 2,..., v p }. Since G is neighbor shared, w and v can be reached from a common vertex y Therefore every vertex in the set {v 1, v 2,..., v p, w} can be reached from y. So every subset of p + 1 vertices in the graph is reachable from a single vertex. So by induction all n vertices are reachable from a single vertex. Every neighbor-shared graph in G is rooted. Converse is false. Verify by constructing an example. Strongly rooted graphs ½ neighbor shared graphs ½ rooted graphs

52 Compositions of rooted and neighbor shared graphs The composition on any n 1 or more self-arced, rooted graphs in G is neighbor shared. Proof: See notes. We will use this fact a little later to get a convergence rate bound for products of flocking matrices whose sequence of graphs is repeatedly jointly rooted. The composition on any n 1 or more neighbor-shared graphs in G is strongly rooted. Let s outline a proof of this: Strongly rooted graphs ½ neighbor shared graphs ½ rooted graphs

53 The composition on any n 1 or more neighbor-shared graphs in G is strongly rooted. A graph G 2 G is k neighbor shared if each set of k distinct vertices in G share a common neighbor. A 2 neighbor shared graph is thus a neighbor shared graph and an n neighbor shared graph is obviously strongly rooted. Suppose G is neighbor shared and H is k neighbor shared for some k < n Let {v 1, v 2,..., v k+1 } be distinct vertices. Since H is k neighbor shared, in H {v 1, v 2,..., v k } share a common neighbor p and {v 2, v 3,..., v k +1 } share a common neighbor q Since G is neighbor shared, in G p and q share a common neighbor w. In H ±G, vertices v 1, v 2,..., v k must have w for a neighbor as must vertices v 2, v 3,..., v k +1 Therefore in H ±G, vertices v 1, v 2,..., v k+1 must have w for a neighbor. Therefore H ±G must be k + 1 neighbor shared. Complete the proof using induction.

54 Convergence rate bounds for products of scrambling matrices Let C be a compact set of n n stochastic matrices which are semi - contractive in the infinity norm. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. Scrambling matrices are semi-contractive in the infinity norm. Let C be a compact set of n n scrambling matrices. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. What can be said about convergence rate for scrambling matrices which are also flocking matrices?

55 Worst Case F 1 for F = D -1 A 0 = Scrambling (F) = N = neighbor shared A = [a ij ] a ij = 1 a ij = 0 if i is a neighbor of j otherwise D = diagonal {d 1, d 2,, d n } n n d i = in-degree of vertex i Since all d i n, all non-zero f ij satisfy

56 Worst Case F 1 for F = D -1 A 0 = Scrambling (F) = N = neighbor shared A = [a ij ] a ij = 1 a ij = 0 if i is a neighbor of j otherwise D = diagonal {d 1, d 2,, d n } n n d i = in-degree of vertex i Since all d i n, all non-zero f ij satisfy Fix distinct i and j and let k be a shared neighbor. Then f ik 0 f jk.

57 Worst Case F 1 for F = D -1 A 0 = Scrambling (F) = N = neighbor shared n 3 Vertex 1 has only itself as a neighbor Vertex 2 has every vertex as a neighbor For i > 2, vertex i has only itself and vertex 1 as neighbors How tight is this bound?

58 Summary Flocking Matrices with Neighbor-Shared Graphs Every infinite product of n n flocking matrices with neighbor-shared graphs converges to a rank-one matrix product 1c at a rate no slower than There exist infinite product of n n flocking matrices with neighbor-shared graphs which actually converge to a rank-one matrix product 1c at this rate.

59 Convergence rates for products of stochastic matrices with rooted graphs The composition on any n 1 or more self-arced, rooted graphs in G is neighbor shared. Let C be a compact set of n n scrambling matrices. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. Let S be a compact set of n n rooted matrices and write C for the compact set of all products of n 1 matrices from S. Then for each infinite sequence of matrices S 1 S 2,... in S, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero What can be said about the convergence rate for the product of an infinite sequence of flocking matrices whose sequence of graphs is repeatedly jointly rooted?

60 We need a few ideas For any nonzero matrix M 0, define Á(M) = smallest nonzero element of M. Note that M can be written as where For S 1 and S 2 n n stochastic matrices By induction

61 Recall that Suppose S is scrambling Claim that Since S is scrambling, for any distinct i and j there must be a k such that

62 If S is scrambling S Á(S) The composition on any n 1 or more self-arced, rooted graphs in G is neighbor shared. F = set of all n n flocking matrices F(p) = { F p F p-1 F 1 : F i 2 F, { (F 1 ), (F 2 ),..., (F p ) } is jointly rooted } Each matrix in F (p) is rooted F k (p) = set of all products of k matrices from F(p) Each matrix in F k (p) is scrambling if k n - 1 For any F 2 F, If S 2 F k (p), then S is the product of kp flocking matrices so If k = n 1, then S is scrambling and Therefore a convergence rate bound for the infinite product of flocking matrices whose sequence of graphs is repeatedly jointly rooted is

63

DATA ANALYSIS II. Matrix Algorithms

DATA ANALYSIS II. Matrix Algorithms DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

More information

Systems of Linear Equations

Systems of Linear Equations Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

More information

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like

More information

Notes on Determinant

Notes on Determinant ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH

SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH 31 Kragujevac J. Math. 25 (2003) 31 49. SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH Kinkar Ch. Das Department of Mathematics, Indian Institute of Technology, Kharagpur 721302, W.B.,

More information

Similarity and Diagonalization. Similar Matrices

Similarity and Diagonalization. Similar Matrices MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

More information

THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok

THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan

More information

Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

More information

(67902) Topics in Theory and Complexity Nov 2, 2006. Lecture 7

(67902) Topics in Theory and Complexity Nov 2, 2006. Lecture 7 (67902) Topics in Theory and Complexity Nov 2, 2006 Lecturer: Irit Dinur Lecture 7 Scribe: Rani Lekach 1 Lecture overview This Lecture consists of two parts In the first part we will refresh the definition

More information

The Characteristic Polynomial

The Characteristic Polynomial Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem

More information

Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

More information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1. MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

More information

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS Stochastic Models, 22:289 317, 2006 Copyright Taylor & Francis Group, LLC ISSN: 1532-6349 print/1532-4214 online DOI: 10.1080/15326340600649045 SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number

More information

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in

More information

BANACH AND HILBERT SPACE REVIEW

BANACH AND HILBERT SPACE REVIEW BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but

More information

DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH

DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH CHRISTOPHER RH HANUSA AND THOMAS ZASLAVSKY Abstract We investigate the least common multiple of all subdeterminants,

More information

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...

More information

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

More information

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

More information

9.2 Summation Notation

9.2 Summation Notation 9. Summation Notation 66 9. Summation Notation In the previous section, we introduced sequences and now we shall present notation and theorems concerning the sum of terms of a sequence. We begin with a

More information

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they

More information

6.3 Conditional Probability and Independence

6.3 Conditional Probability and Independence 222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted

More information

SCORE SETS IN ORIENTED GRAPHS

SCORE SETS IN ORIENTED GRAPHS Applicable Analysis and Discrete Mathematics, 2 (2008), 107 113. Available electronically at http://pefmath.etf.bg.ac.yu SCORE SETS IN ORIENTED GRAPHS S. Pirzada, T. A. Naikoo The score of a vertex v in

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

What is Linear Programming?

What is Linear Programming? Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to

More information

Linear Programming. March 14, 2014

Linear Programming. March 14, 2014 Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010. Chapter 7: Digraphs

Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010. Chapter 7: Digraphs MCS-236: Graph Theory Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010 Chapter 7: Digraphs Strong Digraphs Definitions. A digraph is an ordered pair (V, E), where V is the set

More information

Introduction. Appendix D Mathematical Induction D1

Introduction. Appendix D Mathematical Induction D1 Appendix D Mathematical Induction D D Mathematical Induction Use mathematical induction to prove a formula. Find a sum of powers of integers. Find a formula for a finite sum. Use finite differences to

More information

Vector and Matrix Norms

Vector and Matrix Norms Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

More information

Notes on Symmetric Matrices

Notes on Symmetric Matrices CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.

More information

Inner Product Spaces

Inner Product Spaces Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

More information

MATH 551 - APPLIED MATRIX THEORY

MATH 551 - APPLIED MATRIX THEORY MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

More information

Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

More information

1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES 1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

More information

SECTION 10-2 Mathematical Induction

SECTION 10-2 Mathematical Induction 73 0 Sequences and Series 6. Approximate e 0. using the first five terms of the series. Compare this approximation with your calculator evaluation of e 0.. 6. Approximate e 0.5 using the first five terms

More information

3. INNER PRODUCT SPACES

3. INNER PRODUCT SPACES . INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.

More information

MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform

MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish

More information

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied

More information

Direct Methods for Solving Linear Systems. Matrix Factorization

Direct Methods for Solving Linear Systems. Matrix Factorization Direct Methods for Solving Linear Systems Matrix Factorization Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011

More information

Metric Spaces. Chapter 7. 7.1. Metrics

Metric Spaces. Chapter 7. 7.1. Metrics Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some

More information

Social Media Mining. Graph Essentials

Social Media Mining. Graph Essentials Graph Essentials Graph Basics Measures Graph and Essentials Metrics 2 2 Nodes and Edges A network is a graph nodes, actors, or vertices (plural of vertex) Connections, edges or ties Edge Node Measures

More information

The Ideal Class Group

The Ideal Class Group Chapter 5 The Ideal Class Group We will use Minkowski theory, which belongs to the general area of geometry of numbers, to gain insight into the ideal class group of a number field. We have already mentioned

More information

1 Norms and Vector Spaces

1 Norms and Vector Spaces 008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)

More information

7 Gaussian Elimination and LU Factorization

7 Gaussian Elimination and LU Factorization 7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method

More information

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as

More information

Random graphs with a given degree sequence

Random graphs with a given degree sequence Sourav Chatterjee (NYU) Persi Diaconis (Stanford) Allan Sly (Microsoft) Let G be an undirected simple graph on n vertices. Let d 1,..., d n be the degrees of the vertices of G arranged in descending order.

More information

Some Polynomial Theorems. John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom.

Some Polynomial Theorems. John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom. Some Polynomial Theorems by John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom.com This paper contains a collection of 31 theorems, lemmas,

More information

Solution to Homework 2

Solution to Homework 2 Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if

More information

How To Prove The Dirichlet Unit Theorem

How To Prove The Dirichlet Unit Theorem Chapter 6 The Dirichlet Unit Theorem As usual, we will be working in the ring B of algebraic integers of a number field L. Two factorizations of an element of B are regarded as essentially the same if

More information

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand Notes V General Equilibrium: Positive Theory In this lecture we go on considering a general equilibrium model of a private ownership economy. In contrast to the Notes IV, we focus on positive issues such

More information

Zeros of Polynomial Functions

Zeros of Polynomial Functions Zeros of Polynomial Functions The Rational Zero Theorem If f (x) = a n x n + a n-1 x n-1 + + a 1 x + a 0 has integer coefficients and p/q (where p/q is reduced) is a rational zero, then p is a factor of

More information

Math 181 Handout 16. Rich Schwartz. March 9, 2010

Math 181 Handout 16. Rich Schwartz. March 9, 2010 Math 8 Handout 6 Rich Schwartz March 9, 200 The purpose of this handout is to describe continued fractions and their connection to hyperbolic geometry. The Gauss Map Given any x (0, ) we define γ(x) =

More information

Methods for Finding Bases

Methods for Finding Bases Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,

More information

Lemma 5.2. Let S be a set. (1) Let f and g be two permutations of S. Then the composition of f and g is a permutation of S.

Lemma 5.2. Let S be a set. (1) Let f and g be two permutations of S. Then the composition of f and g is a permutation of S. Definition 51 Let S be a set bijection f : S S 5 Permutation groups A permutation of S is simply a Lemma 52 Let S be a set (1) Let f and g be two permutations of S Then the composition of f and g is a

More information

Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time

Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time Edward Bortnikov & Ronny Lempel Yahoo! Labs, Haifa Class Outline Link-based page importance measures Why

More information

n k=1 k=0 1/k! = e. Example 6.4. The series 1/k 2 converges in R. Indeed, if s n = n then k=1 1/k, then s 2n s n = 1 n + 1 +...

n k=1 k=0 1/k! = e. Example 6.4. The series 1/k 2 converges in R. Indeed, if s n = n then k=1 1/k, then s 2n s n = 1 n + 1 +... 6 Series We call a normed space (X, ) a Banach space provided that every Cauchy sequence (x n ) in X converges. For example, R with the norm = is an example of Banach space. Now let (x n ) be a sequence

More information

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

More information

Factoring Algorithms

Factoring Algorithms Factoring Algorithms The p 1 Method and Quadratic Sieve November 17, 2008 () Factoring Algorithms November 17, 2008 1 / 12 Fermat s factoring method Fermat made the observation that if n has two factors

More information

LS.6 Solution Matrices

LS.6 Solution Matrices LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions

More information

Math 312 Homework 1 Solutions

Math 312 Homework 1 Solutions Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please

More information

CHAPTER 9. Integer Programming

CHAPTER 9. Integer Programming CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

More information

Chapter 3. Distribution Problems. 3.1 The idea of a distribution. 3.1.1 The twenty-fold way

Chapter 3. Distribution Problems. 3.1 The idea of a distribution. 3.1.1 The twenty-fold way Chapter 3 Distribution Problems 3.1 The idea of a distribution Many of the problems we solved in Chapter 1 may be thought of as problems of distributing objects (such as pieces of fruit or ping-pong balls)

More information

Lecture 17 : Equivalence and Order Relations DRAFT

Lecture 17 : Equivalence and Order Relations DRAFT CS/Math 240: Introduction to Discrete Mathematics 3/31/2011 Lecture 17 : Equivalence and Order Relations Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT Last lecture we introduced the notion

More information

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column

More information

Discrete Mathematics. Hans Cuypers. October 11, 2007

Discrete Mathematics. Hans Cuypers. October 11, 2007 Hans Cuypers October 11, 2007 1 Contents 1. Relations 4 1.1. Binary relations................................ 4 1.2. Equivalence relations............................. 6 1.3. Relations and Directed Graphs.......................

More information

Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs

Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Kousha Etessami U. of Edinburgh, UK Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 6) 1 / 13 Overview Graphs and Graph

More information

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions. Chapter 1 Vocabulary identity - A statement that equates two equivalent expressions. verbal model- A word equation that represents a real-life problem. algebraic expression - An expression with variables.

More information

1 Introduction to Matrices

1 Introduction to Matrices 1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

More information

1. Prove that the empty set is a subset of every set.

1. Prove that the empty set is a subset of every set. 1. Prove that the empty set is a subset of every set. Basic Topology Written by Men-Gen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since

More information

3 Some Integer Functions

3 Some Integer Functions 3 Some Integer Functions A Pair of Fundamental Integer Functions The integer function that is the heart of this section is the modulo function. However, before getting to it, let us look at some very simple

More information

Lecture 13 Linear quadratic Lyapunov theory

Lecture 13 Linear quadratic Lyapunov theory EE363 Winter 28-9 Lecture 13 Linear quadratic Lyapunov theory the Lyapunov equation Lyapunov stability conditions the Lyapunov operator and integral evaluating quadratic integrals analysis of ARE discrete-time

More information

Classification of Cartan matrices

Classification of Cartan matrices Chapter 7 Classification of Cartan matrices In this chapter we describe a classification of generalised Cartan matrices This classification can be compared as the rough classification of varieties in terms

More information

Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

More information

Mathematical Induction. Lecture 10-11

Mathematical Induction. Lecture 10-11 Mathematical Induction Lecture 10-11 Menu Mathematical Induction Strong Induction Recursive Definitions Structural Induction Climbing an Infinite Ladder Suppose we have an infinite ladder: 1. We can reach

More information

Zeros of a Polynomial Function

Zeros of a Polynomial Function Zeros of a Polynomial Function An important consequence of the Factor Theorem is that finding the zeros of a polynomial is really the same thing as factoring it into linear factors. In this section we

More information

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

More information

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact

More information

Modeling and Performance Evaluation of Computer Systems Security Operation 1

Modeling and Performance Evaluation of Computer Systems Security Operation 1 Modeling and Performance Evaluation of Computer Systems Security Operation 1 D. Guster 2 St.Cloud State University 3 N.K. Krivulin 4 St.Petersburg State University 5 Abstract A model of computer system

More information

Inner products on R n, and more

Inner products on R n, and more Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +

More information

How To Understand And Solve Algebraic Equations

How To Understand And Solve Algebraic Equations College Algebra Course Text Barnett, Raymond A., Michael R. Ziegler, and Karl E. Byleen. College Algebra, 8th edition, McGraw-Hill, 2008, ISBN: 978-0-07-286738-1 Course Description This course provides

More information

Determinants in the Kronecker product of matrices: The incidence matrix of a complete graph

Determinants in the Kronecker product of matrices: The incidence matrix of a complete graph FPSAC 2009 DMTCS proc (subm), by the authors, 1 10 Determinants in the Kronecker product of matrices: The incidence matrix of a complete graph Christopher R H Hanusa 1 and Thomas Zaslavsky 2 1 Department

More information

Solution of Linear Systems

Solution of Linear Systems Chapter 3 Solution of Linear Systems In this chapter we study algorithms for possibly the most commonly occurring problem in scientific computing, the solution of linear systems of equations. We start

More information

Linear Programming I

Linear Programming I Linear Programming I November 30, 2003 1 Introduction In the VCR/guns/nuclear bombs/napkins/star wars/professors/butter/mice problem, the benevolent dictator, Bigus Piguinus, of south Antarctica penguins

More information

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013 Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

More information

Mathematical finance and linear programming (optimization)

Mathematical finance and linear programming (optimization) Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may

More information

Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011

Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011 Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011 A. Miller 1. Introduction. The definitions of metric space and topological space were developed in the early 1900 s, largely

More information

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8 Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e

More information

GENERATING SETS KEITH CONRAD

GENERATING SETS KEITH CONRAD GENERATING SETS KEITH CONRAD 1 Introduction In R n, every vector can be written as a unique linear combination of the standard basis e 1,, e n A notion weaker than a basis is a spanning set: a set of vectors

More information

( ) which must be a vector

( ) which must be a vector MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are

More information

Suk-Geun Hwang and Jin-Woo Park

Suk-Geun Hwang and Jin-Woo Park Bull. Korean Math. Soc. 43 (2006), No. 3, pp. 471 478 A NOTE ON PARTIAL SIGN-SOLVABILITY Suk-Geun Hwang and Jin-Woo Park Abstract. In this paper we prove that if Ax = b is a partial signsolvable linear

More information

Row Echelon Form and Reduced Row Echelon Form

Row Echelon Form and Reduced Row Echelon Form These notes closely follow the presentation of the material given in David C Lay s textbook Linear Algebra and its Applications (3rd edition) These notes are intended primarily for in-class presentation

More information

160 CHAPTER 4. VECTOR SPACES

160 CHAPTER 4. VECTOR SPACES 160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results

More information

MATH2210 Notebook 1 Fall Semester 2016/2017. 1 MATH2210 Notebook 1 3. 1.1 Solving Systems of Linear Equations... 3

MATH2210 Notebook 1 Fall Semester 2016/2017. 1 MATH2210 Notebook 1 3. 1.1 Solving Systems of Linear Equations... 3 MATH0 Notebook Fall Semester 06/07 prepared by Professor Jenny Baglivo c Copyright 009 07 by Jenny A. Baglivo. All Rights Reserved. Contents MATH0 Notebook 3. Solving Systems of Linear Equations........................

More information

Chapter 6. Orthogonality

Chapter 6. Orthogonality 6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

More information