Introduction to Flocking {Stochastic Matrices}


 Dustin Austen French
 2 years ago
 Views:
Transcription
1 Supelec EECI Graduate School in Control Introduction to Flocking {Stochastic Matrices} A. S. Morse Yale University Gif sur  Yvette May 21, 2012
2
3 CRAIG REYNOLDS BOIDS The Lion King
4 CRAIG REYNOLDS BOIDS neighborhood Flocking Rules separation alignment cohesion
5 Demetri Terzopoulos Flocking Rules separation alignment cohesion
6 Motivated by simulation results reported in
7 V i c s e k e t a l. s i m u l a t e d a o c k o f n a g e n t s f p a r t i c l e s g a l l m o v i n g i n t h e p l a n e a t t h e s a m e s p e e d s, b u t w i t h d i e r e n t h e a d i n g s µ 1 ; µ 2 ; : : : ; µ n. s µ i s = s p e e d µ i = h e a d i n g Each agent s heading is updated at the same time as the rest using a local rule based on the average of its own current heading plus the headings of its neighbors. Vicsek s simulations demonstrated that these nearest neighbor rules can cause all agents to eventually move in the same direction despite 1. the absence of a leader and/or centralized coordination 2. the fact that each agent s set of neighbors changes with time.
8 Vicsek Model
9 r i neighbors of agent i Each agent is a neighbor of itself agent i Each agent has its own sensing radius r i So neighbor relations are not symmetric
10 HEADING UPDATE EQUATIONS s µ i s = s p e e d µ i = h e a d i n g N i (t) = set of indices of agent i s 0 neighbors at time t n i (t) = number of indices in N i (t) Average at time t of headings of neighbors of agent i. Another rule: Vicsek Flocking Problem: Under what conditions do all n headings converge to a common value? Convex combination {Requires collaboration!}
11 Neighbor Graph N of Index Sets N 1, N 2,., N n G = all directed graphs with vertex set V = {1,2,,n} N = graph in G with an arc from j to i whenever j 2 N i, i 2 {1,2,,n} j i j i s a n e i g h b o r o f i A selfarced graph = any graph G with selfarcs at all vertices 3 1 (1,2) 2 N e i g h b o r g r a p h s = s e l f  a r c e d g r a p h s
12 State Space Model Adjacency Matrix A G of a graph G 2 G: An n n matrix of 0 s and 1 s with a ij = 1 whenever there is an arc in G from i to j. _ (1,2) 2 Indegree = 4, outdegree = Indegree of vertex i = number of arcs entering vertex i Outdegree of vertex i = number of arcs leaving vertex i
13 State Space Model Adjacency Matrix A G of a graph G 2 G: An n n matrix of 0 s and 1 s with a ij = 1 whenever there is an arc in G from i to j. Flocking Matrix F N of a neighbor graph N 2 G: bijection where D N = d i a g o n a l f d 1 ; d 2 ; : : : ; d n g a n d n i = d i = i n  d e g r e e o f v e r t e x i = n X j = 1 a j i Update Eqns: State Model:
14 Vicsek flocking problem: Under what conditions do all n headings converge to a common value? A switched linear system No common quadratic Lyapunov function exists B u t t h e n o n  n e g a t i v e f u n c t i o n V ( µ ) = m a x i f µ g i m i n f µ g i i Verify this! is at least nonincreasing along trajectories But it takes much more to conclude that V! 0 µ ( t + 1 ) = F N ( t ) µ ( t )
15 Vicsek flocking problem: Under what conditions do all n headings converge to a common value? µ ( t + 1 ) = F N ( t ) µ ( t ) Problem reduces to determining conditions on the sequence N(0), N(1),... under which where For if this is so, then where and so
16 S n n = stochastic if {Right} Stochastic Matrices 1. i t h a s o n l y n o n  n e g a t i v e e n t r i e s 2. its row sums all equal 1 Stochastic matrices closed under multiplication flocking matrices are not Flocking matrices are stochastic Therefore it is sufficient to determine conditions on an infinite sequence of n n stochastic matrices S 1, S 2,... so that This is a well studied problem in the theory of nonhomogeneous Markov chains If S is a compact set of n n stochastic matrices whose members each have at least one positive column, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. Why is this true?
17 Induced Norms and SemiNorms For M 2 R n n and p > 0, let M p denote the induced matrix p norm on R n n. We will be interested primarily in the cases p = 1, 2, 1 : For any such p, define 1. Nonnegative: M p 0 2. Homogeneous: rm p = r M p 3. Triangle inequality: M 1 + M 2 p M 1 p + M 2 p verify! These three properties mean that p is a seminorm {If M p = 0 were to imply M = 0, then p would be a norm.} M p = 0 ; M = 0
18 Additional Properties of 1. M p 1 if M p 1 Because M p M p M is semi  contractive in the p seminorm if M p < 1 2. Submultiplicative: Suppose M is a subset of R n n such that M1 = 1 for all M 2 M. Then Proof: Let c 0,c 1 and c 2 denote values of c which minimize M 2 M 11c p, M 11c p, and M 21c p respectively. 1 = M 2 1
19 Suppose M is a subset of R n n such that M1 = 1 for all M 2 M. Let p be fixed and let C be a compact set of semi  contractive matrices in M. Let Then for each infinite sequence of matrices M 1, M 2,... in C, the matrix product converges as i! 1 as fast as i converges to zero, to a rank one matrix of the form 1c. Proof: See board We want to use this fact to prove that: If S is a compact set of n n stochastic matrices whose members each have at least one positive column, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. To do this, it is enough to show that: stochastic matrix S is semicontractive in the seminorm 1 if S has a positive column.
20 Any stochastic matrix S can be written as S = 1c + T where c is the largest row vector for which S  1c is nonnegative and T = S 1c so T1 = S1 1c1 = (1  c1)1 all row sums of T = (1  c1) 0 because T 0 Moreover c 0 if and only if S has a positive column. verify! Therefore (1 c1) < 1 if and only if S has a positive column jsj 1 = m in d jjs 1d 0 jj 1 jjs 1cjj 1 = jjt jj 1 = ( 1 c1) stochastic matrix S is semicontractive in the seminorm 1 if S has a positive column.
21 Transitioning from Matrices to Graphs For a nonnegative matrix M n n, (M) is that graph whose adjacency matrix is the transpose of the matrix which results when each nonzero entries in M is replaced by a 1. In other words, for a nonnegative matrix M, (M) is that graph which has an arc (i, j) from i to j whenever m j,i 0.
22 Transitioning from Matrices to Graphs For a nonnegative matrix M n n, (M) is that graph whose adjacency matrix is the transpose of the matrix which results when each nonzero entries in M is replaced by a 1. ( F N ) = ( A 0 N ) = N A graph is strongly rooted if at least one vertex is adjacent to every vertex in the graph strongly rooted graph Motivation for strongly rooted: For any nonnegative matrix M, (M) has an arc (i, j) whenever m j,i 0. (M) is strongly rooted, M has a positive column
23 Transitioning from Matrices to Graphs If S is a compact set of n n stochastic matrices whose members each have at least one positive column, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. If S is a compact set of n n stochastic matrices whose members each have a strongly rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast.
24 Transitioning from Matrices to Graphs When does... T T q T 2 T 1 If then Thus establishing convergence to 1c of an infinite product of stochastic matrices boils down to determining when the graph of a product of stochastic matrices is strongly rooted.
25 Transitioning from Matrices to Graphs As before G = set of all directed graphs with vertex set {1,2,...,n}. By the composition of graph G 2 2 G with graph G 1 2 G, written G 2 ± G 1, is that directed graph in G which has an arc (i, j) from i to j whenever there is an integer k such that (i, k) is an arc in G 1 and (k, j) is an arc in G 2. What motivates this definition? If A and B are nonnegative n n matrices and C = BA, then Thus c ji 0 if and only if for some k, b jk 0 and a ki 0. Therefore (i, j) is an arc in (C) if and only if for some k, (i, k) is an arc in (A) and (k, j) is an arc in (B). (BA) = (B) ± (A)
26 Transitioning from Matrices to Graphs Graph composition is defined so that for any two n n stochastic matrices S 1 and S 2 (S 2 S 1 ) = (S 2 ) ± (S 1 ) Thus deciding when a finite product of stochastic matrices has a strongly rooted graph is the same problem as deciding when a finite composition of graphs is strongly rooted. So... When is the composition of a finite number of graphs strongly rooted?
27 A rooted graph is any graph in G which has has at least one vertex v which, for each vertex i 2 V there is a directed path from v to i. 3 roots When is the composition of a finite number of graphs strongly rooted? rooted graph Every composition of (n 1) 2 or more selfarced, rooted graphs in G is strongly rooted. Proof: See notes. The set of selfarced, rooted graphs in G is the largest set of set of selfarced graphs in G for which every sufficiently long composition is strongly rooted. Proof: See notes.
28 Every composition of (n 1) 2 or more selfarced, rooted graphs in G is strongly rooted.
29 Every composition of (n 1) 2 or more selfarced, rooted graphs in G is strongly rooted. Define A (G ) = set of arcs in G For given graphs G 1, G 2 2 G, G 2 ± G 1, is that graph in G for which (i, j) 2 A (G 2 ± G 1 ) whenever there is an integer k such that (i, k) 2 A (G 1 ) and (k, j) 2 A (G 2 ). If G 1 has a selfarc at i, then (i, i) 2 A (G 1 ). If G 1 has a selfarc at i and (i, j) 2 A (G 2 ) for some j, then (i, j) 2 A (G 2 ± G 1 ) If G 2 has a selfarc at j, then (j, j) 2 A (G 2 ). If G 2 has a selfarc at j and (i, j) 2 A (G 1 ) for some i, then (i, j) 2 A (G 2 ± G 1 ) If G 1 and G 2 both have selfarcs at all vertices, then A (G 2 )[A(G 1 ) ½ A (G 2 ±G 1 ). In general for A (G 2 )[A(G 1 ) A (G 2 ±G 1 ) even if both graphs are selfarced. However for selfarced graphs, if there is a directed path between i and j in G 2 ±G 1 then there is a directed bath between i and j in G 2 [ G 1
30 If S is a compact set of n n stochastic matrices whose members each have a strongly rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast.... rooted strongly rooted strongly rooted strongly rooted If S is a compact set of n n stochastic matrices whose members each have a selfarced, rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast. We can generalize further still...
31 Repeatedly Jointly Rooted Sequences An finite sequence of graphs G 1, G 2,..., G p in G is jointly rooted if the composed graph G p ± G p1 ± ± G 1 is rooted. An infinite sequence of graphs G 1, G 2,... in G is repeatedly jointly rooted if there is a finite positive integer m for which each of the sequences G m(k 1)+1,... G k 1, k 1, is jointly rooted.... repeatedly jointly rooted rooted rooted rooted If S is a compact set of n n stochastic matrices whose members each have a selfarced, rooted graph, then for each sequence of matrices S 1, S 2, from S, and this limit is approached exponentially fast.
32 Suppose S is a compact set of n n stochastic matrices whose members each have a selfarced graph. Suppose that S 1, S 2,... is an infinite sequence of matrices from S whose corresponding sequence of graphs (S 1 ), (S 2 ),... is repeatedly jointly rooted by subsequences of length m. Suppose in addition that the set of all products of m matrices from S with rooted graphs, written C(m), is closed. Then and this limit is approached exponentially fast. Compactness of S does not in general imply compactness of C(m). Construct an example for 2 2 matrices with m = 2. Exception: If S is finite and thus compact {as in flocking applications} so is C(m) Exception: If S is the set of stochastic matrices modeling the convex combo flocking rule, then S and C(m) are both compact. verify this! Flocking Theorem: For each trajectory of the Vicsek flocking system µ(t +1) = F N(t) µ(t) along which the sequence of neighbor graphs N(0), N(1),... is repeatedly jointly rooted, there is a constant steady state heading µ ss which µ(t) approaches exponentially fast, as t! 1.
33 Collectively Rooted Sequences The flocking theorem relies on the notion of jointly rooted sequences: An finite sequence of graphs G 1, G 2,..., G p in G is jointly rooted if the composed graph G p ± G p1 ± ± G 1 is rooted. By the union of G 1, G 2 2 G is meant that graph G 1 [ G 2 in G with arc set A (G 1 ) [ A (G 2 ). An finite sequence of graphs G 1, G 2,..., G p in G is collectively rooted if the union graph G p [ G p1 [ [ G 1 is rooted. In general, for selfarced graphs A ( G p [ G p1 [ [ G 1 ) is a strictly proper subset of A (G p ± G p1 ± ± G 1 ) However, for each arc (i, j) 2 A (G p ± G p1 ± ± G 1 ) there must be a directed path between (i, j) in G p [ G p1 [ [ G 1 Therefore for selfarced graphs, the sequence G 1, G 2,..., G p in G is jointly rooted if and only if it is collectively rooted.
34 Leader Following Suppose that one of the agents in the group, namely agent k, ignores Vicsek s update rule and decides instead to move with some arbitrary but fixed heading θ 0. Suppose that the remaining agents are unaware of this nonconformist s decision and continue to follow Vicsek s rule just as before. Note that under these conditions, agent k must have no neighbors to follow which means that vertex k of any neighbor graph N for the group cannot have any incident arcs. Because of this, the only possible way such a graph N could be rooted or strongly rooted would be if vertex k were the root of N and the only root of N. All of the preceding results are applicable to this case without change. Thus for example, all agents in the group will eventually move in the same direction as agent k if the sequence of neighbor graphs is repeatedly jointly rooted. However more can be said in this special case
35 Leader Following For example, suppose that the neighbor graphs N(1), N(2),... are all rooted. Then each N(t) must be rooted at k. It was noted before that the composition of any (n 1) 2 selfarced rooted graphs in G must be strongly rooted. However in the special case of selfarced, rooted graphs in G which all have a root at the same vertex v, it takes the composition of only (n 1) of them to produce a strongly rooted graph. See notes for a proof Because of this, one would expect faster convergence than in the leaderless case, all other things being equal.
36 FOLLOWING RED LEADER
37 FOLLOWING RED LEADER Leader s Neighbors Yellow
38 FOLLOWING RED LEADER Rectangle Pattern Leader s Neighbors Yellow
39 Symmetric Neighbor Relations The original version of the flocking problem considered the case when all neighbor relations were symmetric that is if agent i is a neighbor of agent j then agent j is a neighbor of agent i. Mathematically, a symmetric neighbor relation means that i 2 N j, j 2 N i The corresponding neighbor graph N would thus be symmetric as well. A directed graph G 2 G is symmetric if (i, j) 2 A (G), (j, i) 2 A (G ) A rooted symmetric graph is the same thing as a strongly connected symmetric graph A graph G 2 G is strongly connected if there is a directed path between any two distinct vertices i and j.
40 Another Way to Write the Vicsek Flocking System L(t) is a symmetric matrix if N(t) is symmetric.
41 Simplified Rule for Symmetric Neighbor Relations Simplified flocking matrix: 1. Symmetric 2. Nonnegative if g > max d i 3. F s 1 = 1 L1 = D1 A1 = 0 F s is stochastic if g > max d i
42 Comparing Flocking Matrices Let N be a given selfarcd directed, symmetric, neighbor graph. A = A 0 (F s ) = (F) = N Therefore all convergence results hold without change for the simplified flocking rule assuming symmetric neighbor relations. Can extend the symmetric case to continuous time.
43 Convergence Rates First we will consider this matter in relation to the seminorm 1
44 Let C be a compact set of n n stochastic matrices which are semi  contractive in the infinity norm. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. So what we d like is a uniform upper bound on S 1 over C. = a convergence rate bound Any stochastic matrix S can be written as S = 1c + T where c is the largest row vector for which S  1c is nonnegative and T = S 1c jsj 1 = m in d jjs 1d 0 jj 1 jjs 1cjj 1 = jjt jj 1 = ( 1 c1) Note that the ith entry c i in c must be the smallest entry in the ith column of S. Since (1 c1) 1  c i for any i, jsj 1 ( 1 c i ) 8i
45 Convergence Rate Bound for Flocking Matrices with Strongly Rooted Graphs If c i is the smallest element in the ith column of S, then S 1 1 c i Suppose that F is a flocking matrix whose graph is strongly rooted at vertex k Then (F) must have an arc from vertex k to each other vertex in the graph which means that the kth row of adjacency matrix A of (F) must be [ ] Since F = D 1 A 0 where D = diagonal {n 1,n 2,..., n n }, the kth column of F must be The smallest entry in this column is bounded below by Therefore for any flocking matrix F with a strongly rooted graph = a convergence rate bound
46 An Explicit Formula for the Infinity SemiNorm For any nonnegative n n matrix M, See notes of a proof of this fact. If M is a stochastic matrix, the quantity on the right is known as the coefficient of ergodicity. For any real numbers x and y So for a stochastic matrix S
47 * 1. S 1 1 Because S 1 S 1 = 1 2. S 1 = 0 if and only if all rows are equal = iff S = 1c For fixed i and j, the kth term in the sum in will be positive iff s ik > 0 and s jk > 0 * Therefore the sum in * will be positive iff s ik > 0 and s jk > 0 for at least one value of k Therefore S 1 < 1 iff for each distinct i and j, s ik > 0 and s jk > 0 for at least one value of k. A stochastic matrix with this property is called a scrambling matrix
48 Summary A stochastic matrix S is a scrambling matrix for each distinct i and j, s ik > 0 and s jk > 0 for at least one value of k. Equivalently, a stochastic matrix is a scrambling matrix if no two rows are orthogonal. A stochastic matrix is a semicontraction in the infinity norm iff it is a scrambling matrix. An explicit formula for the infinity seminorm of any stochastic matrix S is
49 The Graph of a Scrambling Matrix. A stochastic matrix S is a scrambling matrix for each distinct i and j, s ik > 0 and s jk > 0 for at least one value of k. A graph G 2 G is neighbor shared if each two distinct vertices i and j have a common neighbor k A stochastic matrix S is a scrambling matrix if and only if its graph is neighbor shared. In a strongly rooted graph there must be a root which is the neighbor of each vertex in the graph. So... Every strongly rooted graph is neighbor shared. The converse is clearly false.
50 NeighborShared Directed Graph A neighbor shared graph is a directed graph in which each pair of distinct vertices share a common neighbor and 2 share 2 1 and 3 share 2 1 and 4 share 4 2 and 3 share 2 2 and 4 share 4 3 and 4 share 1
51 Suppose G 2 G is neighbor shared. Then any pair of vertices (i, j) must be reachable from a {common neighbor} vertex k. Suppose for some integer p 2 {2, 3,..., n 1}, each subset of p vertices is reachable from a single vertex. Let {v 1, v 2,..., v p } be any any such set and let v be a vertex from which all of the v i can be reached. Let w be any vertex not in the set {v 1, v 2,..., v p }. Since G is neighbor shared, w and v can be reached from a common vertex y Therefore every vertex in the set {v 1, v 2,..., v p, w} can be reached from y. So every subset of p + 1 vertices in the graph is reachable from a single vertex. So by induction all n vertices are reachable from a single vertex. Every neighborshared graph in G is rooted. Converse is false. Verify by constructing an example. Strongly rooted graphs ½ neighbor shared graphs ½ rooted graphs
52 Compositions of rooted and neighbor shared graphs The composition on any n 1 or more selfarced, rooted graphs in G is neighbor shared. Proof: See notes. We will use this fact a little later to get a convergence rate bound for products of flocking matrices whose sequence of graphs is repeatedly jointly rooted. The composition on any n 1 or more neighborshared graphs in G is strongly rooted. Let s outline a proof of this: Strongly rooted graphs ½ neighbor shared graphs ½ rooted graphs
53 The composition on any n 1 or more neighborshared graphs in G is strongly rooted. A graph G 2 G is k neighbor shared if each set of k distinct vertices in G share a common neighbor. A 2 neighbor shared graph is thus a neighbor shared graph and an n neighbor shared graph is obviously strongly rooted. Suppose G is neighbor shared and H is k neighbor shared for some k < n Let {v 1, v 2,..., v k+1 } be distinct vertices. Since H is k neighbor shared, in H {v 1, v 2,..., v k } share a common neighbor p and {v 2, v 3,..., v k +1 } share a common neighbor q Since G is neighbor shared, in G p and q share a common neighbor w. In H ±G, vertices v 1, v 2,..., v k must have w for a neighbor as must vertices v 2, v 3,..., v k +1 Therefore in H ±G, vertices v 1, v 2,..., v k+1 must have w for a neighbor. Therefore H ±G must be k + 1 neighbor shared. Complete the proof using induction.
54 Convergence rate bounds for products of scrambling matrices Let C be a compact set of n n stochastic matrices which are semi  contractive in the infinity norm. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. Scrambling matrices are semicontractive in the infinity norm. Let C be a compact set of n n scrambling matrices. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. What can be said about convergence rate for scrambling matrices which are also flocking matrices?
55 Worst Case F 1 for F = D 1 A 0 = Scrambling (F) = N = neighbor shared A = [a ij ] a ij = 1 a ij = 0 if i is a neighbor of j otherwise D = diagonal {d 1, d 2,, d n } n n d i = indegree of vertex i Since all d i n, all nonzero f ij satisfy
56 Worst Case F 1 for F = D 1 A 0 = Scrambling (F) = N = neighbor shared A = [a ij ] a ij = 1 a ij = 0 if i is a neighbor of j otherwise D = diagonal {d 1, d 2,, d n } n n d i = indegree of vertex i Since all d i n, all nonzero f ij satisfy Fix distinct i and j and let k be a shared neighbor. Then f ik 0 f jk.
57 Worst Case F 1 for F = D 1 A 0 = Scrambling (F) = N = neighbor shared n 3 Vertex 1 has only itself as a neighbor Vertex 2 has every vertex as a neighbor For i > 2, vertex i has only itself and vertex 1 as neighbors How tight is this bound?
58 Summary Flocking Matrices with NeighborShared Graphs Every infinite product of n n flocking matrices with neighborshared graphs converges to a rankone matrix product 1c at a rate no slower than There exist infinite product of n n flocking matrices with neighborshared graphs which actually converge to a rankone matrix product 1c at this rate.
59 Convergence rates for products of stochastic matrices with rooted graphs The composition on any n 1 or more selfarced, rooted graphs in G is neighbor shared. Let C be a compact set of n n scrambling matrices. Then for each infinite sequence of matrices S 1 S 2,... in C, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero. Let S be a compact set of n n rooted matrices and write C for the compact set of all products of n 1 matrices from S. Then for each infinite sequence of matrices S 1 S 2,... in S, the matrix product converges as i! 1 to a rank one matrix as fast as converges to zero What can be said about the convergence rate for the product of an infinite sequence of flocking matrices whose sequence of graphs is repeatedly jointly rooted?
60 We need a few ideas For any nonzero matrix M 0, define Á(M) = smallest nonzero element of M. Note that M can be written as where For S 1 and S 2 n n stochastic matrices By induction
61 Recall that Suppose S is scrambling Claim that Since S is scrambling, for any distinct i and j there must be a k such that
62 If S is scrambling S Á(S) The composition on any n 1 or more selfarced, rooted graphs in G is neighbor shared. F = set of all n n flocking matrices F(p) = { F p F p1 F 1 : F i 2 F, { (F 1 ), (F 2 ),..., (F p ) } is jointly rooted } Each matrix in F (p) is rooted F k (p) = set of all products of k matrices from F(p) Each matrix in F k (p) is scrambling if k n  1 For any F 2 F, If S 2 F k (p), then S is the product of kp flocking matrices so If k = n 1, then S is scrambling and Therefore a convergence rate bound for the infinite product of flocking matrices whose sequence of graphs is repeatedly jointly rooted is
63
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 19967 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationDiagonal, Symmetric and Triangular Matrices
Contents 1 Diagonal, Symmetric Triangular Matrices 2 Diagonal Matrices 2.1 Products, Powers Inverses of Diagonal Matrices 2.1.1 Theorem (Powers of Matrices) 2.2 Multiplying Matrices on the Left Right by
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationMatrix Norms. Tom Lyche. September 28, Centre of Mathematics for Applications, Department of Informatics, University of Oslo
Matrix Norms Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 28, 2009 Matrix Norms We consider matrix norms on (C m,n, C). All results holds for
More informationDETERMINANTS. b 2. x 2
DETERMINANTS 1 Systems of two equations in two unknowns A system of two equations in two unknowns has the form a 11 x 1 + a 12 x 2 = b 1 a 21 x 1 + a 22 x 2 = b 2 This can be written more concisely in
More informationLecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs
CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like
More informationA Distributed Algorithm for Solving a Linear Algebraic Equation
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, ACCEPTED. 1 A Distributed Algorithm for Solving a Linear Algebraic Equation Shaoshuai Mou Ji Liu A. Stephen Morse arxiv:1503.00808v1 [cs.sy] 3 Mar 2015 Abstract
More informationSHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH
31 Kragujevac J. Math. 25 (2003) 31 49. SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH Kinkar Ch. Das Department of Mathematics, Indian Institute of Technology, Kharagpur 721302, W.B.,
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationNotes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 918/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
More informationSection 6.1  Inner Products and Norms
Section 6.1  Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More informationTHE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More information(67902) Topics in Theory and Complexity Nov 2, 2006. Lecture 7
(67902) Topics in Theory and Complexity Nov 2, 2006 Lecturer: Irit Dinur Lecture 7 Scribe: Rani Lekach 1 Lecture overview This Lecture consists of two parts In the first part we will refresh the definition
More informationThe Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
More informationSPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BIDIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIXEXPONENTIAL DISTRIBUTIONS
Stochastic Models, 22:289 317, 2006 Copyright Taylor & Francis Group, LLC ISSN: 15326349 print/15324214 online DOI: 10.1080/15326340600649045 SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BIDIAGONAL
More informationBANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationCONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation
Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in
More informationx if x 0, x if x < 0.
Chapter 3 Sequences In this chapter, we discuss sequences. We say what it means for a sequence to converge, and define the limit of a convergent sequence. We begin with some preliminary results about the
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An ndimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0534405967. Systems of Linear Equations Definition. An ndimensional vector is a row or a column
More informationDETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH
DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH CHRISTOPHER RH HANUSA AND THOMAS ZASLAVSKY Abstract We investigate the least common multiple of all subdeterminants,
More information9.2 Summation Notation
9. Summation Notation 66 9. Summation Notation In the previous section, we introduced sequences and now we shall present notation and theorems concerning the sum of terms of a sequence. We begin with a
More information4. MATRICES Matrices
4. MATRICES 170 4. Matrices 4.1. Definitions. Definition 4.1.1. A matrix is a rectangular array of numbers. A matrix with m rows and n columns is said to have dimension m n and may be represented as follows:
More informationEigenvalues, Eigenvectors, Matrix Factoring, and Principal Components
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they
More informationProbability and Statistics
CHAPTER 2: RANDOM VARIABLES AND ASSOCIATED FUNCTIONS 2b  0 Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute  Systems and Modeling GIGA  Bioinformatics ULg kristel.vansteen@ulg.ac.be
More informationSCORE SETS IN ORIENTED GRAPHS
Applicable Analysis and Discrete Mathematics, 2 (2008), 107 113. Available electronically at http://pefmath.etf.bg.ac.yu SCORE SETS IN ORIENTED GRAPHS S. Pirzada, T. A. Naikoo The score of a vertex v in
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus ndimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More informationA linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form
Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationSYSTEMS OF EQUATIONS
SYSTEMS OF EQUATIONS 1. Examples of systems of equations Here are some examples of systems of equations. Each system has a number of equations and a number (not necessarily the same) of variables for which
More information4. An isosceles triangle has two sides of length 10 and one of length 12. What is its area?
1 1 2 + 1 3 + 1 5 = 2 The sum of three numbers is 17 The first is 2 times the second The third is 5 more than the second What is the value of the largest of the three numbers? 3 A chemist has 100 cc of
More information6.3 Conditional Probability and Independence
222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted
More informationWeek 5: Binary Relations
1 Binary Relations Week 5: Binary Relations The concept of relation is common in daily life and seems intuitively clear. For instance, let X be the set of all living human females and Y the set of all
More informationIntroduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra  1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
More informationWhat is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
More informationNotes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 201112 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationLinear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
More informationLecture Notes 1: Matrix Algebra Part B: Determinants and Inverses
University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 57 Lecture Notes 1: Matrix Algebra Part B: Determinants and Inverses Peter J. Hammond email: p.j.hammond@warwick.ac.uk Autumn 2012,
More informationAppendix F: Mathematical Induction
Appendix F: Mathematical Induction Introduction In this appendix, you will study a form of mathematical proof called mathematical induction. To see the logical need for mathematical induction, take another
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More informationSECTION 102 Mathematical Induction
73 0 Sequences and Series 6. Approximate e 0. using the first five terms of the series. Compare this approximation with your calculator evaluation of e 0.. 6. Approximate e 0.5 using the first five terms
More informationHandout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010. Chapter 7: Digraphs
MCS236: Graph Theory Handout #Ch7 San Skulrattanakulchai Gustavus Adolphus College Dec 6, 2010 Chapter 7: Digraphs Strong Digraphs Definitions. A digraph is an ordered pair (V, E), where V is the set
More informationVector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a nonempty
More informationIntroduction. Appendix D Mathematical Induction D1
Appendix D Mathematical Induction D D Mathematical Induction Use mathematical induction to prove a formula. Find a sum of powers of integers. Find a formula for a finite sum. Use finite differences to
More information3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More informationLecture 3: Linear Programming Relaxations and Rounding
Lecture 3: Linear Programming Relaxations and Rounding 1 Approximation Algorithms and Linear Relaxations For the time being, suppose we have a minimization problem. Many times, the problem at hand can
More informationSpring 2007 Math 510 Hints for practice problems
Spring 2007 Math 510 Hints for practice problems Section 1 Imagine a prison consisting of 4 cells arranged like the squares of an chessboard There are doors between all adjacent cells A prisoner in one
More informationSocial Media Mining. Graph Essentials
Graph Essentials Graph Basics Measures Graph and Essentials Metrics 2 2 Nodes and Edges A network is a graph nodes, actors, or vertices (plural of vertex) Connections, edges or ties Edge Node Measures
More informationThe Ideal Class Group
Chapter 5 The Ideal Class Group We will use Minkowski theory, which belongs to the general area of geometry of numbers, to gain insight into the ideal class group of a number field. We have already mentioned
More informationDirect Methods for Solving Linear Systems. Matrix Factorization
Direct Methods for Solving Linear Systems Matrix Factorization Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011
More information1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let
Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationLecture 3: Finding integer solutions to systems of linear equations
Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture
More informationMATH 551  APPLIED MATRIX THEORY
MATH 55  APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
More informationSolving Linear Diophantine Matrix Equations Using the Smith Normal Form (More or Less)
Solving Linear Diophantine Matrix Equations Using the Smith Normal Form (More or Less) Raymond N. Greenwell 1 and Stanley Kertzner 2 1 Department of Mathematics, Hofstra University, Hempstead, NY 11549
More informationMATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform
MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish
More informationRandom graphs with a given degree sequence
Sourav Chatterjee (NYU) Persi Diaconis (Stanford) Allan Sly (Microsoft) Let G be an undirected simple graph on n vertices. Let d 1,..., d n be the degrees of the vertices of G arranged in descending order.
More information3. Equivalence Relations. Discussion
3. EQUIVALENCE RELATIONS 33 3. Equivalence Relations 3.1. Definition of an Equivalence Relations. Definition 3.1.1. A relation R on a set A is an equivalence relation if and only if R is reflexive, symmetric,
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationDivisor graphs have arbitrary order and size
Divisor graphs have arbitrary order and size arxiv:math/0606483v1 [math.co] 20 Jun 2006 Le Anh Vinh School of Mathematics University of New South Wales Sydney 2052 Australia Abstract A divisor graph G
More informationFUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied
More informationSome Polynomial Theorems. John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom.
Some Polynomial Theorems by John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom.com This paper contains a collection of 31 theorems, lemmas,
More informationEC9A0: Presessional Advanced Mathematics Course
University of Warwick, EC9A0: Presessional Advanced Mathematics Course Peter J. Hammond & Pablo F. Beker 1 of 55 EC9A0: Presessional Advanced Mathematics Course Slides 1: Matrix Algebra Peter J. Hammond
More informationCoefficient of Potential and Capacitance
Coefficient of Potential and Capacitance Lecture 12: Electromagnetic Theory Professor D. K. Ghosh, Physics Department, I.I.T., Bombay We know that inside a conductor there is no electric field and that
More informationMath 181 Handout 16. Rich Schwartz. March 9, 2010
Math 8 Handout 6 Rich Schwartz March 9, 200 The purpose of this handout is to describe continued fractions and their connection to hyperbolic geometry. The Gauss Map Given any x (0, ) we define γ(x) =
More informationThe Dirichlet Unit Theorem
Chapter 6 The Dirichlet Unit Theorem As usual, we will be working in the ring B of algebraic integers of a number field L. Two factorizations of an element of B are regarded as essentially the same if
More information1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
More informationNotes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand
Notes V General Equilibrium: Positive Theory In this lecture we go on considering a general equilibrium model of a private ownership economy. In contrast to the Notes IV, we focus on positive issues such
More informationZeros of Polynomial Functions
Zeros of Polynomial Functions The Rational Zero Theorem If f (x) = a n x n + a n1 x n1 + + a 1 x + a 0 has integer coefficients and p/q (where p/q is reduced) is a rational zero, then p is a factor of
More informationBig Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time
Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time Edward Bortnikov & Ronny Lempel Yahoo! Labs, Haifa Class Outline Linkbased page importance measures Why
More information7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
More informationWe know a formula for and some properties of the determinant. Now we see how the determinant can be used.
Cramer s rule, inverse matrix, and volume We know a formula for and some properties of the determinant. Now we see how the determinant can be used. Formula for A We know: a b d b =. c d ad bc c a Can we
More informationn k=1 k=0 1/k! = e. Example 6.4. The series 1/k 2 converges in R. Indeed, if s n = n then k=1 1/k, then s 2n s n = 1 n + 1 +...
6 Series We call a normed space (X, ) a Banach space provided that every Cauchy sequence (x n ) in X converges. For example, R with the norm = is an example of Banach space. Now let (x n ) be a sequence
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationFactoring Algorithms
Factoring Algorithms The p 1 Method and Quadratic Sieve November 17, 2008 () Factoring Algorithms November 17, 2008 1 / 12 Fermat s factoring method Fermat made the observation that if n has two factors
More informationMath 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
More informationThe determinant of a skewsymmetric matrix is a square. This can be seen in small cases by direct calculation: 0 a. 12 a. a 13 a 24 a 14 a 23 a 14
4 Symplectic groups In this and the next two sections, we begin the study of the groups preserving reflexive sesquilinear forms or quadratic forms. We begin with the symplectic groups, associated with
More informationSolution to Homework 2
Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if
More informationApproximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NPCompleteness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
More informationINTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL
SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics
More informationBasic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011
Basic Concepts of Point Set Topology Notes for OU course Math 4853 Spring 2011 A. Miller 1. Introduction. The definitions of metric space and topological space were developed in the early 1900 s, largely
More informationMATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all ndimensional column
More information1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
More informationMathematical Induction. Lecture 1011
Mathematical Induction Lecture 1011 Menu Mathematical Induction Strong Induction Recursive Definitions Structural Induction Climbing an Infinite Ladder Suppose we have an infinite ladder: 1. We can reach
More information1. Prove that the empty set is a subset of every set.
1. Prove that the empty set is a subset of every set. Basic Topology Written by MenGen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since
More informationDiscrete Mathematics & Mathematical Reasoning Chapter 10: Graphs
Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Kousha Etessami U. of Edinburgh, UK Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 6) 1 / 13 Overview Graphs and Graph
More informationLecture 13 Linear quadratic Lyapunov theory
EE363 Winter 289 Lecture 13 Linear quadratic Lyapunov theory the Lyapunov equation Lyapunov stability conditions the Lyapunov operator and integral evaluating quadratic integrals analysis of ARE discretetime
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 34 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationBasic Terminology for Systems of Equations in a Nutshell. E. L. Lady. 3x 1 7x 2 +4x 3 =0 5x 1 +8x 2 12x 3 =0.
Basic Terminology for Systems of Equations in a Nutshell E L Lady A system of linear equations is something like the following: x 7x +4x =0 5x +8x x = Note that the number of equations is not required
More informationZeros of a Polynomial Function
Zeros of a Polynomial Function An important consequence of the Factor Theorem is that finding the zeros of a polynomial is really the same thing as factoring it into linear factors. In this section we
More informationa 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More informationDeterminants in the Kronecker product of matrices: The incidence matrix of a complete graph
FPSAC 2009 DMTCS proc (subm), by the authors, 1 10 Determinants in the Kronecker product of matrices: The incidence matrix of a complete graph Christopher R H Hanusa 1 and Thomas Zaslavsky 2 1 Department
More information1. LINEAR EQUATIONS. A linear equation in n unknowns x 1, x 2,, x n is an equation of the form
1. LINEAR EQUATIONS A linear equation in n unknowns x 1, x 2,, x n is an equation of the form a 1 x 1 + a 2 x 2 + + a n x n = b, where a 1, a 2,..., a n, b are given real numbers. For example, with x and
More informationTheta Functions. Lukas Lewark. Seminar on Modular Forms, 31. Januar 2007
Theta Functions Lukas Lewark Seminar on Modular Forms, 31. Januar 007 Abstract Theta functions are introduced, associated to lattices or quadratic forms. Their transformation property is proven and the
More informationInner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
More information