A Dual Eigenvector Condition for Strong Lumpability of Markov Chains
|
|
|
- Dwain Blankenship
- 9 years ago
- Views:
Transcription
1 A Dual Eigenvector Condition for Strong Lumpability of Markov Chains Martin Nilsson Jacobi Olof Göernerup SFI WORKING PAPER: SFI Working Papers contain accounts of scientific work of the author(s) and do not necessarily represent the views of the Santa Fe Institute. We accept papers intended for publication in peer-reviewed journals or proceedings volumes, but not papers that have already appeared in print. Except for papers by our external faculty, papers must be based on work done at SFI, inspired by an invited visit to or collaboration at SFI, or funded by an SFI grant. NOTICE: This working paper is included by permission of the contributing author(s) as a means to ensure timely distribution of the scholarly and technical work on a non-commercial basis. Copyright and all rights therein are maintained by the author(s). It is understood that all persons copying this information will adhere to the terms and constraints invoked by each author's copyright. These works may be reposted only with the explicit permission of the copyright holder. SANTA FE INSTITUTE
2 A DUAL EIGENVECTOR CONDITION FOR STRONG LUMPABILITY OF MARKOV CHAINS MARTIN NILSSON JACOBI AND OLOF GÖRNERUP Abstract. Necessary and sufficient conditions for identifying strong lumpability in Markov chains are presented. We show that the states in a lump necessarily correspond to identical elements in eigenvectors of the dual transition matrix. If there exist as many dual eigenvectors that respect the necessary condition as there are lumps in the aggregation, then the condition is also sufficient. The result is demonstrated with two simple examples. Key words. Strong lumpability, aggregated Markov chains, state space reduction AMS subject classifications. 60J22,15A18, 60J99 1. Introduction. In this paper we present a method for coarse graining Markov chains by partitioning, or lumping, the state space into aggregated Markov chains. This approach to coarse grain Markov chains has been studied for a long time, see e.g. [7, 11]. Due to the combinatorial explosion of the number of possible partitions of the state space, the conditions used to define lumpability are not useful for identifying aggregated states even for relatively moderately sized Markov chains. Existing efficient methods for lumping are concerned with the special task of finding the most coarse approximate lumping [4], or special situations where the transition matrix possesses known symmetries (e.g. permutation symmetries) [2, 9, 10]. There also exist efficient methods for testing that a given partition constitute a valid lumping [1, 6, 5]. It should be noted that there can exist multiple different accepted lumpings of a Markov chain. Furthermore, different lumpings are in general not related through successive refinements. This paper addresses the general question: Given the transition matrix of a Markov process, how can the state space be partitioned such that the Markov property is not lost in the aggregated process? 2. Main result. Consider a Markov chain X t, t = 0, 1,..., with a finite state space Σ = {1,..., N} and transition probability matrix P = [p ij ]. The transition matrix operates from the right x t+1 = x t P. A lumping is defined by a partition of the state space Σ = {L 1,...,L M } where L k is a nonempty subset of Σ, L k L l = Ø if k l and k L k = Σ. By necessity M N. A lumping can be defined by an N M matrix Π = [π ik ] where π ik = 1 if i L k and π ik = 0 if i / L k. The lumping Σ induces a quotient process x t = x t Π. If the process x t is a Markov process (which is not usually the case), then we say that the Markov chain is strongly lumpable with respect to Π (or Σ). The following criterion is necessary and sufficient for a Markov chain with transition matrix P to be lumpable with respect to Π or Σ [7]: THEOREM 1. If, for any L k, L l Σ, the total probability of a transition from any state i L k to L l, i.e. j L l p ij, is independent of i, then P is lumpable with respect to Σ. This work was funded by PACE (Programmable Artificial Cell Evolution), a European Integrated Project in the EU FP6-IST-FET Complex Systems Initiative, and by EMBIO (Emergent Organisation in Complex Biomolecular Systems), a European Project in the EU FP6-NEST Initiative. Complex Systems Group, Department of Energy and Environment, Chalmers University of Technology, Göteborg, Sweden ([email protected] and [email protected]). 1
3 2 M. NILSSON JACOBI AND O. GÖRNERUP The transition matrix for the reduced Markov chain is given by p kl = j L l p ij i L k, (2.1) where we note that P = [ p kl ] is well defined since the sum is independent of which representative i L k we chose. As mentioned in the introduction, Theorem 1 is not immediately useful for identifying lumpings of a Markov chain. In previous work by Barr and Thomas [1], a necessary lumpability criterion involving the left eigenvectors of the transition matrix was presented. It was first noted that the spectrum of P must be a subset of the spectrum of P. Furthermore, it was noted that if vp = λv then ṽ P = λṽ, with ṽ = vπ. It follows that if λ is an eigenvalue of both P and P, then ṽ is an eigenvector of P, but if λ is not an eigenvalue of P then ṽ = vπ = 0. Intuitively this observation can be understood as Π eliminating a subset of the eigenvectors of P. It is an example of a general result concerning reduction of a linear operator P, which can be formulated as a commutation relation Π P = PΠ that must be fulfilled for a reduction Π to be well defined [8]. It is implied that ker(π) is P-invariant (i.e. spanned by eigenvectors of P, if we for the moment ignore the problem of redundant eigenvalues). Barr and Thomas result suggests a search for lumpings defined by Π such that v α Π = 0 for some subset of the left eigenvectors of P, {v α } α J, J Σ. Since Π should be a matrix with zeros and ones, v α Π = 0 essentially means searching for eigenvectors with subsets of elements that sums to zero. For lumpings only involving agglomeration of two states this is straightforward since the eigenvector(s) must have pairs of elements vi α = vj α. However, agglomeration of k states means searching for partial sums evaluating to zero and involving k elements. In addition (as we will see) there must be several eigenvectors simultaneously satisfying v α Π = 0 (the criterion is only necessary, not sufficient). As Barr and Thomas point out, there is no obvious algorithm to generate Π based on their result. In this paper we present a result that can easily be used for identifying possible partitions of the state space leading to a reduced Markov process. Our method is based on the observation that the dual of the probability space can be used to identify lumplings. The criterion v α Π = 0 is viewed as an orthogonality condition between the left eigenvectors {v α } α J and the column space of Π. The orthogonal complement of a set of left eigenvectors is spanned by complementary right eigenvectors (defined naturally in the dual vector space). These complementary eigenvectors span the column space of Π. Requiring that Π consists of zeros and ones does, as we will see, correspond to a criterion of repeated elements within each complementary right eigenvector. Clearly, identifying repeated elements in the right eigenvectors is algorithmically straight forward. The precise result reads as follows. THEOREM 2. Assume that P is a diagonalizable transition matrix with full rank describing a Markov process x t+1 = x t P. Consider a set of linearly independent right eigenvectors of P, Pu α = λ α u α. Let I Σ be the set of indices for the eigenvectors. Form state equivalence classes defined by states with identical elements in all eigenvectors u α, i.e. i j iff u α i = u α j α I. The equivalence classes define a partitioning Σ of the state space. This partitioning is a lumping of the Markov chain if the number of partition elements equals the number of eigenvectors, i.e. Σ = I. Conversely, if Σ is a lumping then there exist Σ linearly independent right eigenvectors that are invariant under permutations within the lumps.
4 STRONG LUMPABILITY OF MARKOV CHAINS 3 Proof. Let {v β } β=1,...,n be the linearly independent (1 N) left eigenvectors of P and {u β } β=1,...,n be the (N 1) right eigenvectors. We normalize the left and/or right eigenvectors so that v β u γ = δ βγ (where δ is the Kronecker delta). Let Π define the partitioning Σ = {L 1,...,L I }. Since {u α } α I are linearly independent and Σ = I, by the construction of the equivalence classes, the column space of Π is also spanned by {u α } α I. From the orthogonality of left and right eigenvectors it follows that We now decompose P as v β Π = 0, β / I. (2.2) p ij = N λ β u β i vβ j, (2.3) β=1 where u β i and vβ i denote the ith entry in right respective left eigenvector β. Let i L k. From Eq. 2.3 follows that p ij = j L l N λ β u β i β=1 N λ β u β i j L l v β j = β=1 ( v β Π ) l = λ β u β i β I ( v β Π ) l, (2.4) where we use Eq. 2.2 in the last step. By the construction (u β i = u β j if i, j L k, β I) it is clear that the right hand side of Eq. 2.4 does not depend on i L k. From Theorem 1 it follows that Σ is a lumping of the Markov chain. We now show that if the Markov chain is lumpable with Σ defining the partition, then there exist Σ eigenvectors with constant elements over the lumps. Let Σ = {L 1,..., L I } be a lumping of the Markov chain. Let y be a vector defined as y i = ỹ k i L k, (2.5) i.e. constant elements within the lumps defined by Σ, and ỹ at this point an arbitrary vector on the reduced state space. Now consider p ij y j = p ij y j = p kl ỹ l i L k. (2.6) j l j L l It follows that y is a right eigenvector of P if ỹ is an eigenvector of P. Since P has exactly Σ eigenvectors (clearly, P is diagonalizable if P is) it follows that P has Σ eigenvectors with the structure imposed in Eq The two trivial lumpings: Σ = {L1,..., L N } (all states in different lumps) represented by all the right eigenvectors (since P is full rank there can be no index pair equal in all eigenvectors); and Σ = {L 1 } (all states in one lump) represented by the always present u = (1, 1,...,1) T eigenvector (implied by j p ij = 1). As a consequence of Σ = {L 1 } always being an accepted lumping, any eigenvector suggesting a unique two-lump partition is an accepted lumping. In practice, the right eigenvectors involved in forming the equivalence classes often, but not always, form a hierarchy of refined partitions (always including Σ = {L 1 } as the most coarse partition), and the overall partition is set by the eigenvector suggesting the finest partition. It should be noted however that nothing prevents ũ k = ũ l for k l in Eq. 2.5, which implies l
5 4 M. NILSSON JACOBI AND O. GÖRNERUP that not all lumpings of a Markov chain must correspond to a single eigenvector with exactly that suggested lumping. A lumping can instead be composed of eigenvectors each by themselves suggesting a coarser lumping. Consider for instance (1, 1,...,1) T, which is always a member of a subset of eigenvectors that correspond to a valid lumping. The four-lump partition ( Σ 5 ) in Example 2 below also exemplifies this situation. In general, the total number of possible lumpings of a Markov chain may exceed the number of eigenvectors, i.e. the number of states. Theorem 2 is valid also if P has redundant eigenvalues. However, eigenvectors corresponding to redundant eigenvalues are not unique. By rotating the eigenvectors corresponding to a redundant eigenvalue it can be possible to find different suggested partitions. Especially, any k elements can be set equal by choosing an appropriate linear combination of the k eigenvectors corresponding to a k-fold degenerate eigenvalue. It can happen that forcing a subset of the elements to be equal simultaneously sets some other elements equal as well, creating a more coarse lumping than expected. The degenerate eigenvalues in Example 2 demonstrates this point. The identity matrix is an extreme example of redundant eigenvalues. In this case every partition of the state space is a lumping. The number of possible lumpings equals the Bell number B N and grows extremely fast with N. From this argument it is clear that, in some cases, generating all possible lumpings of a Markov chain is not computationally feasible. The lumpings suggested by the eigenvectors can be viewed as generators that can possibly be combined to create finer partitions of the state space. A more non-trivial example than the identity matrix is given in Example Examples. Theorem 2 is illustrated with two examples. Example 1. Consider the transition matrix P = a + b + (c 1)/2 1 a b (1 c)/2 a + (c + 1)/2 a (1 c)/2 1 b c b c, with 0 a, b, c 1. P has the eigenvalues λ 1 = 1, λ 2 = 2a+b 1 and λ 3 = (3c 1)/2, and the eigenvectors u 1 = (1, 1, 1) T, u 2 = (1 + c 2a 2b, 2(a c), 2b + c 1) T and u 3 = ( 1, 1, 2) T. There are three possible lumpings of P: Σ 1 = {{1, 2, 3}} from {u 1 }, Σ 2 = {{1, 2}, {3}} from {u 1, u 3 } and Σ 3 = {{1}, {2}, {3}} from {u 1, u 2, u 3 }, where Σ 2 is valid if 2a + b 1 0 and 3c 1 0. Example 2. Consider the transition matrix (from [3])
6 STRONG LUMPABILITY OF MARKOV CHAINS 5 Table 3.1 Possible lumpings of P and associated subsets of eigenvectors. Lumping Subset of eigenvectors Σ 1 = {{1, 2, 3, 4, 5, 6, 7, 8}} {u 1 } Σ 2 = {{1, 2, 3, 4}, {5, 6, 7, 8}} {u 1, u 8 } Σ 3 = {{1, 2}, {3, 4}, {5, 6,7, 8}} {u 1, u 7, u 8 } Σ 4 = {{1, 2, 3, 4}, {5, 6}, {7, 8}} {u 1, u 2, u 8 } Σ 5 = {{1, 2}, {3, 4}, {5, 6}, {7, 8}} {u 1, u 2, u 7, u 8 } Σ 6 = {{1, 2}, {3}, {4}, {5, 8}, {6, 7}} {u 1, u 3, u 5, u 7, u 8 } Σ 7 = {{1}, {2}, {3, 4}, {5, 7}, {6, 8}} {u 1, u 4, u 6, u 7, u 8 } Σ 8 = {{1, 3}, {2, 4}, {5, 6}, {7}, {8}} {u 1, u 2, u 3 + u 4, u 5 u 6, u 8 } Σ 9 = {{1, 4}, {2, 3}, {5}, {6}, {7, 8}} {u 1, u 2, u 3 u 4, u 5 + u 6, u 8 } Σ 10 = {{1}, {2},..., {7}, {8}} {u 1, u 2,..., u 7, u 8 } P = 0 a a a a 0 a 0 a 0 a a 0 a 0 a a a 0 a 0 a a 0 a a a 0 a 0 0 a b 0 0 b 0 0 b b 0 b b b b b 0 b 0 b b b 0 b b b 0 0 where a = 1/5 and b = 1/4. P has two pairs of degenerate eigenvalues. The associated (not normalized) right eigenvectors u i of P have the following structure (columns sorted in order of decreasing λ ) T = (u 1 u 2...u 8 ) =, c 1 c 1 c 2 c 2 0 5/4 1 1 c 1 c 1 c 2 c 2 0 5/4 1 1 c 1 c 1 c 2 c 2 0 5/4 1 1 c 1 c 1 c 2 c 2 0 5/4 for some numerical constants c 1 and c 2, (c and c ). Note that u 3 and u 4 share eigenvalue and therefore any linear combination αu 3 + βu 4 is also an eigenvector. The same is true for u 5 and u 6. The possible lumpings of P and the corresponding subsets of right eigenvectors are given in Table 3.1. Note that Σ 5 is a lumping with an agglomeration not suggested by any of the individual eigenvectors involved, but rather by the set of eigenvectors as a whole. Further, partition Σ 6 through Σ 9 involve appropriate linear combinations of eigenvalues that correspond to the two degenerate eigenvalues of P, as indicated in Table Discussion and conclusions. We have presented a sufficient and necessary condition for Markov chain lumpability. The condition can be applied to identify,
7 6 M. NILSSON JACOBI AND O. GÖRNERUP a set of lumpings of a given Markov chain by identifying sets of identical elements in the dual eigenvectors. The results presented could have been expressed naturally using symmetries. The equivalence relation between identical elements in the dual eigenvectors implies invariance under a permutation symmetry. The lumpings are orbits of the permutation group, which acts as a normal subgroup extension to the transition matrix. In some situations this formulation can be advantageous, especially to clarify correspondence with general reduction methods in dynamical systems. In this paper we have chosen not to emphasize the symmetry formulation since it tends to obstruct the clarity of the presentation. REFERENCES [1] D. R. Barr and M. U. Thomas, An eigenvector condition for Markov chain lumpability, Operations Research, 25 (1977), pp [2] P. Buchholz, Hierarchical Markovian models: symmetries and reduction, Performance Evaluation, 22 (1995), pp [3] G. W. Cobb and Y. Chen, An application of Markov chain Monte Carlo to community ecology, The American Mathematical Monthly, 110 (2003), pp [4] S. Derisavi, H. Hermanns, and W. H. Sanders, Optimal state-space lumping in Markov chains, Information Processing Letters, 87 (2003), pp [5] L. Gurvits and J. Ledoux, Markov property for a function of a Markov chain: A linear algebra approach, Linear Algebra and its Applications, 404 (2005), pp [6] R.W. Jernigan and R.H. Baran, Testing lumpability in Markov chains, Statistics and Probability Letters, 64 (2003), pp (7). [7] J. G. Kemeny and J. L. Snell, Finite Markov Chains, Springer, New York, NY, USA, 2nd ed., [8] E.R. Lorch, Spectral theory, Oxford University Press, New York, [9] W.D. Obal, M.G. McQuinn, and W.H. Sanders, Detecting and exploiting symmetry in discrete-state Markov models, in 12th Pacific Rim International Symposium on Dependable Computing (PRDC 06), Los Alamitos, CA, USA, 2006, IEEE Computer Society, pp [10] A. Ring, Symmetric lumping of Markov chains, in Jahrestagung der DMV, Halle, 2002, Deutsche Mathematiker-Vereinigung, p [11] L.C.G. Rogers and J.W. Pitman, Markov functions, Annals of probability, 9 (1981), pp
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
Orthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
MA106 Linear Algebra lecture notes
MA106 Linear Algebra lecture notes Lecturers: Martin Bright and Daan Krammer Warwick, January 2011 Contents 1 Number systems and fields 3 1.1 Axioms for number systems......................... 3 2 Vector
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
Linear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka [email protected] http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
University of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
The Matrix Elements of a 3 3 Orthogonal Matrix Revisited
Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
Notes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
T ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
Systems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
Inner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.
ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?
Lecture 1: Schur s Unitary Triangularization Theorem
Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections
Decision-making with the AHP: Why is the principal eigenvector necessary
European Journal of Operational Research 145 (2003) 85 91 Decision Aiding Decision-making with the AHP: Why is the principal eigenvector necessary Thomas L. Saaty * University of Pittsburgh, Pittsburgh,
1 Symmetries of regular polyhedra
1230, notes 5 1 Symmetries of regular polyhedra Symmetry groups Recall: Group axioms: Suppose that (G, ) is a group and a, b, c are elements of G. Then (i) a b G (ii) (a b) c = a (b c) (iii) There is an
Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
What is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM Carl D Meyer INTRODUCTION Foranonnegative irreducible matrix m m with spectral radius ρ,afundamental problem concerns the determination of the unique normalized
1 Teaching notes on GMM 1.
Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in
Applied Linear Algebra I Review page 1
Applied Linear Algebra Review 1 I. Determinants A. Definition of a determinant 1. Using sum a. Permutations i. Sign of a permutation ii. Cycle 2. Uniqueness of the determinant function in terms of properties
[1] Diagonal factorization
8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact
Chapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
Quadratic forms Cochran s theorem, degrees of freedom, and all that
Quadratic forms Cochran s theorem, degrees of freedom, and all that Dr. Frank Wood Frank Wood, [email protected] Linear Regression Models Lecture 1, Slide 1 Why We Care Cochran s theorem tells us
MAT 242 Test 2 SOLUTIONS, FORM T
MAT 242 Test 2 SOLUTIONS, FORM T 5 3 5 3 3 3 3. Let v =, v 5 2 =, v 3 =, and v 5 4 =. 3 3 7 3 a. [ points] The set { v, v 2, v 3, v 4 } is linearly dependent. Find a nontrivial linear combination of these
COMMUTATIVITY DEGREES OF WREATH PRODUCTS OF FINITE ABELIAN GROUPS
COMMUTATIVITY DEGREES OF WREATH PRODUCTS OF FINITE ABELIAN GROUPS IGOR V. EROVENKO AND B. SURY ABSTRACT. We compute commutativity degrees of wreath products A B of finite abelian groups A and B. When B
Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
Part 2: Community Detection
Chapter 8: Graph Data Part 2: Community Detection Based on Leskovec, Rajaraman, Ullman 2014: Mining of Massive Datasets Big Data Management and Analytics Outline Community Detection - Social networks -
Linear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
Introduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
Examination paper for TMA4205 Numerical Linear Algebra
Department of Mathematical Sciences Examination paper for TMA4205 Numerical Linear Algebra Academic contact during examination: Markus Grasmair Phone: 97580435 Examination date: December 16, 2015 Examination
A Primer on Index Notation
A Primer on John Crimaldi August 28, 2006 1. Index versus Index notation (a.k.a. Cartesian notation) is a powerful tool for manipulating multidimensional equations. However, there are times when the more
6.2 Permutations continued
6.2 Permutations continued Theorem A permutation on a finite set A is either a cycle or can be expressed as a product (composition of disjoint cycles. Proof is by (strong induction on the number, r, of
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
I. GROUPS: BASIC DEFINITIONS AND EXAMPLES
I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called
Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products
Chapter 3 Cartesian Products and Relations The material in this chapter is the first real encounter with abstraction. Relations are very general thing they are a special type of subset. After introducing
Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain
Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal
Chapter 7. Permutation Groups
Chapter 7 Permutation Groups () We started the study of groups by considering planar isometries In the previous chapter, we learnt that finite groups of planar isometries can only be cyclic or dihedral
MATH 551 - APPLIED MATRIX THEORY
MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
Elasticity Theory Basics
G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold
Minimally Infeasible Set Partitioning Problems with Balanced Constraints
Minimally Infeasible Set Partitioning Problems with alanced Constraints Michele Conforti, Marco Di Summa, Giacomo Zambelli January, 2005 Revised February, 2006 Abstract We study properties of systems of
The Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
Finite Dimensional Hilbert Spaces and Linear Inverse Problems
Finite Dimensional Hilbert Spaces and Linear Inverse Problems ECE 174 Lecture Supplement Spring 2009 Ken Kreutz-Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California,
Notes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
ALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
Linear Algebra: Determinants, Inverses, Rank
D Linear Algebra: Determinants, Inverses, Rank D 1 Appendix D: LINEAR ALGEBRA: DETERMINANTS, INVERSES, RANK TABLE OF CONTENTS Page D.1. Introduction D 3 D.2. Determinants D 3 D.2.1. Some Properties of
State of Stress at Point
State of Stress at Point Einstein Notation The basic idea of Einstein notation is that a covector and a vector can form a scalar: This is typically written as an explicit sum: According to this convention,
Techniques algébriques en calcul quantique
Techniques algébriques en calcul quantique E. Jeandel Laboratoire de l Informatique du Parallélisme LIP, ENS Lyon, CNRS, INRIA, UCB Lyon 8 Avril 25 E. Jeandel, LIP, ENS Lyon Techniques algébriques en calcul
Split Nonthreshold Laplacian Integral Graphs
Split Nonthreshold Laplacian Integral Graphs Stephen Kirkland University of Regina, Canada [email protected] Maria Aguieiras Alvarez de Freitas Federal University of Rio de Janeiro, Brazil [email protected]
These axioms must hold for all vectors ū, v, and w in V and all scalars c and d.
DEFINITION: A vector space is a nonempty set V of objects, called vectors, on which are defined two operations, called addition and multiplication by scalars (real numbers), subject to the following axioms
SOLUTIONS TO EXERCISES FOR. MATHEMATICS 205A Part 3. Spaces with special properties
SOLUTIONS TO EXERCISES FOR MATHEMATICS 205A Part 3 Fall 2008 III. Spaces with special properties III.1 : Compact spaces I Problems from Munkres, 26, pp. 170 172 3. Show that a finite union of compact subspaces
CS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010
CS 598CSC: Combinatorial Optimization Lecture date: /4/010 Instructor: Chandra Chekuri Scribe: David Morrison Gomory-Hu Trees (The work in this section closely follows [3]) Let G = (V, E) be an undirected
NOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction
IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL R. DRNOVŠEK, T. KOŠIR Dedicated to Prof. Heydar Radjavi on the occasion of his seventieth birthday. Abstract. Let S be an irreducible
Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS
Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and
Finite dimensional C -algebras
Finite dimensional C -algebras S. Sundar September 14, 2012 Throughout H, K stand for finite dimensional Hilbert spaces. 1 Spectral theorem for self-adjoint opertors Let A B(H) and let {ξ 1, ξ 2,, ξ n
Chapter 20. Vector Spaces and Bases
Chapter 20. Vector Spaces and Bases In this course, we have proceeded step-by-step through low-dimensional Linear Algebra. We have looked at lines, planes, hyperplanes, and have seen that there is no limit
Matrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
Numerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 [email protected] 1 Course G63.2010.001 / G22.2420-001, Fall 2010 September 30th, 2010 A. Donev (Courant Institute)
Elements of Abstract Group Theory
Chapter 2 Elements of Abstract Group Theory Mathematics is a game played according to certain simple rules with meaningless marks on paper. David Hilbert The importance of symmetry in physics, and for
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Markov random fields and Gibbs measures
Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).
Recall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
Inner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Linearly Independent Sets and Linearly Dependent Sets
These notes closely follow the presentation of the material given in David C. Lay s textbook Linear Algebra and its Applications (3rd edition). These notes are intended primarily for in-class presentation
SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS
Stochastic Models, 22:289 317, 2006 Copyright Taylor & Francis Group, LLC ISSN: 1532-6349 print/1532-4214 online DOI: 10.1080/15326340600649045 SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL
. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2
4. Basic feasible solutions and vertices of polyhedra Due to the fundamental theorem of Linear Programming, to solve any LP it suffices to consider the vertices (finitely many) of the polyhedron P of the
2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
fg = f g. 3.1.1. Ideals. An ideal of R is a nonempty k-subspace I R closed under multiplication by elements of R:
30 3. RINGS, IDEALS, AND GRÖBNER BASES 3.1. Polynomial rings and ideals The main object of study in this section is a polynomial ring in a finite number of variables R = k[x 1,..., x n ], where k is an
Duality in Linear Programming
Duality in Linear Programming 4 In the preceding chapter on sensitivity analysis, we saw that the shadow-price interpretation of the optimal simplex multipliers is a very useful concept. First, these shadow
Factorization Theorems
Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization
Separation Properties for Locally Convex Cones
Journal of Convex Analysis Volume 9 (2002), No. 1, 301 307 Separation Properties for Locally Convex Cones Walter Roth Department of Mathematics, Universiti Brunei Darussalam, Gadong BE1410, Brunei Darussalam
CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation
Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in
13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms
arxiv:1203.1525v1 [math.co] 7 Mar 2012
Constructing subset partition graphs with strong adjacency and end-point count properties Nicolai Hähnle [email protected] arxiv:1203.1525v1 [math.co] 7 Mar 2012 March 8, 2012 Abstract Kim defined
Allen Back. Oct. 29, 2009
Allen Back Oct. 29, 2009 Notation:(anachronistic) Let the coefficient ring k be Q in the case of toral ( (S 1 ) n) actions and Z p in the case of Z p tori ( (Z p )). Notation:(anachronistic) Let the coefficient
On an algorithm for classification of binary self-dual codes with minimum distance four
Thirteenth International Workshop on Algebraic and Combinatorial Coding Theory June 15-21, 2012, Pomorie, Bulgaria pp. 105 110 On an algorithm for classification of binary self-dual codes with minimum
NETZCOPE - a tool to analyze and display complex R&D collaboration networks
The Task Concepts from Spectral Graph Theory EU R&D Network Analysis Netzcope Screenshots NETZCOPE - a tool to analyze and display complex R&D collaboration networks L. Streit & O. Strogan BiBoS, Univ.
