Lecture 6: The Group Inverse

Similar documents
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

Similarity and Diagonalization. Similar Matrices

Section Inner Products and Norms

Orthogonal Diagonalization of Symmetric Matrices

Lecture 1: Schur s Unitary Triangularization Theorem

NOTES ON LINEAR TRANSFORMATIONS

by the matrix A results in a vector which is a reflection of the given

Factorization Theorems

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Lecture 2 Matrix Operations

CONTROLLABILITY. Chapter Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Inner Product Spaces and Orthogonality

Applied Linear Algebra I Review page 1

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

UNCOUPLING THE PERRON EIGENVECTOR PROBLEM

LINEAR ALGEBRA. September 23, 2010

The Characteristic Polynomial

Similar matrices and Jordan form

1 Introduction to Matrices

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression

Lecture 4: Partitioned Matrices and Determinants

Notes on Determinant

2 Polynomials over a field

IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction

Linear Algebra Review. Vectors

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

Chapter 6. Orthogonality

Direct Methods for Solving Linear Systems. Matrix Factorization

Vector and Matrix Norms

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

CS3220 Lecture Notes: QR factorization and orthogonal transformations

LECTURE 4. Last time: Lecture outline

ELA

Systems of Linear Equations

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

Recall that two vectors in are perpendicular or orthogonal provided that their dot

Mathematics Course 111: Algebra I Part IV: Vector Spaces

1 Sets and Set Notation.

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS

MATH1231 Algebra, 2015 Chapter 7: Linear maps

IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS

6. Cholesky factorization

Linear Algebra I. Ronald van Luijk, 2012

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

A note on companion matrices

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

Matrix Differentiation

NOV /II. 1. Let f(z) = sin z, z C. Then f(z) : 3. Let the sequence {a n } be given. (A) is bounded in the complex plane

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = i.

Linear Algebra: Determinants, Inverses, Rank

Math 550 Notes. Chapter 7. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010

minimal polyonomial Example

SOLVING LINEAR SYSTEMS

FOUNDATIONS OF ALGEBRAIC GEOMETRY CLASS 22

Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time

Finite dimensional C -algebras

MATH PROBLEMS, WITH SOLUTIONS

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

Continued Fractions and the Euclidean Algorithm

Introduction to Matrix Algebra

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Examination paper for TMA4205 Numerical Linear Algebra

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

[1] Diagonal factorization

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Notes on Symmetric Matrices

How To Prove The Dirichlet Unit Theorem

Numerical Methods I Eigenvalue Problems

Matrix Calculations: Applications of Eigenvalues and Eigenvectors; Inner Products

Chapter 7. Matrices. Definition. An m n matrix is an array of numbers set out in m rows and n columns. Examples. (

Solving Linear Systems, Continued and The Inverse of a Matrix

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

MA106 Linear Algebra lecture notes

Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj

Structure of the Root Spaces for Simple Lie Algebras

T ( a i x i ) = a i T (x i ).

Orthogonal Bases and the QR Algorithm

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Using row reduction to calculate the inverse and the determinant of a square matrix

Methods for Finding Bases

LS.6 Solution Matrices

A Primer on Index Notation

1 VECTOR SPACES AND SUBSPACES

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

Math 312 Homework 1 Solutions

Chapter 5. Banach Spaces

SALEM COMMUNITY COLLEGE Carneys Point, New Jersey COURSE SYLLABUS COVER SHEET. Action Taken (Please Check One) New Course Initiated

Continuity of the Perron Root

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

Finite Dimensional Hilbert Spaces and Linear Inverse Problems

A linear algebraic method for pricing temporary life annuities

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

7 Gaussian Elimination and LU Factorization

THE PROVISION OF SPARE CABLE

A =

These axioms must hold for all vectors ū, v, and w in V and all scalars c and d.

Transcription:

Lecture 6: The Group Inverse

The matrix index Let A C n n, k positive integer. Then R(A k+1 ) R(A k ). The index of A, denoted IndA, is the smallest integer k such that R(A k ) = R(A k+1 ), or equivalently, rank A k = rank A k+1 (1) holds. Then R(A j ) = R(A Ind A ), j > IndA. A matrix A is range Hermitian (or EP matrix) if R(A) = R(A ) Special cases: (a) Nonsingular matrices, (b) normal matrices, in particular (c) A = O. If A is range Hermitian, then Ind A = 1. [ ] The converse is not true, e.g., A = 1 1 0 0 2-a

Index and complementary subspaces Theorem. Let A C n n have index k, and let l be a positive integer. Then R(A l ) and N(A l ) are complementary subspaces if and only if l k. Proof. The theorem is obvious for nonsingular matrices. Let A C n n. Then, for any positive integer l, From dimr(a l ) + dimn(a l ) = rank A l + nullitya l = n. C n = R(A l ) N(A l ) R(A l ) N(A l ) = {0}. it follows that R(A l+1 ) R(A l ) and N(A l ) N(A l+1 ), C n = R(A l ) N(A l ) R(A l ) = dimr(a l+1 ). 3

Nilpotent matrices A matrix N is nilpotent if N k = O for some integer k 0. The smallest such k is called the index of nilpotency of N. Let J k (λ) be a k k Jordan block, J k (λ) = λ 1 0. λ 1............ λ 1 0 λ Then J k (λ) is nilpotent if λ = 0, indj k (λ) = 1, if λ 0; k, if λ = 0. 4

Let A C n n, X = Jordan blocks [x 1 x 2 x k ] AX = XJ k (λ), where J k (λ) = C n k k λ 1 0 0 0 λ 1.. Ax 1 = λx 1, and λ C satisfy............... 0... λ 1 0 0 λ Ax j = λx j + x j 1, j = 2,...,k. C k k, j 1, k : (A λi) j x j = 0, (A λi) j 1 x j = x 1 0, x 1 is an eigenvector of A corresponding to λ x j is a λ vector (principal vector, generalized eigenvector) of A of grade j 5

J k (λ) = Powers of the matrix (J k (λ) λi) C k k λ 1 0 0 0 λ 1................. 0... λ 1 0 0 λ (J k (λ) λi) 2 =, J k (λ) λi = 0 0 1 0 0 0 0 1........ 0 1.... 0 0 0 0 0 0 1 0 0 0 0 1................ 0.... 0 1 0 0 0, (J k (λ) λi) k = O. 6

The Jordan normal form Theorem. Let A C n n. Then A is similar to a block diagonal matrix J with Jordan blocks on its diagonal, i.e. nonsingular X J k1 (λ 1 ) O O O J X 1 k2 (λ 2 ) O AX = J =....... (1) O O J kp (λ p ) The matrix J, called the Jordan normal form of A, is unique up to a rearrangement of its blocks. The scalars {λ 1,...,λ p } in (1) are the eigenvalues of A. The set of eigenvalues, or spectrum of A, is denoted Λ(A). For A as above, (A λ 1 I) k 1 (A λ 2 I) k2 (A λ p I) k p = O. (2) 7

The polynomial Vanishing polynomials c(z) = (z λ 1 ) k 1 (z λ 2 ) k2 (z λ p ) k p is the characteristic polynomial of A. An eigenvalue λ may be repeated, say c(z) = p(z)(z λ) k 1 (z λ) k1 (z λ) k m, p(λ) 0 The algebraic multiplicity of the eigenvalue λ is then k 1 + k 2 + + k m and max {k i : i 1, m} is the geometric multiplicity, or index of the eigenvalue λ, denoted ν(λ). The polynomial m(z) = (z λ) ν(λ) λ Λ(A) is the minimal polynomial of A. 8-a

Spectral inverses Let A C n n. Then X is an S inverse of A if they share the property that, for every λ C and every vector x, x is a λ vector of A of grade p x is a λ vector of X of grade p where λ = 0 if λ = 0, and otherwise 1/λ. Ex. If A C n n is diagonable, A = PJP 1, J = diag(λ 1, λ 2,, λ n ) then X = PJ P 1, J = diag (λ 1, λ 2,, λ n) is an S inverse of A. It is a {1, 2} inverse, and commutes with A, AX = XA. 9

The group inverse Theorem. Let A C n n. Then the system has a solution X iff AXA = A, (1) XAX = X, (2) AX = XA, (5) Ind A = 1 and the solution is unique. It is called the group inverse, or {1, 2, 5} inverse, of A, and is denoted A #. Proof. X A{1, 2} AX = P R(A),N(X), XA = P R(X),N(A) X A{1, 2, 5} AX = P R(A),N(A) C n = R(A) N(A) IndA = 1 10

The group inverse (cont d) Theorem. A is range Hermitian if, and only if, A # = A. Proof. A = A (1,2) R(A),N(A), A# = A (1,2) R(A ),N(A ). Theorem (Erdélyi). Let A have index 1 and Jordan form A = PJP 1, Then A # = PJ P 1. Proof. The relations (1) AXA = A, (2) XAX = X, and (5) AX = XA, are similarity invariants. Therefore J # = P 1 A # P and since IndJ = 1, J # = J. 11-a

The group inverse (cont d) Theorem (Cline). Let a square matrix A have the FRF A = CR. Then A has group inverse if and only if RC is nonsingular, and A # = C(RC) 2 R. Ex. Let A C n n. Then A has index 1 if and only if the limit exists, in which case lim (λi n + A) 1 A λ 0 lim (λi n + A) 1 A = AA #. λ 0 Proof. A full rank factorization A = CR gives (λi n + A) 1 A = C(λI r + RC) 1 R, etc. 12

Properties of the group inverse (a) If A is nonsingular, A # = A 1. (b) A ## = A. (c) A # = A #. (d) A T# = A #T. (e) (A l ) # = (A # ) l for every positive integer l. (f) Let A have index 1 and denote Then, A j := (A # ) j, j = 1, 2,..., A 0 := AA #. A l A m = A l+m, l, m. The powers of A, positive, negative and zero, constitute an Abelian group under matrix multiplication ( group inverse ) 13

Lemma on λ vectors Lemma 1. Let x be a λ vector of A with λ 0. Then x R(A l ) where l is an arbitrary positive integer. Proof. Let (A λi) p x = 0 for some positive integer p, or, using the binomial expapsion, ( ) p x = c 1 Ax + c 2 A 2 x + + c p A p x, c i = ( 1) i 1 λ i. (1) i Ax = c 1 A 2 x + c 2 A 3 x + + c p A p+1 x, A 2 x = c 1 A 3 x + c 2 A 4 x + + c p A p+2 x, = (2) A l 1 x = c 1 A l x + c 2 A l+1 x + + c p A p+l 1 x, Successive substitutions of (2) in (1) give x = A l q(a)x, where q is some polynomial. 14

Lemma on spectral inverses Lemma 2. Let A be a square matrix and let XA l+1 = A l (1) for some positive integer l. Then every λ vector of A of grade p for λ 0 is a λ 1 vector of X of grade p. Proof by induction on the grade p. p = 1. Let Ax = λx, λ 0. Then A l+1 x = λ l+1 x x = λ l 1 A l+1 x Xx = λ l 1 XA l+1 x = λ 1 x. Assume true for p 1, r. Let x be a λ vector of A of grade r + 1. 15

Lemma 2 (cont d) Then, by Lemma 1, x = A l y, for some y. (X λ 1 I)x = (X λ 1 I)A l y = X(A l λ 1 A l+1 )y = X(I λ 1 A)A l y = λ 1 X(A λi)x. By the induction hypothesis, (A λi)x is a λ 1 vector of X of grade r. (X λ 1 I) r (A λi)x = 0, z = (X λ 1 I) r 1 (A λi)x 0 Xz = λ 1 z. (X λ 1 I) r+1 x = λ 1 X(X λ 1 I) r (A λi)x = 0, (X λ 1 I) r x = λ 1 X z = λ 2 z 0. x is a λ 1 vector of X of grade r + 1. 16

The group inverse is spectral Recall: Let A C n n. Then X is an S inverse of A if they share the property that, for every λ C and every vector x, x is a λ vector of A of grade p x is a λ vector of X of grade p where λ = 0 if λ = 0, and otherwise 1/λ. Theorem. Let A C n n have index 1. Then A # is the unique S inverse of A in A{1} A{2}. If A is diagonable, A # is the only S inverse of A. Proof. A # is an S inverse of A. Since X = A # satisfies XA l+1 = A l with l = 1, it follows from Lemma 2 that A # satisfies the = part of the definition of S inverse for λ 0. Replacing A by A # establishes the = part for λ 0, since A ## = A. The rest of the proof omitted. 17

Application to finite Markov chains A system is observed at times t = 0, 1, 2,.... The system has N states, denoted {1, 2,...,N}, the state at time t is denoted X t The system is a (finite) Markov chain (or chain) if a matrix P = (p ij ) R N N such that Prob {X t+1 = j X t = i} = p ij, i, j 1, N, t = 0, 1,.... The numbers p ij are called the transition probabilities, and the matrix P is called the transition matrix. N p ij = 1, i 1, N, (a) j=1 p ij 0, i, j 1, N. (b) A square matrix P = [p ij ] satisfying (a),(b) is called stochastic. Condition (a) is Pe = e, i.e., 1 Λ(P). 18

Let p (n) ij Markov chains (cont d) denote the n step transition probability p (n) ij = Prob {X t+n = j X t = i}, p (0) ij := δ ij. The Chapman Kolmogorov equations p (m+n) ij = N k=1 p (m) ik p(n) kj, m, n Z +. (1) p (n) ij is the (i, j) th element of P n. More terminology: (a) A state i leads to state j, denoted by i j, if p (n) ij > 0 for some n 1. (b) Two states i, j communicate, denoted by i j if each state leads to the other. (c) A set S of states is closed if i, j S = i j 19-a

Terminology (cont d) (d) A chain is irreducible if its only closed set is the set of all states 1, N, and is reducible otherwise. (e) A single state forming a closed set is an absorbing state. A reducible chain is absorbing if each of its closed sets consists of a single (necessarily absorbing) state. (f) A chain is regular if for some k, P k is a positive matrix. (g) A state i has period τ if p (n) ii = 0 except when n = τ, 2τ, 3τ,.... The period of i is denoted τ(i). If τ(i) = 1, i is aperiodic. (h) Let f (n) ij reached for 1st time at step n, and let f ij = n=1 µ ij = = probability that starting from i, state j is f (n) ij probability that j eventually reached from i, n=1 nf (n) ij the mean 1st passage time from i to j. 20

(h) µ kk := n=1 nf (n) kk Terminology (cont d) is the mean return time of state k. (i) A state i is recurrent if f ii = 1, and is transient otherwise. A chain is recurrent if all its states are recurrent. (j) If i is recurrent and µ ii =, i is called a null state. All states in a closed set are of the same type. (k) A state is ergodic if recurrent and aperiodic, but not a null state. If a chain is irreducible, all its states have the same period. An irreducible aperiodic chain is called ergodic. Theorem (Chung). For any states i, j of an irreducible recurrent chain n 1 lim n n+1 p (k) ij = 1. (1) µ jj k=0 Proof. Both sides of (1) give the expected number of returns to state j in unit time. 21-a

Stationary probabilities The probabilities {π 1, π 2,...,π N } are stationary if π k = N i=1 p ik π i, k 1, N, or, in terms of vector of stationary probabilities, P T π = π. π is called a stationary distribution or steady state. In an irreducible recurrent chain, the stationary distribution is simply, π k = 1 µ kk, k 1, N. If a chain is ergodic, the system converges to its stationary distribution from any initial distribution. In terms of P, every row of P n converges to π T. 22

Summary for ergodic chains Theorem (Feller). Let P R N N be the transition matrix of an ergodic chain. Then: (a) j, k 1, N the following limit exists, independent of j, lim n p(n) jk = π k > 0 (1) (b) π k = 1 µ kk (c) The numbers {π k : k 1, N} are probabilities, π k 0, N k=1 and are a stationary distribution of the chain: π k = N i=1 π k = 1, (2) p ik π i, k 1, N. (3) The distribution {π k } is uniquely determined by (2) and (3). 23

Solution using the group inverse Theorem (Meyer). Let P be the transition matrix of a finite Markov chain, and let Q = I P. Then Q has a group inverse, and: I QQ # = lim n 1 n + 1 n k=0 P k, (1) = lim n (αi + (1 α)p)n, for any 0 < α < 1, (2) and if the chain is regular, or absorbing, I QQ # = lim n P n. (3) For proofs and other details see original work. 24