Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components



Similar documents
Notes on Symmetric Matrices

Introduction to Matrix Algebra

Introduction to Matrix Algebra

Lecture 5: Singular Value Decomposition SVD (1)

Chapter 6. Orthogonality

Similarity and Diagonalization. Similar Matrices

Data Mining: Algorithms and Applications Matrix Math Review

by the matrix A results in a vector which is a reflection of the given

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = i.

DATA ANALYSIS II. Matrix Algorithms

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

[1] Diagonal factorization

Similar matrices and Jordan form

Orthogonal Diagonalization of Symmetric Matrices

1 Introduction to Matrices

Inner products on R n, and more

Inner Product Spaces and Orthogonality

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

Math 312 Homework 1 Solutions

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

LINEAR ALGEBRA. September 23, 2010

MATH APPLIED MATRIX THEORY

The Characteristic Polynomial

Notes on Determinant

Linear Algebra Review. Vectors

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Orthogonal Projections

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Linear Algebraic Equations, SVD, and the Pseudo-Inverse

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

α = u v. In other words, Orthogonal Projection

LINEAR ALGEBRA W W L CHEN

COLLEGE ALGEBRA 10 TH EDITION LIAL HORNSBY SCHNEIDER 1.1-1

Eigenvalues and Eigenvectors

Recall that two vectors in are perpendicular or orthogonal provided that their dot

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

NOTES ON LINEAR TRANSFORMATIONS

MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =

3 Orthogonal Vectors and Matrices

Solution to Homework 2

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Continued Fractions and the Euclidean Algorithm

Quadratic forms Cochran s theorem, degrees of freedom, and all that

9 MATRICES AND TRANSFORMATIONS

Multivariate Analysis of Variance (MANOVA): I. Theory

MAT 242 Test 2 SOLUTIONS, FORM T

1 Sets and Set Notation.

x = + x 2 + x

Chapter 17. Orthogonal Matrices and Symmetries of Space

Applied Linear Algebra I Review page 1

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

University of Lille I PC first year list of exercises n 7. Review

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Numerical Analysis Lecture Notes

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Dot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product

CONTROLLABILITY. Chapter Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

Manifold Learning Examples PCA, LLE and ISOMAP

A note on companion matrices

Section Inner Products and Norms

Orthogonal Bases and the QR Algorithm

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison

Chapter 20. Vector Spaces and Bases

Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product

Linear Algebra: Determinants, Inverses, Rank

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

LS.6 Solution Matrices

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Linear Algebra I. Ronald van Luijk, 2012

More than you wanted to know about quadratic forms

Solving Linear Systems, Continued and The Inverse of a Matrix

Systems of Linear Equations

Vector and Matrix Norms

Adding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors

Numerical Analysis Lecture Notes

Math 215 HW #6 Solutions

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression

Solving Systems of Linear Equations

Brief Introduction to Vectors and Matrices

x + y + z = 1 2x + 3y + 4z = 0 5x + 6y + 7z = 3

BILINEAR FORMS KEITH CONRAD

26. Determinants I. 1. Prehistory

160 CHAPTER 4. VECTOR SPACES

1.3. DOT PRODUCT If θ is the angle (between 0 and π) between two non-zero vectors u and v,

Linear Algebra and TI 89

Inner Product Spaces

6. Cholesky factorization

Matrix Calculations: Applications of Eigenvalues and Eigenvectors; Inner Products

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

3. INNER PRODUCT SPACES

Transcription:

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they are intimately connected with the determination of the rank of a matrix, and the factoring of a matrix into a product of matrices. - 1 -

Determinant of a Square Matrix The determinant of a matrix A, denoted A is a scalar function that is zero if and only if a matrix is of deficient rank. This fact is sufficient information about the determinant to allow the reader to continue through much of the remainder of this book. As needed, the reader should consult the more extensive treatment of determinants in the class handout on matrix methods. - 2 -

Eigenvalues Definition (Eigenvalue and Eigenvector of a Square Matrix). For a square matrix A, a scalar c and a vector v v are an eigenvalue and associated eigenvector, v, respectively, if and only if they satisfy the equation, Av = cv (1) - 3 -

Comment. Note that if Av = cv, then of course Av ( k) = c( v k) for any scalar k, so eigenvectors are not uniquely defined. They are defined only up to their shape. To avoid a fundamental indeterminacy, we normally assume them to be normalized, that is satisfy the restriction that vv = 1. - 4 -

Comment. If Av = cv, then Av cv = 0, and ( A ci) v= 0. Look at this last equation carefully. Note that A ci is a square matrix, and a linear combination of its columns is null, which means A ci is not of full rank. This implies that its determinant must be zero. So an eigenvalue c of a square matrix A must satisfy the equation A ci = 0 (2) - 5 -

Key Properties of Eigenvalues and Eigenvectors For N N matrix with eigenvalues c i and associated eigenvectors v i, the following key properties hold: 1. Tr( ) and 2. N A = ci (3) N i= 1 i= 1 A = ci (4) - 6 -

Key Properties of Eigenvalues and Eigenvectors 3. Eigenvalues of a symmetric matrix with real elements are all real. 4. Eigenvalues of a positive definite matrix are all positive. - 7 -

Key Properties of Eigenvalues and Eigenvectors 5. If a N N symmetric matrix A is positive semidefinite and of rank r, it has exactly r positive eigenvalues and p r zero eigenvalues. 6. The nonzero eigenvalues of the product AB are equal to the nonzero eigenvalues of BA. Hence the traces of AB and BA are equal. - 8 -

Key Properties of Eigenvalues and Eigenvectors 7. The eigenvalues of a diagonal matrix are its diagonal elements. 8. The scalar multiple ba has eigenvalue bc i with eigenvector v i. Proof: Avi = civ i implies immediately that ( bav ) = ( bc ) v. i i i - 9 -

Key Properties of Eigenvalues and Eigenvectors 9. Adding a constant b to every diagonal element of A creates a matrix A+ bi with eigenvalues c i + b and associated eigenvectors v i. Proof. ( + b ) = + b = c + b = ( c + b) A I v Av v v v v i i i i i i i i - 10 -

Key Properties of Eigenvalues and Eigenvectors m m i 10. A has c as an eigenvalue, and v i as its eigenvector. Proof: Consider ( 2 ) ( ) ( c ) A v = A Av = A v = Av c = i i i i i i ( 2 ) = cv c = c i i i i i The general case follows by induction. v (5) - 11 -

Key Properties of Eigenvalues and Eigenvectors 1 11. A, if it exists, has 1/ c i as an eigenvalue, and v i as its eigenvector. Proof: Av = cv = v c A Av v A v i i i i i 1 1 i = i = ici But the right side of the previous equation implies that 1 1 (1/ ci) vi = (1/ ci) A vici = A v i, or 1 A v = 1/ c v - 12 - ( ) i i i

Key Properties of Eigenvalues and Eigenvectors 12. For symmetric A, for distinct eigenvalues c i, c j with associated eigenvectors v i, v j we have v i v j. Proof: Avi = civ i, and Av j = c jv j. So vav j i = civv j i and vav i j = c jvv i j. But, since a bilinear form aab is a scalar, it is equal to its transpose, and, remembering that A= A, vav i j = v jav i = v jav. i So placing parentheses around Av expressions, we see that civv j i = cjvv i j = cjvv j i. If c i and c j are different, this implies vv = 0. j i - 13 -

Key Properties of Eigenvalues and Eigenvectors 13. For any real, symmetric A, there exists a V such that VAV = D, where D is diagonal. Moreover, any real, symmetric matrix A can be written as VDV, where contains the eigenvectors v i of A in order in its columns, and D contains the eigenvalues c i of A in the ith diagonal position. - 14 -

Key Properties of Eigenvalues and Eigenvectors 14. Suppose that the eigenvectors and eigenvalues of A are ordered in the matrices V and D in descending order, so that the first element of D is the largest eigenvalue of A, and the first column of V is its corresponding * eigenvector. Define V as the first m columns of V, and * D as an m m diagonal matrix with the corresponding m eigenvalues as diagonal entries. Then * * VDV * (6) is a matrix of rank m that is the best possible (in the least squares sense) rank m approximation of A. - 15 -

Key Properties of Eigenvalues and Eigenvectors 15. Consider all possible normalized quadratic forms in A, i.e., q ( x ) = x Ax (7) i i i with xx i i = 1. The maximum of all quadratic forms is achieved with x i = v 1, where v 1 is the eigenvector corresponding to the largest eigenvalue of A. The minimum is achieved with xi = v m, the eigenvector corresponding to the smallest eigenvalue of A. - 16 -

Applications of Eigenvalues and Eigenvectors 1. Principal Components From property 15 in the preceding section, it follows directly that the maximum variance linear composite of a set of variables is computed with linear weights equal to the first eigenvector of Σ yy, since the variance of this linear combination is a quadratic form in Σ yy. - 17 -

2. Matrix Factorization Diagonal matrices act much more like scalars than most matrices do. For example, we can define fractional powers of diagonal matrices, as well as positive powers. Specifically, if diagonal matrix D has diagonal elements x d i, the matrix D has elements d. If x is negative, it is x assumed D is positive definite. With this definition, the powers of D behave essentially like scalars. For example, 1/2 D D 1/2 = D. x i - 18 -

Example. Suppose we have Then 4 0 D = 0 9 1/2 2 0 D = 0 3-19 -

Example. Suppose you have a variance-covariance matrix Σ for some statistical population. Assuming Σ is positive semidefinite, then (from Property 13 on page 14 it can be 1/2 written in the form Σ = VDV = FF, where F= VD is called a Gram-factor of F. Comment. Gram-factors are not, in general, uniquely defined. - 20 -

Example. Suppose Σ = FF. Then consider any orthogonal matrix T, conformable with F, such that TT = T T = I. There are infinitely many orthogonal matrices of order 2 2 and higher. Then for any such matrix T, we have Σ = FTT F = F * F * (8) where F * = FT. - 21 -

Applications of Gram-Factors Gram-factors have some significant applications. For example, in the field of random number generation, it is relatively easy to generate pseudo-random numbers that mimic p variables that are independent with zero mean and unit variance. But suppose we wish to mimic p variables that are not independent, but have variancecovariance matrix Σ? The following result describes one method for doing this. - 22 -

Result. Given p 1 random vector x having variance-covariance matrix I. Let F be a Gram-factor of Σ = FF. Then y = Fx will have variance-covariance matrix Σ. So if we want to create random numbers with a specific covariance matrix, we take a vector of independent random numbers, and premultiply it by F. - 23 -

Symmetric Powers of a Symmetric Matrix In certain intermediate and advanced derivations in matrix algebra, reference is made to symmetric powers of a symmetric matrix Σ, in particular the symmetric 1/2 square root Σ of Σ, a symmetric matrix which, when multiplied by itself, yields Σ. Recall that 1/2 1/2 1/2 Σ = VDV = VD D V. Note that VD V is a symmetric square root of Σ, i.e., 1/2 1/2 VD V VD V = VDV - 24 -

Orthogonalizing a Set of Variables Consider a random vector x withvar( x) = Σ I. What is 1/2 1/2 Var Σ x? How might you compute Σ? ( ) Suppose a set of variables x have a covariance matrix A, and you want to linearly transform them so that they have a covariance matrix B. How could you do that if you had a computer program that easily gives you the eigenvectors and eigenvalues of A and B? (Hint: First orthogonalize them. Then transform the orthogonalized variables to a covariance matrix you want.) - 25 -