Appendix A. Rayleigh Ratios and the Courant-Fischer Theorem



Similar documents
Chapter 6. Orthogonality

Orthogonal Diagonalization of Symmetric Matrices

Soft Clustering with Projections: PCA, ICA, and Laplacian

Conductance, the Normalized Laplacian, and Cheeger s Inequality

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Math 550 Notes. Chapter 7. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010

SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH

[1] Diagonal factorization

Similarity and Diagonalization. Similar Matrices

Inner Product Spaces and Orthogonality

Section Inner Products and Norms

Linear Algebra Review. Vectors

Spectral Analysis of Signed Graphs for Clustering, Prediction and Visualization

DATA ANALYSIS II. Matrix Algorithms

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Continuity of the Perron Root

Notes on Symmetric Matrices

Applied Linear Algebra I Review page 1

by the matrix A results in a vector which is a reflection of the given

SMOOTHING APPROXIMATIONS FOR TWO CLASSES OF CONVEX EIGENVALUE OPTIMIZATION PROBLEMS YU QI. (B.Sc.(Hons.), BUAA)

17. Inner product spaces Definition Let V be a real vector space. An inner product on V is a function

Numerical Methods I Eigenvalue Problems

Inner Product Spaces

(67902) Topics in Theory and Complexity Nov 2, Lecture 7

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

ON THE DEGREES OF FREEDOM OF SIGNALS ON GRAPHS. Mikhail Tsitsvero and Sergio Barbarossa

CS Master Level Courses and Areas COURSE DESCRIPTIONS. CSCI 521 Real-Time Systems. CSCI 522 High Performance Computing

α = u v. In other words, Orthogonal Projection

Modélisation et résolutions numérique et symbolique

SYMMETRIC EIGENFACES MILI I. SHAH

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

BANACH AND HILBERT SPACE REVIEW

Data Mining: Algorithms and Applications Matrix Math Review

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

Cheeger Inequalities for General Edge-Weighted Directed Graphs

MATH APPLIED MATRIX THEORY

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression

MCS 563 Spring 2014 Analytic Symbolic Computation Wednesday 9 April. Hilbert Polynomials

NEW VERSION OF DECISION SUPPORT SYSTEM FOR EVALUATING TAKEOVER BIDS IN PRIVATIZATION OF THE PUBLIC ENTERPRISES AND SERVICES

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8

Model order reduction via Proper Orthogonal Decomposition

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS

Max-Min Representation of Piecewise Linear Functions

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

Examination paper for TMA4205 Numerical Linear Algebra

ISOMETRIES OF R n KEITH CONRAD

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

Zachary Monaco Georgia College Olympic Coloring: Go For The Gold

Similar matrices and Jordan form

A Tutorial on Spectral Clustering

Decision-making with the AHP: Why is the principal eigenvector necessary

Sketch As a Tool for Numerical Linear Algebra

160 CHAPTER 4. VECTOR SPACES

Inner product. Definition of inner product

Visualization of General Defined Space Data

Stephanie A. Blanda 020 McAllister Building University Park, PA Webpage:

Mean value theorem, Taylors Theorem, Maxima and Minima.

Vector and Matrix Norms

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

2.3 Convex Constrained Optimization Problems

Chapter 6. Cuboids. and. vol(conv(p ))

Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization

Analysis of Internet Topologies

Data visualization and dimensionality reduction using kernel maps with a reference point

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

1 VECTOR SPACES AND SUBSPACES

x + y + z = 1 2x + 3y + 4z = 0 5x + 6y + 7z = 3

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

F. ABTAHI and M. ZARRIN. (Communicated by J. Goldstein)

3. INNER PRODUCT SPACES

Numerical Analysis Lecture Notes

Subspace Analysis and Optimization for AAM Based Face Alignment

Mathematics INDIVIDUAL PROGRAM INFORMATION Macomb1 ( )

On the representability of the bi-uniform matroid

Classifying Manipulation Primitives from Visual Data

Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday.

MATHEMATICS (MATH) 3. Provides experiences that enable graduates to find employment in sciencerelated

Numerical Analysis Lecture Notes

MATH1231 Algebra, 2015 Chapter 7: Linear maps

SALEM COMMUNITY COLLEGE Carneys Point, New Jersey COURSE SYLLABUS COVER SHEET. Action Taken (Please Check One) New Course Initiated

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Review Jeopardy. Blue vs. Orange. Review Jeopardy

WHEN DOES A CROSS PRODUCT ON R n EXIST?

Inner products on R n, and more

Pacific Journal of Mathematics

Chapter 17. Orthogonal Matrices and Symmetries of Space

x = + x 2 + x

1 Introduction to Matrices

Principal components analysis

Tutorial on Exploratory Data Analysis

CITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION

Analysis of Internet Topologies: A Historical View

Math 215 HW #6 Solutions

8.1 Min Degree Spanning Tree

Transcription:

Appendix A Rayleigh Ratios and the Courant-Fischer Theorem The most important property of symmetric matrices is that they have real eigenvalues and that they can be diagonalized with respect to an orthogonal matrix. Thus, if A is an n n symmetric matrix, then it has n real eigenvalues 1,..., n (not necessarily distinct), and there is an orthonormal basis of eigenvectors (u 1,...,u n ) (for a proof, see Gallier [6]). 211

212 APPENDIX A. RAYLEIGH RATIOS AND THE COURANT-FISCHER THEOREM Another fact that is used frequently in optimization problem is that the eigenvalues of a symmetric matrix are characterized in terms of what is known as the Rayleigh ratio, definedby R(A)(x) = x> Ax x > x, x 2 Rn,x6= 0. The following proposition is often used to prove the correctness of various optimization or approximation problems (for example PCA).

Proposition A.1. (Rayleigh Ritz) IfA is a symmetric n n matrix with eigenvalues 1 apple 2 apple apple n and if (u 1,...,u n ) is any orthonormal basis of eigenvectors of A, where u i is a unit eigenvector associated with i, then 213 max x6=0 x > x = n (with the maximum attained for x = u n ), and max x6=0,x2{u n k+1,...,u n }? x > x = n k (with the maximum attained for x = u n 1 apple k apple n 1. k ), where Equivalently, if V k is the subspace spanned by (u 1,...,u k ), then k = max x6=0,x2v k x > x, k =1,...,n.

214 APPENDIX A. RAYLEIGH RATIOS AND THE COURANT-FISCHER THEOREM For our purposes, we also need the version of Proposition A.1 applying to min instead of max. Proposition A.2. (Rayleigh Ritz) IfA is a symmetric n n matrix with eigenvalues 1 apple 2 apple apple n and if (u 1,...,u n ) is any orthonormal basis of eigenvectors of A, where u i is a unit eigenvector associated with i, then min x6=0 x > x = 1 (with the minimum attained for x = u 1 ), and min x6=0,x2{u 1,...,u i 1 }? x > x = i (with the minimum attained for x = u i ), where 2 apple i apple n. Equivalently, if W k = V k? 1 denotes the subspace spanned by (u k,...,u n ) (with V 0 =(0)), then k = min x6=0,x2w k x > x = min x6=0,x2v? k 1 x > x, k =1,...,n.

215 Propositions A.1 and A.2 together are known as the Rayleigh Ritz theorem. As an application of Propositions A.1 and A.2, we give aproofofapropositionwhichisthekeytotheproofof Theorem 2.2. Given an n n symmetric matrix A and an m m symmetric B, withm apple n, if 1 apple 2 apple apple n are the eigenvalues of A and µ 1 apple µ 2 apple appleµ m are the eigenvalues of B, thenwesaythattheeigenvalues of B interlace the eigenvalues of A if i apple µ i apple n m+i, i =1,...,m. The following proposition is known as the Poincaré separation theorem.

216 APPENDIX A. RAYLEIGH RATIOS AND THE COURANT-FISCHER THEOREM Proposition A.3. Let A be an n n symmetric matrix, R be an n m matrix such that R > R = I (with m apple n), and let B = R > AR (an m m matrix). The following properties hold: (a) The eigenvalues of B interlace the eigenvalues of A. (b) If 1 apple 2 apple apple n are the eigenvalues of A and µ 1 apple µ 2 apple appleµ m are the eigenvalues of B, and if i = µ i, then there is an eigenvector v of B with eigenvalue µ i such that Rv is an eigenvector of A with eigenvalue i. Observe that Proposition A.3 implies that 1 + + m apple tr(r > AR) apple n m+1 + + n. The left inequality is used to prove Theorem 2.2.

For the sake of completeness, we also prove the Courant Fischer characterization of the eigenvalues of a symmetric matrix. 217 Theorem A.4. (Courant Fischer) Let A be a symmetric n n matrix with eigenvalues 1 apple 2 apple apple n and let (u 1,...,u n ) be any orthonormal basis of eigenvectors of A, where u i is a unit eigenvector associated with i.ifv k denotes the set of subspaces of R n of dimension k, then k = max min W 2V n k+1 x2w,x6=0 k =min W 2V k max x2w,x6=0 x > x x > x.

218 APPENDIX A. RAYLEIGH RATIOS AND THE COURANT-FISCHER THEOREM

Bibliography [1] Nikhil Bansal, Avrim Blum, and Shuchi Chawla. Correlation clustering. Machine Learning, 56:89 113, 2004. [2] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation, 15:1373 1396,2003. [3] Fan R. K. Chung. Spectral Graph Theory, volume 92 of Regional Conference Series in Mathematics. AMS, first edition, 1997. [4] Eric D. Demaine and Nicole Immorlica. Correlation clustering with partial information. In S. Arora et al., editor, Working Notes of the 6th International Workshop on Approximation Algorithms for Combinatorial Problems, LNCS Vol. 2764, pages 1 13. Springer, 2003. [5] Jean H. Gallier. Discrete Mathematics. Universitext. Springer Verlag, first edition, 2011. 219

220 BIBLIOGRAPHY [6] Jean H. Gallier. Geometric Methods and Applications, For Computer Science and Engineering. TAM, Vol. 38. Springer, second edition, 2011. [7] Chris Godsil and Gordon Royle. Algebraic Graph Theory. GTMNo.207.SpringerVerlag,firstedition, 2001. [8] H. Golub, Gene and F. Van Loan, Charles. Matrix Computations. TheJohnsHopkinsUniversityPress, third edition, 1996. [9] Frank Harary. On the notion of balance of a signed graph. Michigan Math. J., 2(2):143 146,1953. [10] Jao Ping Hou. Bounds for the least laplacian eigenvalue of a signed graph. Acta Mathematica Sinica, 21(4):955 960, 2005. [11] Ravikrishna Kolluri, Jonathan R. Shewchuk, and James F. O Brien. Spectral surface reconstruction from noisy point clouds. In Symposium on Geometry Processing, pages11 21.ACMPress,July2004. [12] Jérôme Kunegis, Stephan Schmidt, Andreas Lommatzsch, Jürgen Lerner, Ernesto William De Luca, and Sahin Albayrak. Spectral analysis of signed graphs for clustering, prediction and visualization. In SDM 10, pages559 559,2010.

BIBLIOGRAPHY 221 [13] Jianbo Shi and Jitendra Malik. Normalized cuts and image segmentation. Transactions on Pattern Analysis and Machine Intelligence, 22(8):888 905, 2000. [14] Daniel Spielman. Spectral graph theory. In Uwe Naumannn and Olaf Schenk, editors, Combinatorial Scientific Computing. CRCPress,2012. [15] von Luxburg Ulrike. A tutorial on spectral clustering. Statistics and Computing, 17(4):395 416,2007. [16] Stella X. Yu. Computational Models of Perceptual Organization. PhD thesis, Carnegie Mellon University, Pittsburgh, PA 15213, USA, 2003. Dissertation. [17] Stella X. Yu and Jianbo Shi. Multiclass spectral clustering. In 9th International Conference on Computer Vision, Nice, France, October 13-16. IEEE, 2003.