The Full Pythagorean Theorem

Similar documents
Recall that two vectors in are perpendicular or orthogonal provided that their dot

Section Inner Products and Norms

The Determinant: a Means to Calculate Volume

Section 1.1. Introduction to R n

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Mathematics Course 111: Algebra I Part IV: Vector Spaces

NOTES ON LINEAR TRANSFORMATIONS

α = u v. In other words, Orthogonal Projection

1 Introduction to Matrices

Orthogonal Projections

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

Classification of Cartan matrices

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

Linear Algebra Notes

Inner product. Definition of inner product

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

CITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION

3. INNER PRODUCT SPACES

28 CHAPTER 1. VECTORS AND THE GEOMETRY OF SPACE. v x. u y v z u z v y u y u z. v y v z

Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday.

Matrix Representations of Linear Transformations and Changes of Coordinates

5.3 The Cross Product in R 3

Adding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors

Orthogonal Projections and Orthonormal Bases

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

Orthogonal Diagonalization of Symmetric Matrices

13.4 THE CROSS PRODUCT

Similarity and Diagonalization. Similar Matrices

6. Vectors Scott Surgent (surgent@asu.edu)

ISOMETRIES OF R n KEITH CONRAD

1 VECTOR SPACES AND SUBSPACES

5. Orthogonal matrices

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

x = + x 2 + x

Solutions to old Exam 1 problems

Numerical Analysis Lecture Notes

Chapter 6. Orthogonality

LINEAR ALGEBRA. September 23, 2010

Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj

Chapter 20. Vector Spaces and Bases

Systems of Linear Equations

Matrix Differentiation

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

WHEN DOES A CROSS PRODUCT ON R n EXIST?

Chapter 17. Orthogonal Matrices and Symmetries of Space

T ( a i x i ) = a i T (x i ).

MATH1231 Algebra, 2015 Chapter 7: Linear maps

8 Square matrices continued: Determinants

THREE DIMENSIONAL GEOMETRY

Section 4.4 Inner Product Spaces

Math 4310 Handout - Quotient Vector Spaces

Solution to Homework 2

1 Symmetries of regular polyhedra

Name: Section Registered In:

BANACH AND HILBERT SPACE REVIEW

MA106 Linear Algebra lecture notes

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

State of Stress at Point

Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product

The Characteristic Polynomial

University of Lille I PC first year list of exercises n 7. Review

Inner Product Spaces

Metric Spaces. Chapter Metrics

Vector Math Computer Graphics Scott D. Anderson

A =

DEFINITIONS. Perpendicular Two lines are called perpendicular if they form a right angle.

11.1. Objectives. Component Form of a Vector. Component Form of a Vector. Component Form of a Vector. Vectors and the Geometry of Space

Data Mining: Algorithms and Applications Matrix Math Review

Linear Algebra Review. Vectors

ON TORI TRIANGULATIONS ASSOCIATED WITH TWO-DIMENSIONAL CONTINUED FRACTIONS OF CUBIC IRRATIONALITIES.

1.3. DOT PRODUCT If θ is the angle (between 0 and π) between two non-zero vectors u and v,

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

x + y + z = 1 2x + 3y + 4z = 0 5x + 6y + 7z = 3

Inner Product Spaces and Orthogonality

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

1 Sets and Set Notation.

Grassmann Algebra in Game Development. Eric Lengyel, PhD Terathon Software

Notes on Determinant

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

BILINEAR FORMS KEITH CONRAD

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices

LINEAR ALGEBRA W W L CHEN

[1] Diagonal factorization

1 Determinants and the Solvability of Linear Systems

CONTROLLABILITY. Chapter Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Math 312 Homework 1 Solutions

4.5 Linear Dependence and Linear Independence

THE DIMENSION OF A VECTOR SPACE

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

Dot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product

Geometry and Measurement

Solving Simultaneous Equations and Matrices

Elementary Linear Algebra

Section The given line has equations. x = 3 + t(13 3) = t, y = 2 + t(3 + 2) = 2 + 5t, z = 7 + t( 8 7) = 7 15t.

17. Inner product spaces Definition Let V be a real vector space. An inner product on V is a function

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

Transcription:

The Full Pythagorean Theorem Charles Frohman January 1, 2010 Abstract This note motivates a version of the generalized pythagorean that says: if A is an n k matrix, then det(a t A) = I det(a I ) 2 where the sum is over all k k minors of A. This is followed by a proof via an elementary computation in exterior algebra. The author hopes it is accessible to anybody who has learned calculus with differential forms. 1 Introduction The pythagorean theorem is one of the first theorems of geometry that people learn. If a right triangle has legs of length a and b and its hypotenuse has length c then a 2 + b 2 = c 2. The Playfair proof of the Pythagorean theorem is easy to explain, but somehow mysterious. 1

b a a c c b b c c a a Behold! b The pythagorean theorem leads to the distance formula in the plane. If (x 1, y 1 ), (x 2, y 2 ) are two points in the plane the distance between them is (x1 x 2 ) 2 + (y 1 y 2 ) 2, which flows from the pythagorean theorem via the diagram below. (x 2, y 2 ) (x 1, y 1 ) (x 2, y 1 ) If you tell someone that the Pythagorean theorem generalizes to any dimension, often they will think that the generalization is just the distance formula for points in R n. However, suppose that T is the tetrahedron with vertices at (0, 0, 0), (a, 0, 0), (0, b, 0), and (0, 0, c). 2

(0, 0, c) (0, b, 0) (a, 0, 0) You can imagine that the tetrahedron has three legs, which are right triangles that lie in the xy-plane, xz-plane, and yz-plane. The hypotenuse is the triangle having vertices (a, 0, 0), (0, b, 0) and (0, 0, c). The sum of the squares of the areas of the three legs is 1 4 a2 b 2 + 1 4 a2 c 2 + 1 4 b2 c 2. The base of the hypotenuse is a 2 + b 2 and its height is area of the hypotenuse is one half of its base times its height, 1 a2 b 2 2 + a 2 c 2 + b 2 c 2. a 2 b 2 + c a 2 +b 2. The 2 Squaring this gives the sum of the squares of the areas of the legs! This computation is sometimes referred to as de Gua s theorem after Jean Paul de Gua de Malves who was an 18th century French mathematician [2]. The statement that this tetrahedron is right, boils down to the legs of the tetrahedron being the orthogonal projections of the hypotenuse into the coordinate hyperplanes. The shapes we have been using so far, triangles and tetrahedra, are examples of simplices. Simplices are well and good, but for the sake of discussing the pythagorean theorem in the language of linear algebra, parallelepipeds are better. Suppose that v = (a, b, c) and w = (d, e, f) are vectors in space. The parallelogram spanned by v and w is everything of the form s v + t w where 3

s, t [0, 1]. The orthogonal projection of the parallelogram into the xyplane is spanned by (a, b) and (d, e). Using the standard area formula via determinants its area is ae bd. Its orthogonal projection into the yz-plane is spanned by (b, c) and (e, f) and has area bf ce. Finally its projection into the xz-plane is spanned by (a, c) and (d, f) and has area af cd. The pythagorean theorem says that the square of the area of the parallelogram in space is the sum of the squares of the areas of the projections into the coordinate hyperplanes. (ae bd) 2 + (bf ce) 2 + (af cd) 2 but this is just the norm squared of the cross product (a, b, c) (d, e, f), which confirms a well known formula for the area of a parallelogram in space. In general, the parallelepiped spanned by vectors v 1, v 2,..., v k is everything of the form k i=1 λ i v i where the λ i vary over the unit interval [0, 1]. The pythagorean theorem says that if P is a parallelepiped in R n spanned by k-vectors v 1, v 2,..., v k then the square of the k-dimensional content of P, is the sum of the squares of the k-dimensional content of the orthogonal projections of P into the k-dimensional coordinate hyperplanes in R n. It is the point of this note to give a statement in linear algebraic terms of this theorem, and prove it. 2 Content Let V, W be innerproduct spaces, and L : V W a linear map. We can restrict L to get, L : ker L iml, where by ker L we mean the subspace of V made up of all vectors that are pependicular to the kernel of L and by iml we mean the image of L. As ker L and iml are subspaces of innerproduct spaces they are themselves innerproduct spaces. Hence we can choose orthonormal bases for them and represent L as a matrix l ij with respect to those bases. The matrix l ij is square. Definition 1. The content of L, denoted c(l) is the absolute value of the determinant of any matrix representation l ij of L with respect to orthonormal bases of the perpendicular to its kernel, and its image. 4

The content is not as well behaved as the determinant, for instance it doesn t have a sign. Also the content of the composition of two linear maps is not in general the product of their contents. However, Proposition 1. If L : V W and M : W U are linear maps of innerproduct spaces and iml = ker M then c(m L) = c(m)c(l). Proof. Since iml = ker M one of the orthonormal bases you use to compute c(l) can be chosen to coincide with one of the orthonormal bases used to compute c(m). If l ij and m jk are the matrices that represent L and M with respect to this choice, then the matrix representing M L is the product of l ij and m jk. The determinant of the product is the product of the determinants. If L : V W is a linear map of innerproduct spaces then it has an adjoint L : W V defined by requiring L to solve the equation, < L( v), w > W =< v, L ( w) > V for all v V and w W where <, > V and <, > W denote the innerproducts in V and W. If we have represented L with respect to orthonormal bases of V and W then the matrix of L with respect to those bases is just the transpose of the matrix representing L. From this it is clear that c(l) = c(l ). Using Proposition 1, and the fact that the image of L is the perpendicular to the kernel of L we arrive at c(l L ) = c(l) 2, and using the fact that L = L we have c(l L) = c(l) 2. Proposition 2. For any linear map of innerproduct spaces L : V W, c(l L) = c(l L ) = c(l) 2. Suppose f : M N is an immersion of manifolds and N is Riemannian. Choose local coordinates x 1, x 2,..., x k at p M. The matrix g ij =< df p ( x i ), df p( x j ) > T f(p) N is the first fundamental form of the induced Riemannian metric on M and the associated volume form is det(g)dx 1... dx k. 5

You can think of detg(p) as the k-dimensional content of the parallelepiped spanned by the vectors df p ( x 1 p),..., df p ( x k p) Choose an orthonormal basis for T f(p) N and let v i be the column vector representing df p ( x p) i with respect to this basis. Let A be the n k-matrix whose columns are the v i. Notice that A t A = g, so c(a) = det(g). In general, you can think of an n k matrix A as a list of k column vectors and c(a) as the k-dimensional content of the parallelepiped spanned by those vectors. Choose I {1, 2,..., n} with k-elements. We use I to denote the cardinality of I. By A I I mean the k k-matrix made from the rows of A corresponding to the subset I. The parallelepiped spanned by A I can be thought of as the orthogonal projection of the parallelepiped spanned by A into the k-dimensional coordinate hyperplane determined by I. This brings us to: Theorem 1 (Pythagorean Theorem). Let A be an n k-matrix, then det(a t A) = det(a I ) 2. I {1,...,n} I =k That is, the square of the content of the parallelepiped spanned by A is equal to the sum of the squares of the orthogonal projections of the parallelepiped into the k-dimensional coordinate hyperplanes. We will prove this theorem after we develop some vocabulary for manipulating determinants. 3 Exterior Algebra and the Proof Let V be a vector space over k. The exterior algebra Λ (V ) of V, is the associative algebra over k that is the quotient of the free unital associative 6

algebra over k on V modulo the relations that v v = 0 for every v V, where we are denoting the multiplication by a wedge [1]. It is an elementary fact that the relation v v = 0 for all v implies that v w = w v for all v and w. We will be working with innerproduct spaces. Suppose that e i i {1,..., n} is an orthonormal basis of V. If I = {i 1, i 2,..., i k } {1, 2,..., n} with i 1 < i 2... < i k let e I = e i1 e i2... e ik. If we let e = 1, then the e I where I ranges over all subsets of {1, 2,..., n} is a basis for Λ (V ) In fact, Λ (V ) is graded by I. We let Λ i (V ) be the subspace spanned by all e I with I = i. Notice Λ i (V ) has dimension ( n i). We make Λ (V ) into an innerproduct space, by declaring that the e I are an orthonormal basis. If W is any innerproduct space we denote the identity map from W to itself by Id W. If L : V W is a linear map of innerproduct spaces then for all i, it induces Λ i (L) : Λ i (V ) Λ i (W ), for all i. It is defined by letting and L( v 1 v 2... v i ) = L( v 1 ) L( v 2 )... L( v i ). The assignment L Λ i (L), is functorial as Λ i (Id V ) = Id Λi (V ), Λ i (L M) = Λ i (L) Λ i (M). where L : V W and M : W U are any linear mappings between innerproduct spaces. Proposition 3. If we choose orthonormal bases e j for V and f i for W and let l ij be the matrix of f with respect to these bases, if J = i then L(e J ) = I =i M IJ f I where M IJ is the determinant of the i i submatrix of l ij whose rows and columns come respectively from the sets I and J in order. 7

Proof. Notice that L(e j ) = i l ijf i. Expanding, Λ(L) i (e J ) = (L(e j1 )... L(e ji )) = ( k l kj1 f k ) ( k l kj2 f k )... ( k l kjk f k ) After a lot of cancellation and reordering we arrive at: = ( k ) sgn(σ) l im,jσ(m ) f I, σ S i m=1 I {1,...,n} where S i denotes the symmetric group on i letters, and sgn(σ) is the sign of the permutation sigma. Notice that the sum over S i of the signed products is a classical formula for the determinant, so it reduces to = I =i M IJ f I where the M IJ is the determinant of the minor whose entries are indexed by ij with i I and j J.. Now suppose that dim(v ) = dim(w ) = k. Notice dimλ k (V ) = dimλ k (W ) = 1, and Λ k (L)(e {1,...,k} ) = det(l ij )f {1,...,k}. Proposition 4. If L : V W is a linear map of innerproduct spaces then the map Λ i (L ) is equal to Λ i (L). Proof. Choose orthonormal bases e j and f i for V and W respectively. That way, if l ij is the matrix of L with respect to those bases then the matrix of L is l ji. Notice e J and f I are orthonormal bases of Λ i (V ) and Λ i (W ) where we let I and J range over subsets with i elements in their respective index sets. Hence if L IJ is the matrix of Λ i (L) with respect to the bases e J and f I then the matrix of Λ i (L ) is L JI. Here I am using the fact that the determinant of the transpose of a matrix is equal to the determinant of the matrix in computing the coefficients of the matrix for Λ i (L ). We are ready. 8

Proof. Theorem 1 Let A be an n k matrix, which we can view as the matrix of a linear map A : R k R n with respect to the standard orthonormal bases. The determinant of A t A is expressed as Λ k (A A)(e {1,...,k} ) = det(a t A)e {1,...,k}. By the functoriality of Λ k the left hand side is However, we have Λ k (A ) Λ k (A)(e {1,...,k} ). Λ k (A)(e {1,...,k} ) = I {1,...n} I =k} det(a I )e I, where det(a I ) is the determinant of the k k-matrix whose rows correspond to the rows of A whose index is in I. Recall, Λ k (A ) = Λ k (A). Going back to the definition of adjoint we arrive at Therefore, Λ k (A )(e I ) = det(a I )e {1,...,k}. Λ k (A ) Λ k (A)(e {1,...,k} ) = Λ k (A )( I {1,...n} I =k} I {1,...n} I =k} Putting it all together we have, det(a t A) = I {1,...n} I =k} det(a I )Λ k (A )(e I ) = det(a I ) 2 e {1,...,k}. I {1,...n} I =k} det(a I ) 2. det(a I )e I ) = 9

4 Epilogue This is a theorem that gets rediscovered over and over again. Via a web search I found a note by Alvarez [3] that proves the theorem for right n-simplices, but better than that has a nice bibliography with some references to proofs and historical texts. He cites evidence that the theorem first appeared in a book on analytic geometry by Monge and Hatchet written in the 19th century. There is also a paper by Atzema [4] that proves the same theorem I proved here, via a more general result about the determinant of a product of matrices due to Cauchy. I would be surprised if a proof along the lines that I gave here, didn t appear elsewhere. I think the theorem might be of pedagogical interest as it gives a unified paradigm for integrals to compute arc length and area that could leave the students in a position to set up integrals to compute higher dimensional content. References [1] [Sp] Spivak, Michael Calculus on Manifolds, Perseus Books, Cambridge MA, 1965 [2] [We] Weisstein, Eric W. de Gua s Theorem., From MathWorld A Wolfram Web Resource, http://mathworld.wolfram.com/deguastheorem.html [3] Alvarez, Sergio Note on an n-dimensional Pythagorean theorem http://www.cs.bc.edu/ alvarez/ndpyt.pdf [4] Atzema, Eisso, Beyond Monge s Theorem, Mathematics Magazine, Vol. 73, No. 4 (oct., 2000) pp. 293-296 10