V = [v 1, v 2,... v n ].


 Rosamund Harrington
 1 years ago
 Views:
Transcription
1 COORDINATES, EIGENVALUES, AND EIGENVECTORS Review of coordinates The standard basis The standard basis for R n is the basis E = {e, e, e n }, where e j is the vector with in the jth position and zeros elsewhere Coordinates with respect to a basis Let B = {v, v, v n } be a basis for R n We know that for each vector x R n there are unique scalar coordinates c, c, c n such that () x = c v + c v + + c n v n We call the vector c R n having coordinates c, c c n the coordinate vector of x with respect to the basis B, and denote it by x B 3 First examples (a) x E = x for each x R n (b) For any basis B = {v, v, v n } of R n : v j B = e j ( j n) (c) 0 B = 0 for any basis B (the zero vector has the same representation with respect to any basis 4 Matrix calculation of coordinates Letting B = {v, v, v n } be a basis for R n we can rewrite equation () in matrix form as () x = V x B where V is the matrix having the vector v j for its jth column, ie V = v, v, v n Date: March 6, 005
2 EIGENVALUES/EIGENVECTORS V is invertible because its columns are linearly independent (see, for example, Summary 39(v) of our textbook, pp 9 0), so we can compute the B coordinate representation of any vector x R n by solving () for x B : (3) x B = V x For example, if in R we take B = {v, v } to be our basis, where v = and 3 v =, then 5 5 V = and V 3 5 = 3 Suppose, for instance, that x = Then 5 x B = V x = 3 More generally, if x = x 7 = 4 is any vector in R x n then the same calculation shows that 5x + x x B = 3x x 5 The matrix of a linear transformation with respect to a basis Recall that each n n matrix induces a linear transformation T : R n R n by means of matrix multiplication: (4) T (x) = Ax (x R n ) We refer to A as the standard matrix for T The lesson of what s to follow is that it s not always the best matrix for T! For a basis B = {v, v, v n } for R n we define (5) T B = T (v ) B, T (v ) B T (v n ) B, ie T B is the matrix that has the Bcoordinate representation of T (v j ) as its jth column ( j n) We call T B the matrix of T with respect to the basis B 6 Example: B = E, the standard basis In this case T B = Ae, Ae, Ae n, the standard matrix for T
3 EIGENVALUES/EIGENVECTORS 3 7 Geometric examples in R In this next group of examples, L is a line through the origin in the direction of a nonzero vector v R, and w is a nonzero vector perpendicular to v B = {v, w} We take as our basis for R, not the standard one, but instead (a) Suppose T : R R is the transformation of orthogonal projection onto L Then T (v) = v and T (w) = 0, so T B = T (v) B, T (w) B = v B, 0 B = e, 0 = where we use the fact that v B = e and 0 B = 0 (see 3 (b) and (c)) Compare this simple and informative matrix for T with the one we got when we first introduced orthogonal projection Then we took v to be a unit vector with coordinates u and u, and did this computation: u T (x) = (x v)v = (x u + x u ) u = u x + u u x u u u x + u = u u x u u u, x, x so the standard matrix for the linear transformation of orthogonal projection onto L is the completely uninformative one u A = u u u u u (b) Suppose T is reflection in L Then T (v) = v as before, but now T (w) = w Thus T B = T (v) B, T (w) B = v B, w B = e, e = 0 0 Note that in this computation we used the fact that w B = w B ( linearity of coordinates Fact 34, page 39 of our textbook) and that v B = e and w B = e ( 3 (b)) (c) Suppose T is a shear in the direction of L, ie that T (v) = v and T (w) = kv + w for some fixed real number k Then using linearity of coordinates and 3(b) we see that k T B = 0 which is the same as the standard matrix for a horizontal shear
4 4 EIGENVALUES/EIGENVECTORS 8 Using T B to compute T (x) The utility of the matrix of a linear transformation T with respect to a basis B lies in the following formula: (6) T B x B = T (x) B (x R n ), which says that, if you re willing to give up the standard basis fro R n, and do all your calculations relative to the basis B, then you can use the matrix of T with respect to B in the same way you used to use the standard matrix Proof of formula (6): For x R n we have the unique representation x = c v + c v + + c n v n where c, c,, c n are the (standard) coordinates of x B By linearity of T : T (x) = c T (v ) + c T (v ) + + c n T (v n ), so by linearity of coordinates: T (x) B = c T (v ) B + c T (v ) B + + c n T (v n ) B, But the expression on the right is just the matrix product T B x B, so we are done Eigenvalues and Eigenvectors Definition Suppose A is an n n matrix and T the associated linear transformation of R n (so T (x) = Ax for every x R n ) We call a real number λ an eigenvalue of T (or of A) whenever there is a nonzero vector x R n such that T (x) = λx The vector x is called an eigenvector of T (or of A) corresponding to the eigenvalue λ Examples (a) Look back at the examples of 7, where L is a line through the origin in the direction of a nonzero vector v, and w is a nonzero vector perpendicular to v If T : R R is orthogonal projection onto L, then v and w are eigenvectors of T corresponding respectively to eigenvalues and 0 If T is reflection in L then v and w are still eigenvectors, but now the corresponding eigenvalues are and If T is a shear along L then v is an eigenvector with eigenvalue, but w is not an eigenvector
5 EIGENVALUES/EIGENVECTORS 5 (b) (see Problem 5, page 46) Suppose A = and v =, w = 4 3 Then one computes that Av = 5v and Aw = w, so v is an eigenvector for A corresponding to the eigenvalue 5 and w is an eigenvector corresponding to the eigenvalue Note that, with respect to the basis B = {v, w} of R, the matrix 5 0 for the linear transformation induced by A is just 0 (c) (see Problem 9 of 34) Here T is the linear transformation of R 3 induced by the matrix A = We are given a basis B = {v, v, v 3 } for R 3, where v =, v =, and v 3 = You can check easily that each of these basis vectors is an eigenvector: Av = 0 (eigenvalue zero), Av = v (eigenvalue ) and Av 3 = v 3 (eigenvalue ), and that therefore the matrix of the associated linear transformation T with respect to the basis B is T B = In each of the examples of the previous two sections, whenever we had a linear transformation T of R n that bequeathed to R n a basis B of eigenvectors, the matrix of T with respect to B turned out to be diagonal, ie all entries zero off the main diagonal 3 Notation The n n diagonal matrix with entry λ j in the jth diagonal position will be denoted diag {λ, λ, λ n } Thus, for example, the 3 3 identity matrix is diag {,, }, while in (c), T B = diag {0,, } Diagonal matrices are the easiest ones for which to find eigenvalues and eigenvectors: 4 Proposition If A = diag {λ, λ, λ n } then the standard unit vector e j is an eigenvector for the eigenvalue λ j ( j n)
6 6 EIGENVALUES/EIGENVECTORS Proof Our hypothesis is that A = λ e, λ e,, λ n e n, ie that A is the matrix with λ j e j as its jth column ( j n) But Ae j is the jth column of A, so Ae j = λ j e j for each j between and n 5 Exercise Prove the converse: If an n n matrix A has the standard basis for R n as eigenvectors, then A is diagonal 6 The Diagonalization Theorem For a linear transformation T : R n R n, and a basis B for R n, the following two statements are equivalent (each implies the other): (a) B consists of eigenvectors of T (b) T B is a diagonal matrix Proof (a) (b): Suppose B = {v, v, v n } consists of eigenvectors of T, so that there are real numbers λ j ( j n) such that T (v j ) = λ j v j (j =,,, n) We must show that T B is diagonal Starting with the definition of T B, we compute: T B = T v B, T v B,, T v n B = λ v B, λ v B,, λ n v n B = λ v B, λ v B,, λ n v n B = λ e, λ e,, λ n e n = diag {λ, λ,, λ n } where, in this computation we used successively: the definition of T B (line ), the fact that T (v j ) = λ j v j (line ), linearity of coordinates (line 3), and finally the fact that v j B = e j (line 4) (b) (a): Now suppose B = {v, v, v n } is a basis for R n, and that T B = diag {λ, λ,, λ n } We must show that each v j is an eigenvector of T In fact, we ll see that T v j = λ j v j for j n Fix such a j, and note that the jth column of T B is just λ j e j, ie, () T B e j = λ j e j
7 EIGENVALUES/EIGENVECTORS 7 Now the righthand side of () is λ j v j B which, by linearity of coordinates, is λ j v j B Similarly, the lefthand side of () is T B v j B which, by (6), is T v j B Thus T v j B = λ j v j B, ie, the both T v j and λ j v j have the same Bcoordinate vectors Thus T v j = λ j v j, as desired 7 Definition Call an n n matrix diagonalizable if R n has a basis consisting of eigenvectors of this matrix Warning: Not all square matrices are diagonalizable! For example it s geometrically clear that shear transfomations of R have (up to multiplication by a nonzero constant) just a single eigenvector, while rotations of R through any angle not an integer multiple of π have no eigenvectors at all! We ll see how to prove this kind of result analytically (ie without reference to pictures) in the next section 3 Efficient eigenvalue computation for matrices We ll base our computations for matrices on the notion of determinant a b 3 Definition The determinant of the matrix A = is: c d det(a) = ad bc The key to all the work of this section is: 3 Theorem For A a matrix, the system of linear equations Ax = 0 has nontrivial solution if and only if det(a) = 0 Proof In scalar form the system of equations is: ax + bx = 0 cx + dx = 0 Suppose first that neither a nor c is zero Then we can replace the second equation by c times the first minus a times the second, yielding the equivalent system: ax + bx = 0 (ad bc)x = 0
8 8 EIGENVALUES/EIGENVECTORS which has a nontrivial solution if and only if ad bc = 0 Suppose a = 0 Then det(a) = bc, which is zero precisely when either b or c is zero If b = 0 then we are back to one equation in two unknowns, so there are nontrivial (infinitely many) solutions If b 0 then any solution must have x = 0, whereupon the second equation implies cx = 0 If c = 0 (ie det(a) = 0) then there are infinitely many solutions: x = 0 and x arbitrary If, however c 0 (ie, det(a) 0) then x is also required to be zero, hence the only solution is the trivial one 33 Corollary The eigenvalues of a matrix A are the real numbers λ that satisfy its characteristic equation det(a λi) = 0 Proof This follows immediately from the above Theorem and the definition of eigenvalue More precisely, to say a real number λ is an eigenvalue of A means that there is a nonzero vector x R such that: Ax = λx ie Ax λx = 0 ie (A λi)x = 0 Thus λ is an eigenvector of A if and only if the homogeneous system (A λi)x = 0 has a nontrivial solution, which, by Theorem 3, happens if and only if det(a λi) = 0 34 Example Show that the matrix is diagonalizable A = 4 3 Proof Of course we know this already from (b) where the eigenvectors and eigenvalues of A were handed to us But suppose they weren t Then we write the characteristic equation: λ 0 = det(a λi) = det( 4 3 λ = ( λ)(3 λ) 8 = λ 4λ 5 = (λ 5)(λ + ) )
9 EIGENVALUES/EIGENVECTORS 9 so that λ = 5 and λ = are the solutions, and so these are the eigenvalues of A Now for λ = 5, 4 A λi = A 5I = 4 and the system (A 5I)x = 0 is easily seen to have solutions x = any scalar multiple of In particular, v = is the unique (up to multiplication by a nonzero scalar) eigenvector for the eigenvalue 5 I leave it to you to show by the same method that v = eigenvector for the eigenvalue is the unique It s easy to check that B = {v, v } is a basis for R, so A is diagonalizable, and the matrix of its linear transformation, with respect to the basis B is diag {5, } 0 35 Example: a nondiagonalizable matrix Let A = The characteristic 0 0 equation for this matrix is λ = 0, which has only λ = 0 as its solution The corresponding eigenvectors are the nontrivial solutions to the equation Ax = 0 (ie the nonzero vectors in the kernel of A), and these all the nonzero scalar multiples of the standard unit vector e Thus there aren t enough eigenvectors here to make a basis of R, so A is not diagonalizable 4 Matrix of T B vs standard matrix of T Suppose T is a linear transformation of R n whows standard matrix is A, ie T (x) = Ax (x R n ) Suppose B = {v, v, v n } is a basis for R n It s time to get serious about the relationship between A and T B Recall that in 8 we derived equation (3): x B = V x where V = v, v,, v n is the matrix having the vector v j as its jth column Recall also that the invertibility of V derives from the fact that its columns are linearly independent
10 0 EIGENVALUES/EIGENVECTORS Thus we may compute: T B = T (v ) B, T (v ) B,, T (v n ) B = Av B, Av B,, Av n B = V Av, V Av,, V Av n = V A v, v,, v n = V AV where the first line is the definition of T B, the second uses the definition of T in terms of A, the third uses (3), and the fourth is the definition of matrix multiplication We have proved: 4 Theorem (Fact 344, Page 44 of our text) Suppose B = {v, v, v n } is a basis for R n and T a linear transformation of R n with standard matrix A Then T B = V AV, where V is the matrix whose jth column is v j ( j n) 5 Application: Powers of matrices Theorem 4 is of great utility in computing large powers of matrices a calculation that would be difficult to do by hand in many circumstances The method depends on two crucial facts 5 Theorem If A, B and V are n n matrices, and V is invertible, then for each positive integer p: (V BV ) p = V B p V Proof We use the method of Mathematical Induction, which asserts that to prove the theorem we need only check that it s true for p =, and then show that: (*) Result true for an integer p Result true for p + The point is that once you ve established the case p = and proved (*), then (*) gives you for free the case p =, and then the case p = 3, and so on forever
11 EIGENVALUES/EIGENVECTORS For our result, the case p = is obviously true, so suppose we have proved the theorem for some p We must show that it s true for p + ; to this end write (V BV ) p+ = (V BV ) p (V BV ) = (V B p V )(V BV ) = (V B p )(V V )(BV ) = (V B p )I(BV ) = V B p+ V where the second line follows from our induction hypothesis that the theorem is true for p, the third line follows from the associate law of matrix multiplication, and the others are clear The next result asserts that for a diagonal matrix you can compute a power by taking the corresponding power of each element of the matrix 5 Theorem For any positive integer p, diag {λ, λ,, λ n } p = diag {λ p, λ p,, λ p n} Proof I leave this one mostly as an exercise First check that diag {λ, λ,, λ n }diag {µ, µ,, µ n } = diag {λ µ, λ µ,, λ n µ n } The result about powers then follows quickly from this by induction 5 53 Example Compute 4 3 Solution Recall that we ve already found that the vectors v = and v = are eigenvectors for A = (for eigenvalues 5 and  respectively) and that B = 4 3 {v, v } is a basis for R Thus, according to Theorem 4 above, 5 0 = T 0 B = V AV, where V = v, v = Thus A = V V,
12 EIGENVALUES/EIGENVECTORS so successive application of Theorem 5 and Theorem 5 yield, (5) A 5 = V V = V 0 ( ) 5 V A little calculation shows that V = /3 /3 /3 /3 more calculation (and the fact that ( ) 5 = ) yield A 5 = , after which (5) and a little 54 Example: A dynamical system Neighboring countries Elbonia and Pilikia are undergoing a relatively peaceful period in their histories Each is reducing its annual defense budget, but because of nervousness about the others intentions, each adds back in to its budget a percentage of its neighbors expenditures More precisely, Pilikia reduces its budget by 5%, but adds back in 5% of Elbonia s budget, while Elbonia reduces its budget by a whopping 75%, but adds back in 75% of Pilikia s defense budget Let P (t) denote Pilikia s defense budget for year t, and E(t) Elbonia s (in 005 Euros times one billion) Suppose initially P (0) = 3 and E(0) = Determine the behavior of P (t) and E(t) as t P (t) Solution The vector x(t) = describes the state of the system at year t We have E(t) (with amounts calculated in billions of Euros) ie, P (t + ) = 75P (t) + 5E(t) E(t + ) = 5P (t) + 75E(t) (5) x(t + ) = Ax(t) where A = Thus defense expenditures for both countries will be: x() = Ax(0) in year one, x() = Ax() = A x(0) in year two, and more generally in year n they will be x(n) = Ax(n ) = A x(n ) = A n x(0)
13 So it all comes down to computing A n EIGENVALUES/EIGENVECTORS 3 For this I leave it to you to check that the eigenvalues of A are and /, and the corresponding eigenvectors are v = for λ = and v = for λ = / Now B = {v, v } is a basis for R relative to which the linear transformation T whose standard matrix is A has a diagonal matrix: 0 T B = = V 0 / AV, where V = v, v = Thus for each positive integer n, n (/) n = 0 / hence solving for A n : Now one calculates easily that A n = V V = n = V A n V, n 0 0 (/) n V / / / / from which follows A n = + n n n + n Thus x(n) = A n x(0) = ( + n ) n 3 n + n = 5 + n 5 n 5/ Thus x(n) as n, ie both nations tend toward a yearly defense budget 5/ of 5 billion Euros I 7 # 4, 8, 9,, 36, 38 6 Exercises II 7 #8; this will be part of a future writing assignment III 73 #40, 4 These will be part of a future writing assignment 54 of these notes works out a special case of #40
1 Eigenvalues and Eigenvectors
Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationLectures notes on orthogonal matrices (with exercises) 92.222  Linear Algebra II  Spring 2004 by D. Klain
Lectures notes on orthogonal matrices (with exercises) 92.222  Linear Algebra II  Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n realvalued matrix A is said to be an orthogonal
More informationPresentation 3: Eigenvalues and Eigenvectors of a Matrix
Colleen Kirksey, Beth Van Schoyck, Dennis Bowers MATH 280: Problem Solving November 18, 2011 Presentation 3: Eigenvalues and Eigenvectors of a Matrix Order of Presentation: 1. Definitions of Eigenvalues
More informationSummary of week 8 (Lectures 22, 23 and 24)
WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry
More informationEIGENVALUES AND EIGENVECTORS
Chapter 6 EIGENVALUES AND EIGENVECTORS 61 Motivation We motivate the chapter on eigenvalues b discussing the equation ax + hx + b = c, where not all of a, h, b are zero The expression ax + hx + b is called
More informationDiagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions
Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential
More informationChapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
More informationUsing determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible:
Cramer s Rule and the Adjugate Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible: Theorem [Cramer s Rule] If A is an invertible
More informationC 1 x(t) = e ta C = e C n. 2! A2 + t3
Matrix Exponential Fundamental Matrix Solution Objective: Solve dt A x with an n n constant coefficient matrix A x (t) Here the unknown is the vector function x(t) x n (t) General Solution Formula in Matrix
More informationMATH 240 Fall, Chapter 1: Linear Equations and Matrices
MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted
More information8 Square matrices continued: Determinants
8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You
More information7  Linear Transformations
7  Linear Transformations Mathematics has as its objects of study sets with various structures. These sets include sets of numbers (such as the integers, rationals, reals, and complexes) whose structure
More informationAu = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
More informationMath 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = 36 + 41i.
Math 5A HW4 Solutions September 5, 202 University of California, Los Angeles Problem 4..3b Calculate the determinant, 5 2i 6 + 4i 3 + i 7i Solution: The textbook s instructions give us, (5 2i)7i (6 + 4i)(
More informationProblems for Advanced Linear Algebra Fall 2012
Problems for Advanced Linear Algebra Fall 2012 Class will be structured around students presenting complete solutions to the problems in this handout. Please only agree to come to the board when you are
More information1 Spherical Kinematics
ME 115(a): Notes on Rotations 1 Spherical Kinematics Motions of a 3dimensional rigid body where one point of the body remains fixed are termed spherical motions. A spherical displacement is a rigid body
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 34 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationMATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006
MATH 511 ADVANCED LINEAR ALGEBRA SPRING 26 Sherod Eubanks HOMEWORK 1 1.1 : 3, 5 1.2 : 4 1.3 : 4, 6, 12, 13, 16 1.4 : 1, 5, 8 Section 1.1: The EigenvalueEigenvector Equation Problem 3 Let A M n (R). If
More informationChapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
More informationSolutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
More informationNotes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 918/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
More informationby the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the yaxis We observe that
More informationLS.6 Solution Matrices
LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions
More information1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
More informationSimilar matrices and Jordan form
Similar matrices and Jordan form We ve nearly covered the entire heart of linear algebra once we ve finished singular value decompositions we ll have seen all the most central topics. A T A is positive
More information1 2 3 1 1 2 x = + x 2 + x 4 1 0 1
(d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which
More informationUNIT 2 MATRICES  I 2.0 INTRODUCTION. Structure
UNIT 2 MATRICES  I Matrices  I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress
More informationNotes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
More informationUniversity of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationWHICH LINEARFRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE?
WHICH LINEARFRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE? JOEL H. SHAPIRO Abstract. These notes supplement the discussion of linear fractional mappings presented in a beginning graduate course
More informationSolution based on matrix technique Rewrite. ) = 8x 2 1 4x 1x 2 + 5x x1 2x 2 2x 1 + 5x 2
8.2 Quadratic Forms Example 1 Consider the function q(x 1, x 2 ) = 8x 2 1 4x 1x 2 + 5x 2 2 Determine whether q(0, 0) is the global minimum. Solution based on matrix technique Rewrite q( x1 x 2 = x1 ) =
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationis in plane V. However, it may be more convenient to introduce a plane coordinate system in V.
.4 COORDINATES EXAMPLE Let V be the plane in R with equation x +2x 2 +x 0, a twodimensional subspace of R. We can describe a vector in this plane by its spatial (D)coordinates; for example, vector x 5
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationInverses and powers: Rules of Matrix Arithmetic
Contents 1 Inverses and powers: Rules of Matrix Arithmetic 1.1 What about division of matrices? 1.2 Properties of the Inverse of a Matrix 1.2.1 Theorem (Uniqueness of Inverse) 1.2.2 Inverse Test 1.2.3
More information(a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular.
Theorem.7.: (Properties of Triangular Matrices) (a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular. (b) The product
More information1 Orthogonal projections and the approximation
Math 1512 Fall 2010 Notes on least squares approximation Given n data points (x 1, y 1 ),..., (x n, y n ), we would like to find the line L, with an equation of the form y = mx + b, which is the best fit
More informationMatrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc.
2 Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Theorem 8: Let A be a square matrix. Then the following statements are equivalent. That is, for a given A, the statements are either all true
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus ndimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More information4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
More informationDeterminants. Dr. Doreen De Leon Math 152, Fall 2015
Determinants Dr. Doreen De Leon Math 52, Fall 205 Determinant of a Matrix Elementary Matrices We will first discuss matrices that can be used to produce an elementary row operation on a given matrix A.
More informationIterative Methods for Computing Eigenvalues and Eigenvectors
The Waterloo Mathematics Review 9 Iterative Methods for Computing Eigenvalues and Eigenvectors Maysum Panju University of Waterloo mhpanju@math.uwaterloo.ca Abstract: We examine some numerical iterative
More information4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns
L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows
More information2.1: Determinants by Cofactor Expansion. Math 214 Chapter 2 Notes and Homework. Evaluate a Determinant by Expanding by Cofactors
2.1: Determinants by Cofactor Expansion Math 214 Chapter 2 Notes and Homework Determinants The minor M ij of the entry a ij is the determinant of the submatrix obtained from deleting the i th row and the
More informationSergei Silvestrov, Christopher Engström, Karl Lundengård, Johan Richter, Jonas Österberg. November 13, 2014
Sergei Silvestrov,, Karl Lundengård, Johan Richter, Jonas Österberg November 13, 2014 Analysis Todays lecture: Course overview. Repetition of matrices elementary operations. Repetition of solvability of
More informationINTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL
SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics
More informationPROVING STATEMENTS IN LINEAR ALGEBRA
Mathematics V2010y Linear Algebra Spring 2007 PROVING STATEMENTS IN LINEAR ALGEBRA Linear algebra is different from calculus: you cannot understand it properly without some simple proofs. Knowing statements
More informationMATH10212 Linear Algebra B Homework 7
MATH22 Linear Algebra B Homework 7 Students are strongly advised to acquire a copy of the Textbook: D C Lay, Linear Algebra and its Applications Pearson, 26 (or other editions) Normally, homework assignments
More informationNON SINGULAR MATRICES. DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that
NON SINGULAR MATRICES DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that AB = I n = BA. Any matrix B with the above property is called
More informationAdditional Topics in Linear Algebra Supplementary Material for Math 540. Joseph H. Silverman
Additional Topics in Linear Algebra Supplementary Material for Math 540 Joseph H Silverman Email address: jhs@mathbrownedu Mathematics Department, Box 1917 Brown University, Providence, RI 02912 USA Contents
More informationCHARACTERISTIC ROOTS AND VECTORS
CHARACTERISTIC ROOTS AND VECTORS 1 DEFINITION OF CHARACTERISTIC ROOTS AND VECTORS 11 Statement of the characteristic root problem Find values of a scalar λ for which there exist vectors x 0 satisfying
More information(January 14, 2009) End k (V ) End k (V/W )
(January 14, 29) [16.1] Let p be the smallest prime dividing the order of a finite group G. Show that a subgroup H of G of index p is necessarily normal. Let G act on cosets gh of H by left multiplication.
More information4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
More informationFacts About Eigenvalues
Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v
More informationExamination paper for TMA4115 Matematikk 3
Department of Mathematical Sciences Examination paper for TMA45 Matematikk 3 Academic contact during examination: Antoine Julien a, Alexander Schmeding b, Gereon Quick c Phone: a 73 59 77 82, b 40 53 99
More informationDot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product
Dot product and vector projections (Sect. 12.3) Two definitions for the dot product. Geometric definition of dot product. Orthogonal vectors. Dot product and orthogonal projections. Properties of the dot
More informationChapter 5. Matrices. 5.1 Inverses, Part 1
Chapter 5 Matrices The classification result at the end of the previous chapter says that any finitedimensional vector space looks like a space of column vectors. In the next couple of chapters we re
More informationSolutions to Linear Algebra Practice Problems
Solutions to Linear Algebra Practice Problems. Find all solutions to the following systems of linear equations. (a) x x + x 5 x x x + x + x 5 (b) x + x + x x + x + x x + x + 8x Answer: (a) We create the
More informationOn the general equation of the second degree
On the general equation of the second degree S Kesavan The Institute of Mathematical Sciences, CIT Campus, Taramani, Chennai  600 113 email:kesh@imscresin Abstract We give a unified treatment of the
More informationDeterminants. Chapter Properties of the Determinant
Chapter 4 Determinants Chapter 3 entailed a discussion of linear transformations and how to identify them with matrices. When we study a particular linear transformation we would like its matrix representation
More informationMATH 551  APPLIED MATRIX THEORY
MATH 55  APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationIterative Techniques in Matrix Algebra. Jacobi & GaussSeidel Iterative Techniques II
Iterative Techniques in Matrix Algebra Jacobi & GaussSeidel Iterative Techniques II Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin
More informationThe Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
More information160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationLinear Dependence Tests
Linear Dependence Tests The book omits a few key tests for checking the linear dependence of vectors. These short notes discuss these tests, as well as the reasoning behind them. Our first test checks
More information1 Quiz on Linear Equations with Answers. This is Theorem II.3.1. There are two statements of this theorem in the text.
1 Quiz on Linear Equations with Answers (a) State the defining equations for linear transformations. (i) L(u + v) = L(u) + L(v), vectors u and v. (ii) L(Av) = AL(v), vectors v and numbers A. or combine
More informationSection 6.1  Inner Products and Norms
Section 6.1  Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
More informationName: Section Registered In:
Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are
More informationMarkov Chains, Stochastic Processes, and Advanced Matrix Decomposition
Markov Chains, Stochastic Processes, and Advanced Matrix Decomposition Jack Gilbert Copyright (c) 2014 Jack Gilbert. Permission is granted to copy, distribute and/or modify this document under the terms
More informationCONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation
Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in
More information4.6 Null Space, Column Space, Row Space
NULL SPACE, COLUMN SPACE, ROW SPACE Null Space, Column Space, Row Space In applications of linear algebra, subspaces of R n typically arise in one of two situations: ) as the set of solutions of a linear
More informationd(f(g(p )), f(g(q))) = d(g(p ), g(q)) = d(p, Q).
10.2. (continued) As we did in Example 5, we may compose any two isometries f and g to obtain new mappings f g: E E, (f g)(p ) = f(g(p )), g f: E E, (g f)(p ) = g(f(p )). These are both isometries, the
More informationB such that AB = I and BA = I. (We say B is an inverse of A.) Definition A square matrix A is invertible (or nonsingular) if matrix
Matrix inverses Recall... Definition A square matrix A is invertible (or nonsingular) if matrix B such that AB = and BA =. (We say B is an inverse of A.) Remark Not all square matrices are invertible.
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An ndimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0534405967. Systems of Linear Equations Definition. An ndimensional vector is a row or a column
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More information2.1: MATRIX OPERATIONS
.: MATRIX OPERATIONS What are diagonal entries and the main diagonal of a matrix? What is a diagonal matrix? When are matrices equal? Scalar Multiplication 45 Matrix Addition Theorem (pg 0) Let A, B, and
More informationTopic 1: Matrices and Systems of Linear Equations.
Topic 1: Matrices and Systems of Linear Equations Let us start with a review of some linear algebra concepts we have already learned, such as matrices, determinants, etc Also, we shall review the method
More informationFactorization Theorems
Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization
More informationLecture 1: Schur s Unitary Triangularization Theorem
Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections
More informationMatrix Inverse and Determinants
DM554 Linear and Integer Programming Lecture 5 and Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1 2 3 4 and Cramer s rule 2 Outline 1 2 3 4 and
More informationLinear Codes. In the V[n,q] setting, the terms word and vector are interchangeable.
Linear Codes Linear Codes In the V[n,q] setting, an important class of codes are the linear codes, these codes are the ones whose code words form a subvector space of V[n,q]. If the subspace of V[n,q]
More informationLecture 10: Invertible matrices. Finding the inverse of a matrix
Lecture 10: Invertible matrices. Finding the inverse of a matrix Danny W. Crytser April 11, 2014 Today s lecture Today we will Today s lecture Today we will 1 Single out a class of especially nice matrices
More information1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each)
Math 33 AH : Solution to the Final Exam Honors Linear Algebra and Applications 1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each) (1) If A is an invertible
More informationInner product. Definition of inner product
Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product
More informationOrthogonal Matrices. u v = u v cos(θ) T (u) + T (v) = T (u + v). It s even easier to. If u and v are nonzero vectors then
Part 2. 1 Part 2. Orthogonal Matrices If u and v are nonzero vectors then u v = u v cos(θ) is 0 if and only if cos(θ) = 0, i.e., θ = 90. Hence, we say that two vectors u and v are perpendicular or orthogonal
More informationEigenvalues, Eigenvectors, Matrix Factoring, and Principal Components
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they
More informationMath 018 Review Sheet v.3
Math 018 Review Sheet v.3 Tyrone Crisp Spring 007 1.1  Slopes and Equations of Lines Slopes: Find slopes of lines using the slope formula m y y 1 x x 1. Positive slope the line slopes up to the right.
More informationInner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
More informationSubspaces of R n LECTURE 7. 1. Subspaces
LECTURE 7 Subspaces of R n Subspaces Definition 7 A subset W of R n is said to be closed under vector addition if for all u, v W, u + v is also in W If rv is in W for all vectors v W and all scalars r
More informationMath 315: Linear Algebra Solutions to Midterm Exam I
Math 35: Linear Algebra s to Midterm Exam I # Consider the following two systems of linear equations (I) ax + by = k cx + dy = l (II) ax + by = 0 cx + dy = 0 (a) Prove: If x = x, y = y and x = x 2, y =
More informationMath 54. Selected Solutions for Week Is u in the plane in R 3 spanned by the columns
Math 5. Selected Solutions for Week 2 Section. (Page 2). Let u = and A = 5 2 6. Is u in the plane in R spanned by the columns of A? (See the figure omitted].) Why or why not? First of all, the plane in
More information