V = [v 1, v 2,... v n ].

Size: px
Start display at page:

Download "V = [v 1, v 2,... v n ]."

Transcription

1 COORDINATES, EIGENVALUES, AND EIGENVECTORS Review of coordinates The standard basis The standard basis for R n is the basis E = {e, e, e n }, where e j is the vector with in the j-th position and zeros elsewhere Coordinates with respect to a basis Let B = {v, v, v n } be a basis for R n We know that for each vector x R n there are unique scalar coordinates c, c, c n such that () x = c v + c v + + c n v n We call the vector c R n having coordinates c, c c n the coordinate vector of x with respect to the basis B, and denote it by x B 3 First examples (a) x E = x for each x R n (b) For any basis B = {v, v, v n } of R n : v j B = e j ( j n) (c) 0 B = 0 for any basis B (the zero vector has the same representation with respect to any basis 4 Matrix calculation of coordinates Letting B = {v, v, v n } be a basis for R n we can rewrite equation () in matrix form as () x = V x B where V is the matrix having the vector v j for its j-th column, ie V = v, v, v n Date: March 6, 005

2 EIGENVALUES/EIGENVECTORS V is invertible because its columns are linearly independent (see, for example, Summary 39(v) of our textbook, pp 9 0), so we can compute the B coordinate representation of any vector x R n by solving () for x B : (3) x B = V x For example, if in R we take B = {v, v } to be our basis, where v = and 3 v =, then 5 5 V = and V 3 5 = 3 Suppose, for instance, that x = Then 5 x B = V x = 3 More generally, if x = x 7 = 4 is any vector in R x n then the same calculation shows that 5x + x x B = 3x x 5 The matrix of a linear transformation with respect to a basis Recall that each n n matrix induces a linear transformation T : R n R n by means of matrix multiplication: (4) T (x) = Ax (x R n ) We refer to A as the standard matrix for T The lesson of what s to follow is that it s not always the best matrix for T! For a basis B = {v, v, v n } for R n we define (5) T B = T (v ) B, T (v ) B T (v n ) B, ie T B is the matrix that has the B-coordinate representation of T (v j ) as its j-th column ( j n) We call T B the matrix of T with respect to the basis B 6 Example: B = E, the standard basis In this case T B = Ae, Ae, Ae n, the standard matrix for T

3 EIGENVALUES/EIGENVECTORS 3 7 Geometric examples in R In this next group of examples, L is a line through the origin in the direction of a non-zero vector v R, and w is a non-zero vector perpendicular to v B = {v, w} We take as our basis for R, not the standard one, but instead (a) Suppose T : R R is the transformation of orthogonal projection onto L Then T (v) = v and T (w) = 0, so T B = T (v) B, T (w) B = v B, 0 B = e, 0 = where we use the fact that v B = e and 0 B = 0 (see 3 (b) and (c)) Compare this simple and informative matrix for T with the one we got when we first introduced orthogonal projection Then we took v to be a unit vector with coordinates u and u, and did this computation: u T (x) = (x v)v = (x u + x u ) u = u x + u u x u u u x + u = u u x u u u, x, x so the standard matrix for the linear transformation of orthogonal projection onto L is the completely uninformative one u A = u u u u u (b) Suppose T is reflection in L Then T (v) = v as before, but now T (w) = w Thus T B = T (v) B, T (w) B = v B, w B = e, e = 0 0 Note that in this computation we used the fact that w B = w B ( linearity of coordinates Fact 34, page 39 of our textbook) and that v B = e and w B = e ( 3 (b)) (c) Suppose T is a shear in the direction of L, ie that T (v) = v and T (w) = kv + w for some fixed real number k Then using linearity of coordinates and 3(b) we see that k T B = 0 which is the same as the standard matrix for a horizontal shear

4 4 EIGENVALUES/EIGENVECTORS 8 Using T B to compute T (x) The utility of the matrix of a linear transformation T with respect to a basis B lies in the following formula: (6) T B x B = T (x) B (x R n ), which says that, if you re willing to give up the standard basis fro R n, and do all your calculations relative to the basis B, then you can use the matrix of T with respect to B in the same way you used to use the standard matrix Proof of formula (6): For x R n we have the unique representation x = c v + c v + + c n v n where c, c,, c n are the (standard) coordinates of x B By linearity of T : T (x) = c T (v ) + c T (v ) + + c n T (v n ), so by linearity of coordinates: T (x) B = c T (v ) B + c T (v ) B + + c n T (v n ) B, But the expression on the right is just the matrix product T B x B, so we are done Eigenvalues and Eigenvectors Definition Suppose A is an n n matrix and T the associated linear transformation of R n (so T (x) = Ax for every x R n ) We call a real number λ an eigenvalue of T (or of A) whenever there is a non-zero vector x R n such that T (x) = λx The vector x is called an eigenvector of T (or of A) corresponding to the eigenvalue λ Examples (a) Look back at the examples of 7, where L is a line through the origin in the direction of a non-zero vector v, and w is a non-zero vector perpendicular to v If T : R R is orthogonal projection onto L, then v and w are eigenvectors of T corresponding respectively to eigenvalues and 0 If T is reflection in L then v and w are still eigenvectors, but now the corresponding eigenvalues are and If T is a shear along L then v is an eigenvector with eigenvalue, but w is not an eigenvector

5 EIGENVALUES/EIGENVECTORS 5 (b) (see Problem 5, page 46) Suppose A = and v =, w = 4 3 Then one computes that Av = 5v and Aw = w, so v is an eigenvector for A corresponding to the eigenvalue 5 and w is an eigenvector corresponding to the eigenvalue Note that, with respect to the basis B = {v, w} of R, the matrix 5 0 for the linear transformation induced by A is just 0 (c) (see Problem 9 of 34) Here T is the linear transformation of R 3 induced by the matrix A = We are given a basis B = {v, v, v 3 } for R 3, where v =, v =, and v 3 = You can check easily that each of these basis vectors is an eigenvector: Av = 0 (eigenvalue zero), Av = v (eigenvalue ) and Av 3 = v 3 (eigenvalue ), and that therefore the matrix of the associated linear transformation T with respect to the basis B is T B = In each of the examples of the previous two sections, whenever we had a linear transformation T of R n that bequeathed to R n a basis B of eigenvectors, the matrix of T with respect to B turned out to be diagonal, ie all entries zero off the main diagonal 3 Notation The n n diagonal matrix with entry λ j in the j-th diagonal position will be denoted diag {λ, λ, λ n } Thus, for example, the 3 3 identity matrix is diag {,, }, while in (c), T B = diag {0,, } Diagonal matrices are the easiest ones for which to find eigenvalues and eigenvectors: 4 Proposition If A = diag {λ, λ, λ n } then the standard unit vector e j is an eigenvector for the eigenvalue λ j ( j n)

6 6 EIGENVALUES/EIGENVECTORS Proof Our hypothesis is that A = λ e, λ e,, λ n e n, ie that A is the matrix with λ j e j as its j-th column ( j n) But Ae j is the j-th column of A, so Ae j = λ j e j for each j between and n 5 Exercise Prove the converse: If an n n matrix A has the standard basis for R n as eigenvectors, then A is diagonal 6 The Diagonalization Theorem For a linear transformation T : R n R n, and a basis B for R n, the following two statements are equivalent (each implies the other): (a) B consists of eigenvectors of T (b) T B is a diagonal matrix Proof (a) (b): Suppose B = {v, v, v n } consists of eigenvectors of T, so that there are real numbers λ j ( j n) such that T (v j ) = λ j v j (j =,,, n) We must show that T B is diagonal Starting with the definition of T B, we compute: T B = T v B, T v B,, T v n B = λ v B, λ v B,, λ n v n B = λ v B, λ v B,, λ n v n B = λ e, λ e,, λ n e n = diag {λ, λ,, λ n } where, in this computation we used successively: the definition of T B (line ), the fact that T (v j ) = λ j v j (line ), linearity of coordinates (line 3), and finally the fact that v j B = e j (line 4) (b) (a): Now suppose B = {v, v, v n } is a basis for R n, and that T B = diag {λ, λ,, λ n } We must show that each v j is an eigenvector of T In fact, we ll see that T v j = λ j v j for j n Fix such a j, and note that the j-th column of T B is just λ j e j, ie, () T B e j = λ j e j

7 EIGENVALUES/EIGENVECTORS 7 Now the right-hand side of () is λ j v j B which, by linearity of coordinates, is λ j v j B Similarly, the left-hand side of () is T B v j B which, by (6), is T v j B Thus T v j B = λ j v j B, ie, the both T v j and λ j v j have the same B-coordinate vectors Thus T v j = λ j v j, as desired 7 Definition Call an n n matrix diagonalizable if R n has a basis consisting of eigenvectors of this matrix Warning: Not all square matrices are diagonalizable! For example it s geometrically clear that shear transfomations of R have (up to multiplication by a non-zero constant) just a single eigenvector, while rotations of R through any angle not an integer multiple of π have no eigenvectors at all! We ll see how to prove this kind of result analytically (ie without reference to pictures) in the next section 3 Efficient eigenvalue computation for matrices We ll base our computations for matrices on the notion of determinant a b 3 Definition The determinant of the matrix A = is: c d det(a) = ad bc The key to all the work of this section is: 3 Theorem For A a matrix, the system of linear equations Ax = 0 has nontrivial solution if and only if det(a) = 0 Proof In scalar form the system of equations is: ax + bx = 0 cx + dx = 0 Suppose first that neither a nor c is zero Then we can replace the second equation by c times the first minus a times the second, yielding the equivalent system: ax + bx = 0 (ad bc)x = 0

8 8 EIGENVALUES/EIGENVECTORS which has a nontrivial solution if and only if ad bc = 0 Suppose a = 0 Then det(a) = bc, which is zero precisely when either b or c is zero If b = 0 then we are back to one equation in two unknowns, so there are nontrivial (infinitely many) solutions If b 0 then any solution must have x = 0, whereupon the second equation implies cx = 0 If c = 0 (ie det(a) = 0) then there are infinitely many solutions: x = 0 and x arbitrary If, however c 0 (ie, det(a) 0) then x is also required to be zero, hence the only solution is the trivial one 33 Corollary The eigenvalues of a matrix A are the real numbers λ that satisfy its characteristic equation det(a λi) = 0 Proof This follows immediately from the above Theorem and the definition of eigenvalue More precisely, to say a real number λ is an eigenvalue of A means that there is a non-zero vector x R such that: Ax = λx ie Ax λx = 0 ie (A λi)x = 0 Thus λ is an eigenvector of A if and only if the homogeneous system (A λi)x = 0 has a nontrivial solution, which, by Theorem 3, happens if and only if det(a λi) = 0 34 Example Show that the matrix is diagonalizable A = 4 3 Proof Of course we know this already from (b) where the eigenvectors and eigenvalues of A were handed to us But suppose they weren t Then we write the characteristic equation: λ 0 = det(a λi) = det( 4 3 λ = ( λ)(3 λ) 8 = λ 4λ 5 = (λ 5)(λ + ) )

9 EIGENVALUES/EIGENVECTORS 9 so that λ = 5 and λ = are the solutions, and so these are the eigenvalues of A Now for λ = 5, 4 A λi = A 5I = 4 and the system (A 5I)x = 0 is easily seen to have solutions x = any scalar multiple of In particular, v = is the unique (up to multiplication by a non-zero scalar) eigenvector for the eigenvalue 5 I leave it to you to show by the same method that v = eigenvector for the eigenvalue is the unique It s easy to check that B = {v, v } is a basis for R, so A is diagonalizable, and the matrix of its linear transformation, with respect to the basis B is diag {5, } 0 35 Example: a non-diagonalizable matrix Let A = The characteristic 0 0 equation for this matrix is λ = 0, which has only λ = 0 as its solution The corresponding eigenvectors are the nontrivial solutions to the equation Ax = 0 (ie the non-zero vectors in the kernel of A), and these all the non-zero scalar multiples of the standard unit vector e Thus there aren t enough eigenvectors here to make a basis of R, so A is not diagonalizable 4 Matrix of T B vs standard matrix of T Suppose T is a linear transformation of R n whows standard matrix is A, ie T (x) = Ax (x R n ) Suppose B = {v, v, v n } is a basis for R n It s time to get serious about the relationship between A and T B Recall that in 8 we derived equation (3): x B = V x where V = v, v,, v n is the matrix having the vector v j as its j-th column Recall also that the invertibility of V derives from the fact that its columns are linearly independent

10 0 EIGENVALUES/EIGENVECTORS Thus we may compute: T B = T (v ) B, T (v ) B,, T (v n ) B = Av B, Av B,, Av n B = V Av, V Av,, V Av n = V A v, v,, v n = V AV where the first line is the definition of T B, the second uses the definition of T in terms of A, the third uses (3), and the fourth is the definition of matrix multiplication We have proved: 4 Theorem (Fact 344, Page 44 of our text) Suppose B = {v, v, v n } is a basis for R n and T a linear transformation of R n with standard matrix A Then T B = V AV, where V is the matrix whose j-th column is v j ( j n) 5 Application: Powers of matrices Theorem 4 is of great utility in computing large powers of matrices a calculation that would be difficult to do by hand in many circumstances The method depends on two crucial facts 5 Theorem If A, B and V are n n matrices, and V is invertible, then for each positive integer p: (V BV ) p = V B p V Proof We use the method of Mathematical Induction, which asserts that to prove the theorem we need only check that it s true for p =, and then show that: (*) Result true for an integer p Result true for p + The point is that once you ve established the case p = and proved (*), then (*) gives you for free the case p =, and then the case p = 3, and so on forever

11 EIGENVALUES/EIGENVECTORS For our result, the case p = is obviously true, so suppose we have proved the theorem for some p We must show that it s true for p + ; to this end write (V BV ) p+ = (V BV ) p (V BV ) = (V B p V )(V BV ) = (V B p )(V V )(BV ) = (V B p )I(BV ) = V B p+ V where the second line follows from our induction hypothesis that the theorem is true for p, the third line follows from the associate law of matrix multiplication, and the others are clear The next result asserts that for a diagonal matrix you can compute a power by taking the corresponding power of each element of the matrix 5 Theorem For any positive integer p, diag {λ, λ,, λ n } p = diag {λ p, λ p,, λ p n} Proof I leave this one mostly as an exercise First check that diag {λ, λ,, λ n }diag {µ, µ,, µ n } = diag {λ µ, λ µ,, λ n µ n } The result about powers then follows quickly from this by induction 5 53 Example Compute 4 3 Solution Recall that we ve already found that the vectors v = and v = are eigenvectors for A = (for eigenvalues 5 and - respectively) and that B = 4 3 {v, v } is a basis for R Thus, according to Theorem 4 above, 5 0 = T 0 B = V AV, where V = v, v = Thus A = V V,

12 EIGENVALUES/EIGENVECTORS so successive application of Theorem 5 and Theorem 5 yield, (5) A 5 = V V = V 0 ( ) 5 V A little calculation shows that V = /3 /3 /3 /3 more calculation (and the fact that ( ) 5 = ) yield A 5 = , after which (5) and a little 54 Example: A dynamical system Neighboring countries Elbonia and Pilikia are undergoing a relatively peaceful period in their histories Each is reducing its annual defense budget, but because of nervousness about the others intentions, each adds back in to its budget a percentage of its neighbors expenditures More precisely, Pilikia reduces its budget by 5%, but adds back in 5% of Elbonia s budget, while Elbonia reduces its budget by a whopping 75%, but adds back in 75% of Pilikia s defense budget Let P (t) denote Pilikia s defense budget for year t, and E(t) Elbonia s (in 005 Euros times one billion) Suppose initially P (0) = 3 and E(0) = Determine the behavior of P (t) and E(t) as t P (t) Solution The vector x(t) = describes the state of the system at year t We have E(t) (with amounts calculated in billions of Euros) ie, P (t + ) = 75P (t) + 5E(t) E(t + ) = 5P (t) + 75E(t) (5) x(t + ) = Ax(t) where A = Thus defense expenditures for both countries will be: x() = Ax(0) in year one, x() = Ax() = A x(0) in year two, and more generally in year n they will be x(n) = Ax(n ) = A x(n ) = A n x(0)

13 So it all comes down to computing A n EIGENVALUES/EIGENVECTORS 3 For this I leave it to you to check that the eigenvalues of A are and /, and the corresponding eigenvectors are v = for λ = and v = for λ = / Now B = {v, v } is a basis for R relative to which the linear transformation T whose standard matrix is A has a diagonal matrix: 0 T B = = V 0 / AV, where V = v, v = Thus for each positive integer n, n (/) n = 0 / hence solving for A n : Now one calculates easily that A n = V V = n = V A n V, n 0 0 (/) n V / / / / from which follows A n = + n n n + n Thus x(n) = A n x(0) = ( + n ) n 3 n + n = 5 + n 5 n 5/ Thus x(n) as n, ie both nations tend toward a yearly defense budget 5/ of 5 billion Euros I 7 # 4, 8, 9,, 36, 38 6 Exercises II 7 #8; this will be part of a future writing assignment III 73 #40, 4 These will be part of a future writing assignment 54 of these notes works out a special case of #40

1 Eigenvalues and Eigenvectors

1 Eigenvalues and Eigenvectors Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x

More information

Similarity and Diagonalization. Similar Matrices

Similarity and Diagonalization. Similar Matrices MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

More information

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal

More information

Presentation 3: Eigenvalues and Eigenvectors of a Matrix

Presentation 3: Eigenvalues and Eigenvectors of a Matrix Colleen Kirksey, Beth Van Schoyck, Dennis Bowers MATH 280: Problem Solving November 18, 2011 Presentation 3: Eigenvalues and Eigenvectors of a Matrix Order of Presentation: 1. Definitions of Eigenvalues

More information

Summary of week 8 (Lectures 22, 23 and 24)

Summary of week 8 (Lectures 22, 23 and 24) WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry

More information

EIGENVALUES AND EIGENVECTORS

EIGENVALUES AND EIGENVECTORS Chapter 6 EIGENVALUES AND EIGENVECTORS 61 Motivation We motivate the chapter on eigenvalues b discussing the equation ax + hx + b = c, where not all of a, h, b are zero The expression ax + hx + b is called

More information

Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions

Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential

More information

Chapter 6. Orthogonality

Chapter 6. Orthogonality 6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

More information

Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible:

Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible: Cramer s Rule and the Adjugate Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible: Theorem [Cramer s Rule] If A is an invertible

More information

C 1 x(t) = e ta C = e C n. 2! A2 + t3

C 1 x(t) = e ta C = e C n. 2! A2 + t3 Matrix Exponential Fundamental Matrix Solution Objective: Solve dt A x with an n n constant coefficient matrix A x (t) Here the unknown is the vector function x(t) x n (t) General Solution Formula in Matrix

More information

MATH 240 Fall, Chapter 1: Linear Equations and Matrices

MATH 240 Fall, Chapter 1: Linear Equations and Matrices MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS

More information

A matrix over a field F is a rectangular array of elements from F. The symbol

A matrix over a field F is a rectangular array of elements from F. The symbol Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted

More information

8 Square matrices continued: Determinants

8 Square matrices continued: Determinants 8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You

More information

7 - Linear Transformations

7 - Linear Transformations 7 - Linear Transformations Mathematics has as its objects of study sets with various structures. These sets include sets of numbers (such as the integers, rationals, reals, and complexes) whose structure

More information

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively. Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry

More information

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = 36 + 41i.

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = 36 + 41i. Math 5A HW4 Solutions September 5, 202 University of California, Los Angeles Problem 4..3b Calculate the determinant, 5 2i 6 + 4i 3 + i 7i Solution: The textbook s instructions give us, (5 2i)7i (6 + 4i)(

More information

Problems for Advanced Linear Algebra Fall 2012

Problems for Advanced Linear Algebra Fall 2012 Problems for Advanced Linear Algebra Fall 2012 Class will be structured around students presenting complete solutions to the problems in this handout. Please only agree to come to the board when you are

More information

1 Spherical Kinematics

1 Spherical Kinematics ME 115(a): Notes on Rotations 1 Spherical Kinematics Motions of a 3-dimensional rigid body where one point of the body remains fixed are termed spherical motions. A spherical displacement is a rigid body

More information

Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

More information

MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006

MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006 MATH 511 ADVANCED LINEAR ALGEBRA SPRING 26 Sherod Eubanks HOMEWORK 1 1.1 : 3, 5 1.2 : 4 1.3 : 4, 6, 12, 13, 16 1.4 : 1, 5, 8 Section 1.1: The Eigenvalue-Eigenvector Equation Problem 3 Let A M n (R). If

More information

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

More information

Solutions to Math 51 First Exam January 29, 2015

Solutions to Math 51 First Exam January 29, 2015 Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not

More information

Notes on Determinant

Notes on Determinant ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

More information

by the matrix A results in a vector which is a reflection of the given

by the matrix A results in a vector which is a reflection of the given Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

More information

LS.6 Solution Matrices

LS.6 Solution Matrices LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions

More information

1 Sets and Set Notation.

1 Sets and Set Notation. LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most

More information

Similar matrices and Jordan form

Similar matrices and Jordan form Similar matrices and Jordan form We ve nearly covered the entire heart of linear algebra once we ve finished singular value decompositions we ll have seen all the most central topics. A T A is positive

More information

1 2 3 1 1 2 x = + x 2 + x 4 1 0 1

1 2 3 1 1 2 x = + x 2 + x 4 1 0 1 (d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which

More information

UNIT 2 MATRICES - I 2.0 INTRODUCTION. Structure

UNIT 2 MATRICES - I 2.0 INTRODUCTION. Structure UNIT 2 MATRICES - I Matrices - I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress

More information

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013 Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

More information

University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7. Review University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE?

WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE? WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE? JOEL H. SHAPIRO Abstract. These notes supplement the discussion of linear fractional mappings presented in a beginning graduate course

More information

Solution based on matrix technique Rewrite. ) = 8x 2 1 4x 1x 2 + 5x x1 2x 2 2x 1 + 5x 2

Solution based on matrix technique Rewrite. ) = 8x 2 1 4x 1x 2 + 5x x1 2x 2 2x 1 + 5x 2 8.2 Quadratic Forms Example 1 Consider the function q(x 1, x 2 ) = 8x 2 1 4x 1x 2 + 5x 2 2 Determine whether q(0, 0) is the global minimum. Solution based on matrix technique Rewrite q( x1 x 2 = x1 ) =

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

is in plane V. However, it may be more convenient to introduce a plane coordinate system in V. .4 COORDINATES EXAMPLE Let V be the plane in R with equation x +2x 2 +x 0, a two-dimensional subspace of R. We can describe a vector in this plane by its spatial (D)coordinates; for example, vector x 5

More information

LINEAR ALGEBRA W W L CHEN

LINEAR ALGEBRA W W L CHEN LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,

More information

Inverses and powers: Rules of Matrix Arithmetic

Inverses and powers: Rules of Matrix Arithmetic Contents 1 Inverses and powers: Rules of Matrix Arithmetic 1.1 What about division of matrices? 1.2 Properties of the Inverse of a Matrix 1.2.1 Theorem (Uniqueness of Inverse) 1.2.2 Inverse Test 1.2.3

More information

(a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular.

(a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular. Theorem.7.: (Properties of Triangular Matrices) (a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular. (b) The product

More information

1 Orthogonal projections and the approximation

1 Orthogonal projections and the approximation Math 1512 Fall 2010 Notes on least squares approximation Given n data points (x 1, y 1 ),..., (x n, y n ), we would like to find the line L, with an equation of the form y = mx + b, which is the best fit

More information

Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc.

Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc. 2 Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Theorem 8: Let A be a square matrix. Then the following statements are equivalent. That is, for a given A, the statements are either all true

More information

Orthogonal Diagonalization of Symmetric Matrices

Orthogonal Diagonalization of Symmetric Matrices MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

More information

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

More information

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION 4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:

More information

Determinants. Dr. Doreen De Leon Math 152, Fall 2015

Determinants. Dr. Doreen De Leon Math 152, Fall 2015 Determinants Dr. Doreen De Leon Math 52, Fall 205 Determinant of a Matrix Elementary Matrices We will first discuss matrices that can be used to produce an elementary row operation on a given matrix A.

More information

Iterative Methods for Computing Eigenvalues and Eigenvectors

Iterative Methods for Computing Eigenvalues and Eigenvectors The Waterloo Mathematics Review 9 Iterative Methods for Computing Eigenvalues and Eigenvectors Maysum Panju University of Waterloo mhpanju@math.uwaterloo.ca Abstract: We examine some numerical iterative

More information

4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows

More information

2.1: Determinants by Cofactor Expansion. Math 214 Chapter 2 Notes and Homework. Evaluate a Determinant by Expanding by Cofactors

2.1: Determinants by Cofactor Expansion. Math 214 Chapter 2 Notes and Homework. Evaluate a Determinant by Expanding by Cofactors 2.1: Determinants by Cofactor Expansion Math 214 Chapter 2 Notes and Homework Determinants The minor M ij of the entry a ij is the determinant of the submatrix obtained from deleting the i th row and the

More information

Sergei Silvestrov, Christopher Engström, Karl Lundengård, Johan Richter, Jonas Österberg. November 13, 2014

Sergei Silvestrov, Christopher Engström, Karl Lundengård, Johan Richter, Jonas Österberg. November 13, 2014 Sergei Silvestrov,, Karl Lundengård, Johan Richter, Jonas Österberg November 13, 2014 Analysis Todays lecture: Course overview. Repetition of matrices elementary operations. Repetition of solvability of

More information

INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL

INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics

More information

PROVING STATEMENTS IN LINEAR ALGEBRA

PROVING STATEMENTS IN LINEAR ALGEBRA Mathematics V2010y Linear Algebra Spring 2007 PROVING STATEMENTS IN LINEAR ALGEBRA Linear algebra is different from calculus: you cannot understand it properly without some simple proofs. Knowing statements

More information

MATH10212 Linear Algebra B Homework 7

MATH10212 Linear Algebra B Homework 7 MATH22 Linear Algebra B Homework 7 Students are strongly advised to acquire a copy of the Textbook: D C Lay, Linear Algebra and its Applications Pearson, 26 (or other editions) Normally, homework assignments

More information

NON SINGULAR MATRICES. DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that

NON SINGULAR MATRICES. DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that NON SINGULAR MATRICES DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that AB = I n = BA. Any matrix B with the above property is called

More information

Additional Topics in Linear Algebra Supplementary Material for Math 540. Joseph H. Silverman

Additional Topics in Linear Algebra Supplementary Material for Math 540. Joseph H. Silverman Additional Topics in Linear Algebra Supplementary Material for Math 540 Joseph H Silverman E-mail address: jhs@mathbrownedu Mathematics Department, Box 1917 Brown University, Providence, RI 02912 USA Contents

More information

CHARACTERISTIC ROOTS AND VECTORS

CHARACTERISTIC ROOTS AND VECTORS CHARACTERISTIC ROOTS AND VECTORS 1 DEFINITION OF CHARACTERISTIC ROOTS AND VECTORS 11 Statement of the characteristic root problem Find values of a scalar λ for which there exist vectors x 0 satisfying

More information

(January 14, 2009) End k (V ) End k (V/W )

(January 14, 2009) End k (V ) End k (V/W ) (January 14, 29) [16.1] Let p be the smallest prime dividing the order of a finite group G. Show that a subgroup H of G of index p is necessarily normal. Let G act on cosets gh of H by left multiplication.

More information

4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

More information

Facts About Eigenvalues

Facts About Eigenvalues Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v

More information

Examination paper for TMA4115 Matematikk 3

Examination paper for TMA4115 Matematikk 3 Department of Mathematical Sciences Examination paper for TMA45 Matematikk 3 Academic contact during examination: Antoine Julien a, Alexander Schmeding b, Gereon Quick c Phone: a 73 59 77 82, b 40 53 99

More information

Dot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product

Dot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product Dot product and vector projections (Sect. 12.3) Two definitions for the dot product. Geometric definition of dot product. Orthogonal vectors. Dot product and orthogonal projections. Properties of the dot

More information

Chapter 5. Matrices. 5.1 Inverses, Part 1

Chapter 5. Matrices. 5.1 Inverses, Part 1 Chapter 5 Matrices The classification result at the end of the previous chapter says that any finite-dimensional vector space looks like a space of column vectors. In the next couple of chapters we re

More information

Solutions to Linear Algebra Practice Problems

Solutions to Linear Algebra Practice Problems Solutions to Linear Algebra Practice Problems. Find all solutions to the following systems of linear equations. (a) x x + x 5 x x x + x + x 5 (b) x + x + x x + x + x x + x + 8x Answer: (a) We create the

More information

On the general equation of the second degree

On the general equation of the second degree On the general equation of the second degree S Kesavan The Institute of Mathematical Sciences, CIT Campus, Taramani, Chennai - 600 113 e-mail:kesh@imscresin Abstract We give a unified treatment of the

More information

Determinants. Chapter Properties of the Determinant

Determinants. Chapter Properties of the Determinant Chapter 4 Determinants Chapter 3 entailed a discussion of linear transformations and how to identify them with matrices. When we study a particular linear transformation we would like its matrix representation

More information

MATH 551 - APPLIED MATRIX THEORY

MATH 551 - APPLIED MATRIX THEORY MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

More information

α = u v. In other words, Orthogonal Projection

α = u v. In other words, Orthogonal Projection Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v

More information

Iterative Techniques in Matrix Algebra. Jacobi & Gauss-Seidel Iterative Techniques II

Iterative Techniques in Matrix Algebra. Jacobi & Gauss-Seidel Iterative Techniques II Iterative Techniques in Matrix Algebra Jacobi & Gauss-Seidel Iterative Techniques II Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin

More information

The Characteristic Polynomial

The Characteristic Polynomial Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem

More information

160 CHAPTER 4. VECTOR SPACES

160 CHAPTER 4. VECTOR SPACES 160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results

More information

Systems of Linear Equations

Systems of Linear Equations Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

More information

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

More information

Linear Dependence Tests

Linear Dependence Tests Linear Dependence Tests The book omits a few key tests for checking the linear dependence of vectors. These short notes discuss these tests, as well as the reasoning behind them. Our first test checks

More information

1 Quiz on Linear Equations with Answers. This is Theorem II.3.1. There are two statements of this theorem in the text.

1 Quiz on Linear Equations with Answers. This is Theorem II.3.1. There are two statements of this theorem in the text. 1 Quiz on Linear Equations with Answers (a) State the defining equations for linear transformations. (i) L(u + v) = L(u) + L(v), vectors u and v. (ii) L(Av) = AL(v), vectors v and numbers A. or combine

More information

Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,

More information

Name: Section Registered In:

Name: Section Registered In: Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are

More information

Markov Chains, Stochastic Processes, and Advanced Matrix Decomposition

Markov Chains, Stochastic Processes, and Advanced Matrix Decomposition Markov Chains, Stochastic Processes, and Advanced Matrix Decomposition Jack Gilbert Copyright (c) 2014 Jack Gilbert. Permission is granted to copy, distribute and/or modify this document under the terms

More information

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in

More information

4.6 Null Space, Column Space, Row Space

4.6 Null Space, Column Space, Row Space NULL SPACE, COLUMN SPACE, ROW SPACE Null Space, Column Space, Row Space In applications of linear algebra, subspaces of R n typically arise in one of two situations: ) as the set of solutions of a linear

More information

d(f(g(p )), f(g(q))) = d(g(p ), g(q)) = d(p, Q).

d(f(g(p )), f(g(q))) = d(g(p ), g(q)) = d(p, Q). 10.2. (continued) As we did in Example 5, we may compose any two isometries f and g to obtain new mappings f g: E E, (f g)(p ) = f(g(p )), g f: E E, (g f)(p ) = g(f(p )). These are both isometries, the

More information

B such that AB = I and BA = I. (We say B is an inverse of A.) Definition A square matrix A is invertible (or nonsingular) if matrix

B such that AB = I and BA = I. (We say B is an inverse of A.) Definition A square matrix A is invertible (or nonsingular) if matrix Matrix inverses Recall... Definition A square matrix A is invertible (or nonsingular) if matrix B such that AB = and BA =. (We say B is an inverse of A.) Remark Not all square matrices are invertible.

More information

Inner Product Spaces

Inner Product Spaces Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

More information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1. MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

More information

ISOMETRIES OF R n KEITH CONRAD

ISOMETRIES OF R n KEITH CONRAD ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x

More information

2.1: MATRIX OPERATIONS

2.1: MATRIX OPERATIONS .: MATRIX OPERATIONS What are diagonal entries and the main diagonal of a matrix? What is a diagonal matrix? When are matrices equal? Scalar Multiplication 45 Matrix Addition Theorem (pg 0) Let A, B, and

More information

Topic 1: Matrices and Systems of Linear Equations.

Topic 1: Matrices and Systems of Linear Equations. Topic 1: Matrices and Systems of Linear Equations Let us start with a review of some linear algebra concepts we have already learned, such as matrices, determinants, etc Also, we shall review the method

More information

Factorization Theorems

Factorization Theorems Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization

More information

Lecture 1: Schur s Unitary Triangularization Theorem

Lecture 1: Schur s Unitary Triangularization Theorem Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections

More information

Matrix Inverse and Determinants

Matrix Inverse and Determinants DM554 Linear and Integer Programming Lecture 5 and Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1 2 3 4 and Cramer s rule 2 Outline 1 2 3 4 and

More information

Linear Codes. In the V[n,q] setting, the terms word and vector are interchangeable.

Linear Codes. In the V[n,q] setting, the terms word and vector are interchangeable. Linear Codes Linear Codes In the V[n,q] setting, an important class of codes are the linear codes, these codes are the ones whose code words form a sub-vector space of V[n,q]. If the subspace of V[n,q]

More information

Lecture 10: Invertible matrices. Finding the inverse of a matrix

Lecture 10: Invertible matrices. Finding the inverse of a matrix Lecture 10: Invertible matrices. Finding the inverse of a matrix Danny W. Crytser April 11, 2014 Today s lecture Today we will Today s lecture Today we will 1 Single out a class of especially nice matrices

More information

1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each)

1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each) Math 33 AH : Solution to the Final Exam Honors Linear Algebra and Applications 1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each) (1) If A is an invertible

More information

Inner product. Definition of inner product

Inner product. Definition of inner product Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product

More information

Orthogonal Matrices. u v = u v cos(θ) T (u) + T (v) = T (u + v). It s even easier to. If u and v are nonzero vectors then

Orthogonal Matrices. u v = u v cos(θ) T (u) + T (v) = T (u + v). It s even easier to. If u and v are nonzero vectors then Part 2. 1 Part 2. Orthogonal Matrices If u and v are nonzero vectors then u v = u v cos(θ) is 0 if and only if cos(θ) = 0, i.e., θ = 90. Hence, we say that two vectors u and v are perpendicular or orthogonal

More information

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they

More information

Math 018 Review Sheet v.3

Math 018 Review Sheet v.3 Math 018 Review Sheet v.3 Tyrone Crisp Spring 007 1.1 - Slopes and Equations of Lines Slopes: Find slopes of lines using the slope formula m y y 1 x x 1. Positive slope the line slopes up to the right.

More information

Inner products on R n, and more

Inner products on R n, and more Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +

More information

Subspaces of R n LECTURE 7. 1. Subspaces

Subspaces of R n LECTURE 7. 1. Subspaces LECTURE 7 Subspaces of R n Subspaces Definition 7 A subset W of R n is said to be closed under vector addition if for all u, v W, u + v is also in W If rv is in W for all vectors v W and all scalars r

More information

Math 315: Linear Algebra Solutions to Midterm Exam I

Math 315: Linear Algebra Solutions to Midterm Exam I Math 35: Linear Algebra s to Midterm Exam I # Consider the following two systems of linear equations (I) ax + by = k cx + dy = l (II) ax + by = 0 cx + dy = 0 (a) Prove: If x = x, y = y and x = x 2, y =

More information

Math 54. Selected Solutions for Week Is u in the plane in R 3 spanned by the columns

Math 54. Selected Solutions for Week Is u in the plane in R 3 spanned by the columns Math 5. Selected Solutions for Week 2 Section. (Page 2). Let u = and A = 5 2 6. Is u in the plane in R spanned by the columns of A? (See the figure omitted].) Why or why not? First of all, the plane in

More information