LINEAR ALGEBRA AND MATRICES M3P9

Save this PDF as:

Size: px
Start display at page:

Transcription

1 LINEAR ALGEBRA AND MATRICES M3P9 December 11, 2000 PART ONE. LINEAR TRANSFORMATIONS Let k be a field. Usually, k will be the field of complex numbers C, but it can also be any other field, e.g., the field of real numbers R. Let V be an n-dimensional vector space over k, and f : V V be a linear transformation. Given a basis e 1,..., e n of V we can write the matrix of f in this basis: n A = (a ij ), where f(e j ) = a ij e i This means that j-th column of A is the column vector f(e j ). In another basis f will be given by a different matrix. How is it related to A? Lemma. Let p 1,..., p n be another basis of V, and let X be the matrix of f in this basis. Then P 1 AP = X where P = (p ij ) is a n n-matrix whose columns are the column vectors p 1,..., p n in order; in other words, p j = n p ij e i. Proof. By definition, A(p j ) = n x ij p i. The l.h.s. of this expression is n n n n n n A( p ij e i ) = p ij Ae i = p ij a ki e k = ( a ki p ij )e k. k=1 k=1 The r.h.s. is n n n n x ij p ki e k = ( p ki x ij )e k. k=1 k=1 On comparing these expressions we conculde that n n a ki p ij = p ki x ij 1

2 for any k and j. Thus we have AP = P X. The matrix P is invertible because p 1,..., p n is a basis, hence we finally get what we claimed. QED Defintion. Two square n n-matrices A and B are similar if for some invertible n n-matrix P we have P 1 AP = B. This is an equivalence relation. The previous lemma shows that the matrices in a similarity class are precisely the matrices of one and the same linear transformation written with respect to all possible bases of V. The first part of the course is devoted to the study of canonical forms of matrices: the classification of matrices up to similarity, or, what is the same, to finding a basis in which a given linear transformation has the matrix of the simplest possible form. We want this canonical form to be unique. The simplest possible matrices are diagonal matrices. One can ask a natural question: is any matrix similar to a diagonal one? Unfortunately, the answer is negative, as we shall see shortly. We shall use the following result from the first year linear algebra course: The following conditions are equivalent: (i) Ker(A) 0, (ii) det(a) = 0, (iii) A is not invertible. Definition. The characteristic polynomial of A is defined as f A (t) := det(t.id A), where Id = δ ij is the identity matrix. The roots of f A (t) are called the eigenvalues of A. Similar matrices have the same characteristic polynomial, so it gives us a means to distinguish between different similarity classes. Equivalently, λ k is an eigenvalue of A if and only if λ.id A is not invertible. Example. Find the characteristic polynomial and the eigenvalues of an upper triangular matrix. Definiton. A non-zero vector v V is called an eigenvector of A if Av = λv for some λ k. Properties. If λ is an eigenvalue of A, then there is an eigenvector with this eigenvalue. If k = C, then any matrix has an eigenvalue, and hence an eigenvector. If k = R, this is no longer true, for example the matrix ( )

3 has no real eigenvalue, and hence no real eigenvector. Theorem 1. A matrix is diagonalizable (that is, is similar to a diagonal matrix) if and only if it has a basis consisting of eigenvectors. If this is the case, then the entries of the corresponding diagonal matrix are the eigenvalues of A. Proof. We shall usually denote by e 1,..., e n the standard basis of V, e i = (0,..., 1 i,..., 0). Suppose that X = P 1 AP is diagonal. Then e 1,..., e n are eigenvectors of X. Then p 1 := P e 1,..., p n := P e n, is a basis of V consisting of eigenvectors of A. Conversely, let p 1,..., p n be a basis of A consisting of eigenvectors with eigenvalues λ 1,..., λ n, respectively. By Lemma above X = P 1 AP is the matrix of A in the basis p 1,..., p n, where p i = P e i. Then Xe i = λ i.e i. QED Examples. Let us consider the matrix ( ) 0 1 S = 0 0 If v = (v 1, v 2 ), then Sv = (v 2, 0). Thus all eigenvectors of S are proportional to e 1. Hence S is not diagonalizable. Theorem 2. Let v 1,..., v m be eigenvectors of A such that their eigenvalues λ i are pairwise different, λ i λ j if i j. Then v 1,..., v m are linearly independent. Proof. Define B i as the product of λ j.id A for all j i. (These matrices commute so the order is not important.) Then B i sends v j for j i to 0, and multiplies v i by a non-zero number (λ 1 λ i )... (λ i 1 λ i )(λ i+1 λ i )... (λ n λ i ), because by assumption the λ i are all different. Suppose that μ 1 v μ n v n = 0. Applying B i to this we conclude that μ i = 0. Hence there is no non-trivial relation among the v i. QED A corollary of this theorem is a sufficient condition of diagonalizability. Corollary. If f A (t) has n different roots, then A is diagonalizable. A diagonal matrix is clearly upper triangular. Over an algebraically closed field any matrix is similar to an upper triangular matrix, as shows the following theorem. 3

4 Theorem 3. Let A be a matrix such that f A (t) is a product of linear factors over k, then A is similar to an upper triangular matrix (with eigenvalues on the main diagonal). Proof. We proceed by induction in dim(v ). The assertion is trivially true for dim(v ) = 1. Suppose it is true for vector spaces of dimension n 1 Let λ 1 k be a root of f A (t), and v 1 V be an eigenvector with eigenvalue λ 1. Let v 2,..., v n be vectors in V such that v 1,..., v n is a basis of V. Then the matrix of A in this basis is X = λ b b 1,n 1 0 b b 2,n b n 1,1... b n 1,n 1 Then f X (t) = f A (t) = (t λ 1 )f B (t), where B = (b ij ), hence f B (t) is a product of linear factors. By the inductive assumption, C := Q 1 BQ is upper triangular for some (n 1) (n 1)-matrix Q. Let T be the direct sum of Id of size 1 1 and Q. Then λ c T c 1,n 1 XT = 0 c c 2,n c n 1,1... c n 1,n 1 This matrix is upper triangular. QED This theorem shows that when k is an algebraically closed field, say k = C, then any matrix is similar to an upper triangular matrix. This is often a big help in calculations, as shows the following proof. Cayley Hamilton theorem. f A (A) = 0. First proof. Before starting the proof we note that for any polynomial g(t) we have g(p 1 AP ) = P 1 g(a)p. Thus it is enough to check the statement on any matrix in the similarity class of A. To check that f A (A) = 0 we can go over to an extension of the ground field k over which f A (t) is a product of linear factors. Then by the previous theorem we can assume that A = λ λ λ λ n 4

5 One proves by induction that (A λ 1 Id)... (A λ m Id) kills e 1,...,e m. (Note that (A λ m Id)e m is a linear combination of e 1,..., e m 1.) Thus f A (A) maps V to 0, that is, f A (A) is a zero matrix. QED Second proof. Recall that the matrix adj(b) (whose entries are principal minors of B with appropriate signs) is a matrix of the same size as B such that det(b).id = adj(b)b. Let adj(t.id A) be the adjacent matrix of t.id A. Then we have f A (t).id = adj(t.id A).(t.Id A). Both sides here are polynomials in t with matrix coefficients. Hence we can substitute t by A. Then the right hand side vanishes, and we get f A (A) = 0. QED There exists a unique monic polynomial m A (t) with coefficients in k such that m A (A) = 0 and deg(m A (t)) is minimal among the polynomials annihilating A. Indeed, by the Cayley Hamilton theorem the set of non-trivial polynomials p(t) such that p(a) = 0 is not empty, therefore it contains a monic polynomial of minimal degree. If there are two such polynomials, the degree of their difference is less than deg(m A (t)), which is a contradiction. Hence such a polynomial is unique. Definition. The unique polynomial m A (t) with coefficients in k such that m A (A) = 0 and deg(m A (t)) is minimal among the polynomials annihilating A is called the minimal polynomial of A. It is clear that m A (t) divides f A (t). (We have f A (t) = g(t)m A (t) + r(t) for some polynomials g(t) and r(t), deg(r(t)) < deg(m A (t)); this implies that r(a) = 0 which is a contradiction.) Note that m A (t) = m P 1 AP (t). Example. If A is diagonal with paiwise different diagonal entries, then m A (t) = f A (t). Note, however, that m Id (t) = t 1, whereas f Id (t) = (t 1) n. Proposition. Every eigenvalue of A is a root of m A (t). Proof. If Av = λv, v 0, then 0 = m A (A)v = m A (λ)v, hence m A (λ) = 0. QED Exercise. The minimal polynomial of the matrix is t n + c n 1 t n c 1 t + c c c c c n 1 5

6 From now on we suppose that f A (t) is a product of linear factors, in other words, that all eigenvalues of A are k. We fix our notation. Let m f A (t) = (t λ i ) d i, where λ i λ j, if i j, m d i = n; f j (t) = m,i j (t λ i ) d i. Define V i := Ker((λ i.id A) d i ). It is clear that AV i V i, that is, the subspaces V i are A-invariant. Lemma. V i = Im(f i (A)). Proof. Cayley Hamilton theorem implies that Im(f i (A)) V i, so it remains to show the opposite inclusion. The ideal in k[t] generated by the polynomials f i (t) is the whole ring k[t], hence 1 = m h i (t)f i (t) for some polynomials h i (t). (This follows from the fact that k[t] has Euclid s algorithm.) Then we have for any v V m m v = h i (A)f i (A)v = f i (A)h i (A)v. Applying this to v V i, and taking into account that f j (t) for j i is divisible by (t λ i ) d i, we conclude that v = h i (A)f i (A)v = f i (A)h i (A)v Im(f i (A)). QED Theorem. V = m V i. Proof. The last displayed formula shows that V is generated by the sum of the V i = Im(f i (A)). Let us show that the sum is direct. Suppose that v v m = 0, where v i V i. Since f i (A) kills all the v j except possibly v i, it must also kill v i, that is, f i (A)v i = 0. Since Id = m h i (A)f i (A) we have v i = h i (A)f i (A)v i. This implies that v i = 0. QED This reduces the classification of arbitrary linear transformations to that of nilpotent ones (a matrix M is nilpotent if M d = 0 for some d). Let us explain this. Recall that AV i V i. If we choose some bases, say p (i) 1,..., p (i) k i, in each of the V i, we get a basis of V in which the matrix of A is the direct sum of blocks. Each block is the matrix of A restricted to V i (in the basis p (i) 1,..., p (i) k i ). Let B i be the restriction of A λ i.id to V i. Then B d i i = 0. 6

7 Remark. We have f Bi (t) = t dim(vi) because the only eigenvalue of any nilpotent matrix is 0. This implies that the characteristic polynomial of A restricted to V i is (t λ i ) dim(vi). Since f A (t) = m (t λ i ) d i is the product of these, we conclude that dim(v i ) = d i. It remains to construct a good basis of a nilpotent linear transformation. Let B be a linear transformation of a vector space W such that B d = 0 whereas B d 1 0. Define Then BW i W i 1. We have W i = Ker(B i ), i = 0,..., d. W = W d W d 1... W 1 W 0 = 0 Lemma. For any j > i if w W j is not in W j 1, then B j i w is in W i \ W i 1 (obvious). (In particular, taking j = d we see that all the inclusions in the dispayed formula above are strict, that is, W i W i 1.) Based on this Lemma we construct a basis of W as follows. Choose w (1) 1,..., w s (1) 1 to be linearly independent vectors in W such that W =< w (1) 1,..., w (1) s 1 > W d 1. By the previous Lemma we conclude that Bw (1) 1,..., Bw s (1) 1 are linearly independent vectors in W d 1, and the intersecion of the space generated by them with W d 2 is zero. Let us choose linearly independent vectors w (2) 1,..., w s (2) 2 in W d 1 so that W d 1 =< w (2) 1,..., w (2) s 2 > < Bw (1) 1,..., Bw (1) s 1 > W d 2. Then we continue like this, adding new vectors at each step to generate W i by W i 1 and the images of vectors already chosen. In the end we get the set of vectors {B k w (i) j }, where i = 1,..., d 1, j = 1,..., s i, k = 0,..., d i. It is clear by construction that it generates the whole space W. We assert that the vectors {B k w (i) j }, i = 1,..., d 1, j = 1,..., s i, k = 0,..., d i, are linearly independent. Let aijk B k w (i) j = 0 7

8 be a non-trivial linear combination. Apply B d 1 to it, the only possibly surviving terms are a 1j0 B d 1 w (1) j = 0, but then a 1j0 w (1) j = 0 (by the Lemma). However, these vectors are linearly independent, hence all a 1j0 = 0. Next, we apply B d 2, then the only possibly non-vanishing terms are a1j1 B d 1 w (1) j + a 2j0 B d 2 w (2) j = 0. By the Lemma this implies that a1j1 Bw (1) j + a 2j0 w (2) j = 0, and then again all coeficients must be zero since these vectors are linearly independent by constrution. Continuing in the same spirit we prove that all our vectors are linearly independent, and so form a basis of W. Although the construction of this basis was somewhat involved, the matrix of B now takes a very simple form. Let us rearrange the basis vectors as follows: B d 1 e (1) 1, B d 2 e (1) 1,..., e (1) 1, B d 1 e (1) 2, B d 2 e (1) 2,..., e (1) 2,..., B d 1 e (1) s 1, B d 2 e (1) s 1,..., e (1) s 1, B d 2 e (2) 1, B d 3 e (2) 1,..., e (2) 1, B d 2 e (2) 2, B d 3 e (2) 2,..., e (2) 2,..., B d 2 e (2) s 2, B d 3 e (2) s 2,..., e (2) s 2,..., Be (d 1) 1, e (d 1) 1, Be (d 1) 2, e (d 1) 2,..., Be (d 1) s d 1, e (d 1) s d 1 Then the matrix of B is the direct sum of blocks of the form , e (d) 1,..., e (d) s d. Such a matrix is called a nilpotent Jordan block. There are s 1 such blocks of size d, s 2 such blocks of size d 1, and so on, finally, there are s d 1 blocks of the form ( 0 1 ) 0 0 and s d zero blocks of size 1. Note that the form of this matrix implies that rk(b i ) = (d i)s 1 + (d i 1)s s d i. this implies that the numbers s i are uniquely determined by the similarity class of B. We have proved that a nilpotent matrix B is similar to a direct sum of nilpotent Jordan blocks uniquely determined up to a permutation. 8

9 A matrix of the form λ λ λ λ λ is called a Jordan block. Summarizing our preceding discussion we obtain the following statement. Theorem. (The Jordan normal form.) Any matrix is similar to a direct sum of Jordan blocks. The Jordan normal form in each similarity class is unique up to a permutation of Jordan blocks. Exercises - corollaries of this theorem. (a) Let A be a Jordan normal form, then m A (t) = m (t λ i ) δ i, where δ i is the size of a maximal Jordan block with eigenvalue λ i. (Use the fact that the minimal polynomial of a direct sum of matrices is the l.c.m. of their minimal polynomials.) (b) A matrix is diagonalizable iff its minimal polynomial is a product of distinct linear factors. (This is read directly from the Jordan normal form.) A corollary of this fact: if the direct sum of two matrices is diagonalizable, then these matrices are also diagonalizable. Definition. Let A be a linear transformation on a vector space V. The set of vectors v V such that Av = λv is called the eigenspace with eigenvalue λ, and is denoted by V λ. Reformulating an earlier result we have the following criterion of diagonalizability: A is diagonalizable iff V is the direct sum of eigenspaces of A. (=iff V has a basis consisting of eigenvectors of A) Another corollary of the theorem. The number of Jordan blocks with eigenvalue λ equals dim(v λ ). (Each block contributes one eigenvector.) Proposition. Two commuting diagonalizable matrices are simultaneously diagonalizable. (In other words, there exists an invertible matrix P such that P 1 AP and P 1 BP are both diagonal.) Proof. If V λ is an eigenspace of A, then AB = BA implies that BV λ V λ. We have seen that the restriction of B to V λ is diagonalizable (its minimal polynomial is a divisor of m B (t) which is a product of diffirent linear factors, hence it is also a product of diffirent linear factors). Since the restriction of A 9

10 to V λ is scalar, we can find a basis of V λ in which both A and B are diagonal. Putting such bases of V λ together we get a desired basis of V. QED Actually, the same is true for any number of pairwise commuting diagonalizable matrices (induction on the number of matrices). The following result has many applications in the theory of Lie groups and Lie algebras. Theorem. (Jordan decomposition) Any (square) matrix A can be represented as a sum A = D A + N A, where D A is diagonalizable and N A is nilpotent, and D A = g 1 (A), N A = g 2 (A), for some polynomials g 1 (t) and g 2 (t). This implies that D A N A = N A D A, and, more generally, D A and N A commute with any matrix B such that AB = BA. Hint. In our previous notation take D A = m λ i h i (A)f i (A), and use the fact that h i (A)f i (A) is the projector from V to V i (that is, it is identity on V i and zero on V j if j i). 10

11 PART TWO. BILINEAR FORMS Definition. Let V be a vector space over a field k. A function <, > from V V to k is called a bilinear form if it is linear in each argument. Definition. If p 1,..., p n is a basis of V, then the Gram matrix of <, > in this basis is the matrix A = (a ij ) such that a ij =< p i, p j >. Then if v and u are column vectors v = v i p i = (v 1,..., v n ) t, u = u i p i = (u 1,..., u n ) t we have < v, u >= v t Au. Lemma. If f 1,..., f n is another basis of V, and F = (f ij ) is a matrix such that f j = n f ij p i, then the Gram matrix of <, > in the new basis f 1,..., f n is F t AF, where F t is the transpose of F. The proof is a direct check. Matrices A and B (square, of the same size) such that B = F t AF for some invertible matrix F are called congruent. Note that then we have rk(b) = rk(a). If rk(a) = n, the bilinear form <, > is called non-singular (or non-degenerate). Theorem. <, > is non-singular iff for any non-zero v V there exists u V such that < v, u > 0. Proof. The set of v V such that < v, u >= 0 for any u V is just the kernel of A t. QED Definition. If U V is a subspace, then U := {v V such that < v, u >= 0 for all u U} Lemma. If U is a subspace of V such that the restriction of <, > to U is non-singular, then V = U U. Proof. We need to prove that U U = 0 (an immediate consequence of the non-singularity assumption), and that V is generated by U and U. Take any v V, then by non-singularity we can find w U such that < u, v >=< u, w > for all u U. Such a w can be defined by the formula Bw = (< p 1, v >,..., < p m, v >) t, 11

12 where p 1,... p m is a basis of U, and B is the Gram matrix of <, > restricted to U. Then v w U, and v = w + (v w) is the desired decomposition. QED Symmetric bilinear forms Definition. A bilinear form is called symmetric if < v, u >=< u, v > for all v, u V. Bilinear forms correspond to symmetric Gram matrices (such that A t = A). Lemma. Assume char(k) 2. A symmetric form is non-zero iff there exists v V such that < v, v > 0. Proof. The form is non-zero iff there exist v, u V such that < v, u > 0. Then the formula < u, v >= 1 (< v + u, v + u > < v, v > < u, u >) 2 implies that at least one of < v + u, v + u >, < v, v >, < u, u > is non-zero. QED The previous formula implies that provided the characteristic of the ground field is different from 2, a symmetric bilinear form < u, v > can be recovered from its restriction to the diagonal, that is the function V k given by < v, v >. This function is called the associated quadratic form, in coordinates it is simply the expression n i,j=1 a ij x i x j. The well known process of reducing a quadratic polynomial to a sum of squares is what lies behind the idea of proof of the following theorem. Theorem. Diagonalization of a symmetric bilinear form. Assume char(k) 2. For any symmetric bilinear form there exist a basis of V in which its Gram matrix is diagonal. Proof. If our form is zero there is nothing to prove. Suppose it is not, then by the previous lemma we can find u V such that the restriction of <, > to the subspace generated by u is non-singular. Call this subspace U. Then V = U U, and we repeat this argument for <, > restricted to U, and so on. QED Corollary 1. Assume char(k) 2. Every symmetric matrix is congruent to a diagonal one. Although in general the coefficients of this diagonal matrix are not uniquely defined, over some fields one can be more precise. 12

13 Corollary 2. If k = C, then any symmetric matrix is congruent to the diagonal matrix diag(1,..., 1, 0,..., 0). In particular, symmetric matrices A and B are congruent over C iff rk(a) = rk(b). Hence there are n + 1 congruency classes. Corollary 3. If k = R, then any symmetric matrix is congruent to the diagonal matrix diag(1,..., 1, 1,..., 1, 0,..., 0). Theorem. Sylvester s law of inertia. The number of 1 s and 1 s in a diagonal form of a real symmetric matrix are well defined (i.e., these numbers only depend on the congruency class of our matrix). Proof. Let p be the number of 1 s and let q be the number of 1 s. The fact that p + q is well defined is clear - this is just the rank of A. We need a definition. A symmetric bilinear form <, > is positive definite if < u, u > is positive for any non-zero u. (Likewise <, > is negative definite if < u, u > is negative for any non-zero u. If the form is neither positive definite nor negative definite it is called indeterminate.) Consider the following number which clearly does not depend on the choice of a basis: max U V {dim(u)}, where U is a subspace of V such that <, > restricted to U is positive definite. Let us compute this number. Let W V be generated by the last n p basis vectors of the basis in which <, > has the form of Corollary 3. Then if dim(u) + dim(w ) > n, then U W contains a non-zero vector, say u. But then < u, u > 0 by the construction of W. Hence our number cannot be greater than n (n p) = p. On the other hand, the restriction of <, > to the subspace generated by the first p vectors of the above basis is clearly positive definite. Hence this number is p, which therefore is well defined. QED Definition. The number p q is called the signature of the symmetric bilinear form <, >. Alternating bilinear forms Definition. A bilinear form <, > is alternating if < u, u >= 0 for any u V. A bilinear form <, > is skew-symmetric if < u, v >= < v, u > for any u, v V. An alternating form is skew-symmetric, but the converse is only true when char(k) 2 (take u = v and divide by 2), and is wrong when char(k) = 2 (symmetric form over such a field is the same thing as skew-symmetric!). 13

14 A standard example of an alternating form is < (v 1, v 2 ), (u 1, u 2 ) >= v 1 u 2 v 2 u 1. Its Gram matrix is ( ) 0 1 J = 1 0 Theorem. Let k be a field of characteristic different from 2. Then for any alternating form there exists a basis in which its Gram matrix is the direct sum of copies of J and a zero matrix. In other words, any skew-symmetric matrix (that is, A t = A) with zeros on the main diagonal is congruent to a matrix of the above form. Note that this theorem is valid over any field. Proof. If <, > is not zero, take a non-zero vector v such that there exists u with the property that < v, u > 0. Then v and u/ < v, u > are linearly independent. Let U be the space spanned by them. The Gram matrix of <, > restricted to U is J. Since this restriction is non-singular we have V = U U. We repeat this process for U until we get a zero form. QED Corollary. The rank of an alternating form is even. In particular, nonsingular alternating forms exist only on even-dimensional vector spaces. The determinant of a skew-symmetric matrix is a square in k. In particular, if k = R, it is always non-negative. Exercise: The Pfaffian of a skew-symmetric matrix. The last corollary leaves open the question: what is the square root of det(a), where A = (a ij ) is a skew-symmetric matrix of size n = 2m? If m = 1, then obviously det(a) = a 2 12, so here the answer is clear. Show by direct computation that for m = 2 we have det(a) = (a 12 a 34 a 13 a 24 + a 14 a 23 ) 2 This formula has an analogue for any m. In fact, there exists a polynomial in matrix entries of A, called the Pfaffian P f(a), such that det(a) = P f(a) 2 (over any field). Inner product spaces. Geometry of quadric surfaces Let V be a vector space over R. An inner product is a positively definite symmetric bilinear form <, > on V. An example of such a form is the standard dot product u t.v. One defines orthonormal bases and proves the Cauchy Schwarz inequality. In terms of the dot product one writes < u, v >= u t.a.v, where A is the Gram matrix of <, >. One has u t.(a.v) = (A t.u) t.v. 14

15 Definition. A square matrix P is an orthogonal matrix if P t P = Id. Such matrices preserve the dot product, and hence preserve the distances. Recall that the distance between v and v is defined as v v := (v v ) t.(v v ). Proposition. All eigenvalues of a real symmetric matrix (considered as complex numbers) are real. Proof. Let z = (z 1,..., z n ) be a complex eigenvector, then z t.a.z = λ n z i 2. Since A is symmetric and real, this also equals (A.z) t.z = λ n z i 2. Hence λ R. QED Theorem. Reduction to principal axes. For any real symmetric matrix there exists an orthonormal basis of V consisting of its eigenvectors. Such a basis is not necessarily unique. Proof. We proceed by induction on n = dim(v ). When n = 1 there is nothing to prove. Suppose the statement is true for n 1. We can find an eigenvector u of length 1 (that is, u t.u = 1) with some eigenvalue λ, and call U its linear span. Let U be its orthogonal complement with respect to the dot product. Since u t.a.v = (A.u) t.v = λ.u t.v, we see that if v is orthogonal to u with respect to the dot product, then Av is also. Hence AU U. Then V = U U (since the restriction of the dot product to U is nonsingular), and U and U are A-invariant. Now take any othonormal basis in U, which exists by the inductive assumption. Together with u this gives an orthonormal basis of V. QED The lines passing through the vectors of the basis constructed in this theorem are called principal axes. Corollary. For any real symmetric matrix there exists an orthogonal matrix P such that P 1 AP = P t AP is diagonal (with eigenvalues on the diagonal). Definition. A map f : R n R n which preserves distances, that is, such that f(v) f(v ) = v v for all v, v R, is called an isometry. Proposition. Any isometry is a composition of an orthogonal transformation v P v and a translation v v + w. Proof. Let f : R n R n be an isometry. By composing f with the translation by f(0) we arrange that 0 goes to 0. So we can assume that f(0) = 0. Now taking v = 0 in the definition of isometry, we see that f(u) = u for any u. Now note that u t.v = (1/2)( u 2 + v 2 u v 2 ) 15

16 and all the terms in the right hand side are preserved by f. Therefore, we have (f(u) t.f(v)) = u t.v for any u and v. (This is, in fact, the standard observation that the angles of a triangle are uniquely determined by its sides.) In other words, f preserves the dot product. In particular, the image of an orthonormal basis is again an orthonormal basis, say p i = f(e i ). Now any vector u is written as n u = u i e i, u i = (u t.e i ). If we write f(u) = n x i p i we must have x i = f(u) t.p i = f(u) t.f(e i ) = u t.e i = u i. Therefore, we have f : n u i e i n u i p i, hence f is linear, say f(u) = F u. Then (F u) t.f v = u t.v implies that F is an orthogonal matrix. QED Proposition. Any equation of degree two n n a ij x i x j + b i x i + c = 0 i,j=1 can be reduced by an isometry to one of the following forms, where all the coefficients are non-zero: I : m λ i x 2 i 1 = 0, III : II : m 1 m λ i x 2 i = 0, λ i x 2 i + x m = 0, where 1 m n, and in the last case m 1 1. Proof. By the principal axes theorem we can assume that the terms of order 2 have diagonal form. Use translations to complete the squares. If there are no linear terms left, after possibly multiplying by an overall constant, we find ourselves in case I or II. If there are some linear terms, change the coordinates by an orthogonal transformation in x m+1,..., x n, which is identity on x 1,..., x m, so that there is only one linear term. (This uses the fact that any vector of unit length can be completed to an orthonormal basis.) After a translation (to eliminate the constant term) and a multiplication by a constant we arrive at case III. QED 16

17 If m < n, we say that the graph of the corresponding equation is a cylinder, such a graph can be obtained from a graph in a space of dimension n 1. In case II the graph is a cone. Recall that if n = m = 2, then the quadric curve (a conic) is an ellipse if both coefficients are positive, a hyperbola if just one is positive, and an imaginary ellipse if both are negative. In case III we get a parabola. Let us now consider the case n = m = 3. In case I if all the coefficients are positive, the quadric is called an ellipsoid, if just one coefficient is negative we get a hyperboloid of one sheet, if exactly two coefficients are negative, we have a hyperboloid of two sheets, and, finally, if all the coefficients are negative, we have an imaginary ellipsoid. In case III, if the coefficients have the same sign, we have an elliptic paraboloid, and in the opposite case an hyperbolic paraboloid. It is clear that cylinders and cones are made of lines (there is a line contained in our surface passing through any point of it). One can prove that an ellipsoid, an elliptic paraboloid and a hypeboloid of 2 sheets do not contain lines. (In fact, they do not contain a subset of a straight line if it has at least 3 points!) On the contrary, there are lines on a hyperboloid of 1 sheet and on a hyperbolic paraboloid. Actually, we have the following stronger statement. Theorem. Through every point of a hyperbolic paraboloid pass two different lines entirely contained in it. (And the same is true for the hyperboloid of 1 sheet.) Proof. We have an equation ax 2 by 2 = z, where a, b > 0. Let (x 0, y 0, z 0 ) be any solution of this equation. Consider the plane given by the equation 2ax 0 x 2by 0 y (ax 2 0 by 2 0) = z (this is a tangent plane). Substituting this expression for z into the first equation we obtain a(x x 0 ) 2 = b(y y 0 ) 2, which is equivalent to the union of the line a(x x 0 ) = b(y y 0 ) and the line a(x x 0 ) = b(y y 0 ) (in the plane given by the above equation). QED Thus we have a really curved surface entirely made of absolutely straight lines! This has important applications in architecture: one can build a curved surface by using only straight metal rails. Hyperboloids of 1 sheet can be used to build towers... Note that the lines we have found are never parallel to the plane given by y = 0. Thus such a line intersects this plane, and the resulting point is (r, 0, ar 2 ) for some r R. The two lines on our paraboloid passing through this point (which we just have found) are L ± r given by the equations y = ± a/b(x r), z = ar(2x r). We have found two one-parameter families of 17

18 lines entirely contained in our surface such that through every point of the surface passes exactly one line of each family. Lemma. For any r and s the lines L + r and L s meet in exactly one point. The lines L + r and L + s never meet unless r = s. (easy) One can rephrase this by saying that points of L + r (for any r) parametrize lines of the family {L s }, and vice versa. This gives a bijective parametrization of the hyperbolic paraboloid by polynomials in two variables r and s: x = 1 2 (r + s), y = 1 a (s r), z = ars. 2 b With this parametrization the point (r, s) is mapped to the intersection point of L + r and L s. Perhaps a more conceptual way of constructing the lines of a hyperbolic paraboloid is this: the equation ax 2 by 2 = z is satisfied by the points of the following two lines ax by = rz, ax + by = 1/r, ax + by = sz, ax by = 1/s, for real parameters r and s. This can also be used on a hyperboloid of 1 sheet. Indeed, it is given by the equation ax 2 cz 2 = 1 by 2, a, b, c > 0, and this equation holds for any point of the following two lines: ax cz = r(1 by), ax + cz = 1/r(1 + by), ax + cz = s(1 by), ax cz = 1/s(1 + by). These families can be used to prove the analogue of the previous theorem for hyperboloids of 1 sheet. Note also that parabolic mirrors (elliptic paraboloids of round shape) are used from cars lights to radiotelescopes, whenever one needs to focus a parallel beam of light (or vice versa). 18

Similarity and Diagonalization. Similar Matrices

MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?

Orthogonal Diagonalization of Symmetric Matrices

MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

MATH 240 Fall, Chapter 1: Linear Equations and Matrices

MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS

University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients

Tangent and normal lines to conics

4.B. Tangent and normal lines to conics Apollonius work on conics includes a study of tangent and normal lines to these curves. The purpose of this document is to relate his approaches to the modern viewpoints

Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function

17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function, : V V R, which is symmetric, that is u, v = v, u. bilinear, that is linear (in both factors):

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

Exercise 1 1. Let A be an n n orthogonal matrix. Then prove that (a) the rows of A form an orthonormal basis of R n. (b) the columns of A form an orthonormal basis of R n. (c) for any two vectors x,y R

1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each)

Math 33 AH : Solution to the Final Exam Honors Linear Algebra and Applications 1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each) (1) If A is an invertible

The determinant of a skew-symmetric matrix is a square. This can be seen in small cases by direct calculation: 0 a. 12 a. a 13 a 24 a 14 a 23 a 14

4 Symplectic groups In this and the next two sections, we begin the study of the groups preserving reflexive sesquilinear forms or quadratic forms. We begin with the symplectic groups, associated with

Summary of week 8 (Lectures 22, 23 and 24)

WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:

Lecture 1: Schur s Unitary Triangularization Theorem

Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections

1 Eigenvalues and Eigenvectors

Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x

WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE?

WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE? JOEL H. SHAPIRO Abstract. These notes supplement the discussion of linear fractional mappings presented in a beginning graduate course

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal

Definition 12 An alternating bilinear form on a vector space V is a map B : V V F such that

4 Exterior algebra 4.1 Lines and 2-vectors The time has come now to develop some new linear algebra in order to handle the space of lines in a projective space P (V ). In the projective plane we have seen

ISOMETRIES OF R n KEITH CONRAD

ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x

(January 14, 2009) End k (V ) End k (V/W )

(January 14, 29) [16.1] Let p be the smallest prime dividing the order of a finite group G. Show that a subgroup H of G of index p is necessarily normal. Let G act on cosets gh of H by left multiplication.

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar

Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions

Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

Chapter 6. Orthogonality

6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

Applied Linear Algebra I Review page 1

Applied Linear Algebra Review 1 I. Determinants A. Definition of a determinant 1. Using sum a. Permutations i. Sign of a permutation ii. Cycle 2. Uniqueness of the determinant function in terms of properties

INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL

SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics

Practice Math 110 Final. Instructions: Work all of problems 1 through 5, and work any 5 of problems 10 through 16.

Practice Math 110 Final Instructions: Work all of problems 1 through 5, and work any 5 of problems 10 through 16. 1. Let A = 3 1 1 3 3 2. 6 6 5 a. Use Gauss elimination to reduce A to an upper triangular

Problems for Advanced Linear Algebra Fall 2012

Problems for Advanced Linear Algebra Fall 2012 Class will be structured around students presenting complete solutions to the problems in this handout. Please only agree to come to the board when you are

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

NOTES on LINEAR ALGEBRA 1

School of Economics, Management and Statistics University of Bologna Academic Year 205/6 NOTES on LINEAR ALGEBRA for the students of Stats and Maths This is a modified version of the notes by Prof Laura

Solution based on matrix technique Rewrite. ) = 8x 2 1 4x 1x 2 + 5x x1 2x 2 2x 1 + 5x 2

8.2 Quadratic Forms Example 1 Consider the function q(x 1, x 2 ) = 8x 2 1 4x 1x 2 + 5x 2 2 Determine whether q(0, 0) is the global minimum. Solution based on matrix technique Rewrite q( x1 x 2 = x1 ) =

LINEAR ALGEBRA. September 23, 2010

LINEAR ALGEBRA September 3, 00 Contents 0. LU-decomposition.................................... 0. Inverses and Transposes................................. 0.3 Column Spaces and NullSpaces.............................

Linear Algebra I. Ronald van Luijk, 2012

Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.

MATH 551 - APPLIED MATRIX THEORY

MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

Chapter 20. Vector Spaces and Bases

Chapter 20. Vector Spaces and Bases In this course, we have proceeded step-by-step through low-dimensional Linear Algebra. We have looked at lines, planes, hyperplanes, and have seen that there is no limit

Inner Product Spaces

Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

by the matrix A results in a vector which is a reflection of the given

Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

Notes on Determinant

ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

Inner products on R n, and more

Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

Math 210 Quadratic Polynomials Jerry L. Kazdan Polynomials in One Variable. After studying linear functions y = ax + b, the next step is to study quadratic polynomials, y = ax 2 + bx + c, whose graphs

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number

Notes on Symmetric Matrices

CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.

10. Graph Matrices Incidence Matrix

10 Graph Matrices Since a graph is completely determined by specifying either its adjacency structure or its incidence structure, these specifications provide far more efficient ways of representing a

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

Linear Least Squares

Linear Least Squares Suppose we are given a set of data points {(x i,f i )}, i = 1,...,n. These could be measurements from an experiment or obtained simply by evaluating a function at some points. One

We call this set an n-dimensional parallelogram (with one vertex 0). We also refer to the vectors x 1,..., x n as the edges of P.

Volumes of parallelograms 1 Chapter 8 Volumes of parallelograms In the present short chapter we are going to discuss the elementary geometrical objects which we call parallelograms. These are going to

Linear Algebra Review. Vectors

Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two

October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix

Linear Algebra & Properties of the Covariance Matrix October 3rd, 2012 Estimation of r and C Let rn 1, rn, t..., rn T be the historical return rates on the n th asset. rn 1 rṇ 2 r n =. r T n n = 1, 2,...,

Vector and Matrix Norms

Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

LINEAR ALGEBRA W W L CHEN

LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,

α = u v. In other words, Orthogonal Projection

Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v

Additional Topics in Linear Algebra Supplementary Material for Math 540. Joseph H. Silverman

Additional Topics in Linear Algebra Supplementary Material for Math 540 Joseph H Silverman E-mail address: jhs@mathbrownedu Mathematics Department, Box 1917 Brown University, Providence, RI 02912 USA Contents

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

MAT067 University of California, Davis Winter 2007 Linear Maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007) As we have discussed in the lecture on What is Linear Algebra? one of

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = 36 + 41i.

Math 5A HW4 Solutions September 5, 202 University of California, Los Angeles Problem 4..3b Calculate the determinant, 5 2i 6 + 4i 3 + i 7i Solution: The textbook s instructions give us, (5 2i)7i (6 + 4i)(

9.3 Advanced Topics in Linear Algebra

548 93 Advanced Topics in Linear Algebra Diagonalization and Jordan s Theorem A system of differential equations x = Ax can be transformed to an uncoupled system y = diag(λ,, λ n y by a change of variables

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

On the general equation of the second degree

On the general equation of the second degree S Kesavan The Institute of Mathematical Sciences, CIT Campus, Taramani, Chennai - 600 113 e-mail:kesh@imscresin Abstract We give a unified treatment of the

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0

APPLICATIONS. are symmetric, but. are not.

CHAPTER III APPLICATIONS Real Symmetric Matrices The most common matrices we meet in applications are symmetric, that is, they are square matrices which are equal to their transposes In symbols, A t =

Matrix Representations of Linear Transformations and Changes of Coordinates

Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under

Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj

Section 5. l j v j = [ u u j u m ] l jj = l jj u j + + l mj u m. l mj Section 5. 5.. Not orthogonal, the column vectors fail to be perpendicular to each other. 5..2 his matrix is orthogonal. Check that

A matrix over a field F is a rectangular array of elements from F. The symbol

Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted

Section 1.1. Introduction to R n

The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to

PYTHAGOREAN TRIPLES KEITH CONRAD 1. Introduction A Pythagorean triple is a triple of positive integers (a, b, c) where a + b = c. Examples include (3, 4, 5), (5, 1, 13), and (8, 15, 17). Below is an ancient

d(f(g(p )), f(g(q))) = d(g(p ), g(q)) = d(p, Q).

10.2. (continued) As we did in Example 5, we may compose any two isometries f and g to obtain new mappings f g: E E, (f g)(p ) = f(g(p )), g f: E E, (g f)(p ) = g(f(p )). These are both isometries, the

PROJECTIVE GEOMETRY. b3 course 2003. Nigel Hitchin

PROJECTIVE GEOMETRY b3 course 2003 Nigel Hitchin hitchin@maths.ox.ac.uk 1 1 Introduction This is a course on projective geometry. Probably your idea of geometry in the past has been based on triangles

2.5 Gaussian Elimination

page 150 150 CHAPTER 2 Matrices and Systems of Linear Equations 37 10 the linear algebra package of Maple, the three elementary 20 23 1 row operations are 12 1 swaprow(a,i,j): permute rows i and j 3 3

Section 2.1. Section 2.2. Exercise 6: We have to compute the product AB in two ways, where , B =. 2 1 3 5 A =

Section 2.1 Exercise 6: We have to compute the product AB in two ways, where 4 2 A = 3 0 1 3, B =. 2 1 3 5 Solution 1. Let b 1 = (1, 2) and b 2 = (3, 1) be the columns of B. Then Ab 1 = (0, 3, 13) and

Eigenvalues and eigenvectors of a matrix

Eigenvalues and eigenvectors of a matrix Definition: If A is an n n matrix and there exists a real number λ and a non-zero column vector V such that AV = λv then λ is called an eigenvalue of A and V is

DETERMINANTS. b 2. x 2

DETERMINANTS 1 Systems of two equations in two unknowns A system of two equations in two unknowns has the form a 11 x 1 + a 12 x 2 = b 1 a 21 x 1 + a 22 x 2 = b 2 This can be written more concisely in

MATH36001 Background Material 2015

MATH3600 Background Material 205 Matrix Algebra Matrices and Vectors An ordered array of mn elements a ij (i =,, m; j =,, n) written in the form a a 2 a n A = a 2 a 22 a 2n a m a m2 a mn is said to be

UNIT 2 MATRICES - I 2.0 INTRODUCTION. Structure

UNIT 2 MATRICES - I Matrices - I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress

3. INNER PRODUCT SPACES

. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.

4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows

MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006

MATH 511 ADVANCED LINEAR ALGEBRA SPRING 26 Sherod Eubanks HOMEWORK 1 1.1 : 3, 5 1.2 : 4 1.3 : 4, 6, 12, 13, 16 1.4 : 1, 5, 8 Section 1.1: The Eigenvalue-Eigenvector Equation Problem 3 Let A M n (R). If

Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible:

Cramer s Rule and the Adjugate Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible: Theorem [Cramer s Rule] If A is an invertible

9 Multiplication of Vectors: The Scalar or Dot Product

Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation

6. ISOMETRIES isometry central isometry translation Theorem 1: Proof:

6. ISOMETRIES 6.1. Isometries Fundamental to the theory of symmetry are the concepts of distance and angle. So we work within R n, considered as an inner-product space. This is the usual n- dimensional

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with

T ( a i x i ) = a i T (x i ).

Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)

8 Square matrices continued: Determinants

8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You

4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

They are congruent since the rst can be moved to coincide with the second. However, the next pair of triangles are not congruent.

Chapter 9 Symmetry In this chapter we explore a connection between algebra and geometry. One of the main topics of plane geometry is that of congruence; roughly, two geometric gures are said to be congruent

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

THE DIMENSION OF A VECTOR SPACE

THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field

Notes on Linear Algebra. Peter J. Cameron

Notes on Linear Algebra Peter J. Cameron ii Preface Linear algebra has two aspects. Abstractly, it is the study of vector spaces over fields, and their linear maps and bilinear forms. Concretely, it is

Solutions to Math 51 First Exam January 29, 2015

Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not