T ( a i x i ) = a i T (x i ).


 Winfred Carroll
 2 years ago
 Views:
Transcription
1 Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b) T (cx) = ct (x). Fact 1. (p. 65) 1. If T is linear, then T (0) = 0. Note that the first 0 is in V and the second one is in W. 2. T is linear if and only if T (cx + y) = ct (x) + T (y) for all x, y V and c F. 3. If T is linear, then T (x y) = T (x) T (y) for all x, y V. 4. T is linear if and only if, for x 1, x 2,..., x n V and a 1, a 2,..., a n F, we have T ( a i x i = a i T (x i ). Proof.. 1. We know 0 x = 0 for all vectors x. Hence, 0 = 0 T (x) = T (0 x) = T (0) 2. Assume T is linear and c F and x, y V, then T (cx) = ct (x) and T (cx + y) = T (cx) + T (y) = ct (x) + T (y). Assume T (x + y) = T (x) + T (y) for all x, y V. Let a F and x, y V. Then T (ax) = T (ax + 0) = at (x) + T (0) = at (x) and T (x + y) = T (1 x + y) = 1 T (x) + T (y) = T (x) + T (y). 3. To start with, the definition of subtraction is: x y= x + ( y). We also wish to show: x V, 1 x = x. Given 0 x = 0 which we proved earlier, we have (1 + ( 1))x = 0 1 x + ( 1) x = 0 Since the additive inverse in unique, we conclude that ( 1) x = x. AssumeT is linear. Then T (x y) = T (x + 1 y) = T (x) + ( 1)T (y) = T (x) + ( T (y)) = T (x) T (y) for all x, y V. 4. ( ) Let n = 2, a 1 = c, and a 2 = 1. We have that T (cx 1 + x 2 ) = ct (x 1 ) + T (x 2 ). ( ) This is done by induction. Using the definition of linear transformation, we get the base case (n = 2). That is, T (a 1 x 1 + a 2 x 2 ) = T (a 1 x 1 ) + T (a 2 x 2 ) = a 1 T (x 1 ) + a 2 T (x 2 ). Now assume that n 2 and for all x 1, x 2,..., x n V and a 1, a 2,..., a n F, we have T ( a i x i ) = a i T (x i ). 1
2 2 Then for all x 1, x 2,..., x n, x n+1 V and a 1, a 2,..., a n a n+1 F, we have n+1 T ( a i x i ) = T ( a i x i + a n+1 x n+1 ) = T ( = = a i x i ) + T (a n+1 x n+1 ) a i T (x i ) + a n+1 T (x n+1 ) n+1 a i T (x i ). Defn 2. (p. 67) Let V and W be vector spaces, and let T : V W be linear. We define the null space (or kernel) N(T ) of T to be the set of all vectors x in V such that T (x) = 0; N(T ) = {x V : T (x) = 0}. We define the range (or image) R(T ) of T to be the subset W consisting of all images (under T ) of vectors in V ; that is, R(T ) = {T (x) : x V }. Theorem 2.1. Let V and W be vector spaces and T : V W be linear. Then N(T ) and R(T ) are subspaces of V and W, respectively. Proof. To prove that N(T ) is a subspace of V, we let x, y N(T ) and c F. We will show that 0 N(T ) and cx + y N(T ). We know that T (0) = 0. So, 0 N(T ). We have that T (x) = 0 and T (y) = 0 so T (cx + y) = c T (x) + T (y) = c = c 0 = 0. (The last equality is a fact that can be proved from the axioms of vector space. To prove that R(T ) is a subspace of W, assume T (x), T (y) R(T ), for some x, y V and c F. Then, ct (x)+t (y) = T (cx+y), where cx+y V, since T is a linear transformation and so, ct (x) + T (y) R(T ). Also, since 0 V and 0 = T (0), we know that 0 R(T ). Let A be a set of vectors in V. Then T (A) = {T (x) : x A}. Theorem 2.2. Let V and W be vector spaces, and let T : V W be linear. If β = {v 1, v 2,..., v n } is a basis for V, then R(T ) = span(t (β)). Proof. T (β) = {T (v 1 ), T (v 2 ),..., T (v n )}. We will show R(T ) = span(t (β)). We showed that R(T ) is a subspace of W. We know that T (β) R(T ). So, by Theorem 1.5, span(t (β)) R(T ). Let x R(T ). Then there is some x V such that T (x ) = x. β is a basis of V, so there exist scalars a 1, a 2,..., a n such that x = a 1 x 1 + a 2 x a n x n. which is in span(t (β)). Thus R(T ) span(t (β)). x = T (x ) = T (a 1 x 1 + a 2 x a n x n ) = a 1 T (x 1 ) + a 2 T (x 2 ) + + a n T (x n )
3 Defn 3. (p. 69) Let V and W be vector spaces, and let T : V W be linear. If N(T ) and R(T ) are finitedimensional, then we define the nullity of T, denoted nullity(t ), and the rank of T, denoted rank(t ), to be the dimensions of N(T ) and R(T ), respectively. Theorem 2.3. (Dimension Theorem). Let V and W be vector spaces, and let T : V W be linear. If V is finitedimensional, then nullity(t )+ rank(t ) = dim(v ). Proof. Suppose nullity(t ) = k and β = {n 1,..., n k } is a basis of N(T ). Since β is linearly independent in V, then β extends to a basis of V, β = {n 1,..., n k, x k+1,..., x n }, where, by Corollary 2c. to Theorem 1.10, i, x i N(T ). Notice that possibly β =. It suffices to show: β = {T (x k+1 ),..., T (x n )} is a basis for R(T ). By Theorem 2.2, T (β ) spans R(T ). But T (β ) = {0} β. So β spans R(T ). Suppose β is not linearly independent. Then there exist c 1, c 2,..., c n F, not all 0 such that 0 = c 1 T (x k+1 ) + + c n T (x n ). 3 Then and But then 0 = T (c 1 x k c n x n ) c 1 x k c n x n N(T ). c 1 x k c n x n = a 1 n 1 + a 2 n a k n k for some scalars a 1, a 2,..., a k, since β = {n 1,..., n k } is a basis of N(T ). Then 0 = a 1 n 1 + a 2 n a k n k c 1 x k+1 c n x n. The c is are not all zero, so this contradicts that β is a linearly independent set. Theorem 2.4. Let V and W be vector spaces, and let T : V W be linear. Then T is onetoone if and only if N(T ) = {0}. Proof. Suppose T is onetoone. Let x N(T ). Then T (x) = 0 also T (0) = 0. T being onetoone implies that x = 0. Therefore, N(T ) {0}. It is clear that {0} N(T ). Therefore, N(T ) = {0}. Suppose N(T ) = {0}. Suppose for x, y V, T (x) = T (y). We have that T (x) T (y) = 0. T (x y) = 0. Then it must be that x y = 0. This implies that y is the additive inverse of x and x is the additive inverse of y. But also y is the additive inverse of y. By uniqueness of additive inverses we have that x = y. Thus, T is onetoone. Theorem 2.5. Let V and W be vector spaces of equal (finite) dimension, and let T : V W be linear. Then the following are equivalent. (a) T is onetoone. (b) T is onto. (c) rank(t ) = dim(v ).
4 4 Proof. Assume dim(v ) = dim(w ) = n. (a) (b) : To prove T is onto, we show that R(T ) = W. Since T is onetoone, we know N(T ) = {0} and so nullity(t ) = 0. By Theorem 2.3, nullity(t ) + rank(t ) = dim(v ). So we have dim(w ) = n = rank(t ). By Theorem 1.11, R(T ) = W. (b) (c) : T is onto implies that R(T ) = W. So rank(t ) = dim(w ) = dim(v ). (c) (a) : If we assume rank(t ) = dim(v ), By Theorem 2.3 again, we have that nullity(t ) = 0. But then we know N(T ) = {0}. By Theorem 2.4, T is onetoone. Theorem 2.6. Let V and W be vector spaces over F, and suppose that β = {v 1, v 2,..., v n } is a basis for V. For w 1, w 2,..., w n in W, there exists exactly one linear transformation T : V W such that T (v i ) = w i for i = 1, 2,..., n. Proof. Define a function T : V W by T (v i ) = w i for i = 1, 2,..., n and since β is a basis, we can express every vector in V as a linear combination of vectors in β. For x V, x = a 1 v 1 + a 2 v a n v n, we define T (x) = a 1 w 1 + a 2 w a 2 w n. We will show that T is linear. Let c F and x, y V. Assume x = a 1 v 1 +a 2 v 2 + +a n v n and y = b 1 v 1 + b 2 v b n v n. Then cx + y = (ca i + b i )v i (1) T (cx + y) = T ( (ca i + b i )v i ) = (ca i + b i )w i = c a i w i + b i w i = ct (x) + T (y) Thus, T is linear. To show it is unique, we need to show that if F : V W is a linear transformation such that F (v i ) = w i for all i [n], then for all x V, F (x) = T (x). Assume x = a 1 v 1 + a 2 v a n v n. Then T (x) = a 1 w 1 + a 2 w a n w n = a 1 F (v 1 ) + a 2 F (v 2 ) + + a n F (v n ) = F (x) Cor 1. Let V and W be vector spaces, and suppose that V has a finite basis {v 1, v 2,..., v n }. If U, T : V W are linear and U(v i ) = T (v i ) for i = 1, 2,..., n, then U = T. Defn 4. (p. 80) Given a finite dimensional vector space V, an ordered basis is a basis where the vectors are listed in a particular order, indicated by subscripts. For F n, we have {e 1, e 2,..., e n } where e i has a 1 in position i and zeros elsewhere. This is known as the standard ordered basis and e i is the i th characteristic vector. Let β = {u 1, u 2,..., u n } be an ordered basis for a finitedimensional vector space V. For x V, let a 1, a 2,..., a n be the
5 unique scalars such that x = [x] β = a i u i We define the coordinate vector of x relative to β, denoted [x] β, by a 1 a 2 Defn 5. (p. 80) Let V and W be finitedimensional vector spaces with ordered bases β = {v 1, v 2,..., v n } and γ = {w 1, w 2,..., w n }, respectively. Let T : V W be linear. Then for each j, 1 j n, there exists unique scalars a i,j F, a i m, such that m T (v j ) = t i,j w i 1 j n. So that [T (v j )] γ = We call the m n matrix A defined by A i,j = a i,j the matrix representation of T in the ordered bases β and γ and write A = [T ] γ β. If V = W and β = γ, then we write A = [T ] β.. a n. t 1,j t 2,j. t m,j. Fact 2. [a 1 x 1 + a 2 x a n x n ] B = a 1 [x 1 ] B + a 2 [x 2 ] B + + a n [x n ] B Proof. Let β = {v 1, v 2,..., v n }. Say x j = b j,1 v 1 + b j,2 v b j,n v n. The LHS: a 1 x 1 + a 2 x a n x n = a 1 (b 1,1 v 1 + b 1,2 v b 1,n v n ) +a 2 (b 2,1 v 1 + b 2,2 v b 2,n v n ) + +a n (b n,1 v 1 + b n,2 v b n,n v n ) = (a 1 b 1,1 + a 2 b 2,1 + + a n b n,1 )v 1 +(a 1 b 1,2 + a 2 b 2,2 + + a n b n,2 )v 2 + The i th position of the LHS is the coefficient c i of v i in: +(a 1 b 1,n + a 2 b 2,n + + a n b n,n )v n a 1 x 1 + a 2 x a n x n = c 1 v 1 + c 2 v c n v n And so, c i = a 1 b 1,i + a 2 b 2,i + + a n b n,i. The RHS: The i th position of the RHS is the sum of the i th position of each a j [x j ] β, which is the coefficient of v i in a j x j = a j (b j,1 v 1 + b j,2 v b j,n v n ) and thus, a j b j,i. We have the i th position of the RHS is a 1 b 1,i + a 2 b 2,i + + a n b n,i. 5
6 6 Fact 2. (a) Let V and W be finite dimensional vector spaces with bases, β and γ, respectively. [T ] γ β [x] β = [T (x)] γ. Proof. Let β = {v 1, v 2,..., v n }. Let x = a 1 v 1 + a 2 v a n v n. Then [T (x)] γ = [T (α 1 v 1 + α 2 v α n v n )] γ = [α 1 T (v 1 ) + α 2 T (v 2 ) + + α n T (v n ))] γ = α 1 [T (v 1 )] γ + α 2 [T (v 2 )] γ + + α n [T (v n )] γ by fact 2 t 1,1 t 1,n α 1... = t n,1 t n,n α n = [T ] γ β [x] γ Fact 3. Let V be a vector space of dimension n with basis β. {x 1, x 2,..., x k } is linearly independent in V if and only if {[x 1 ] β, [x 2 ] β,..., [x k ] β } is linearly independent in F n. Proof. {x 1, x 2,..., x k } is linearly dependent a 1, a 2,..., a k, not all zero, such that a 1 x 1 + a 2 x 2 + a k x k = 0 a 1, a 2,..., a k, not all zero, such that [a 1 x 1 + a 2 x 2 + a k x k ] β = [0] β = 0 since the only way to represent 0( V ) in any basis is 0( F n ) a 1, a 2,..., a k, not all zero, such that a 1 [x 1 ] β + a 2 [x 2 ] β + a k [x k ] β = 0 {[x 1 ] β, [x 2 ] β,..., [x k ] β }is a linearly dependent set Defn 6. (p. 99) Let V and W be vector spaces, and let T : V W be linear. A function U : W V is said to be an inverse of T if T U = I W and UT = I V. If T has an inverse, then T is said to be invertible. Fact 4. v and W are vector spaces with T : V W. If T is invertible, then the inverse of T is unique. Proof. Suppose U : W V is such that T U = I W and UT = I V and X : W V is such that T X = I W and XT = I V. To show U = X, we must show that w W, U(w) = X(w). We know I W (w) = w and I V (X(w)) = X(w). U(w) = U(I W (w)) = UT X(w) = I V X(w) = X(w). Defn 7. We denote the inverse of T by T 1. Theorem Let V and W be vector spaces, and let T : V W be linear and invertible. Then T 1 : W V is linear.
7 Proof. By the definition of invertible, we have: w W, T (T 1 (w)) = w. Let x, y W. Then, (2) (3) (4) (5) T 1 (cx + y) = T 1 (ct (T 1 (x)) + T (T 1 (y))) = T 1 (T (ct 1 (x) + T 1 y)) = T 1 T (ct 1 (x) + T 1 y) = T 1 (x) + T 1 y 7 Fact 5. If T is a linear transformation between vector spaces of equal (finite) dimension, then the conditions of being a.) invertible, b.) onetoone, and c.) onto are all equivalent. Proof. We start with a) c). We have T T 1 = I V. To show onto, let v V then for x = T 1 (v), we have: T (x) = T (T 1 (v)) = I V (v) = v. Therefore, T is onto. By Theorem 2.5, we have b) c). We will show b) a). Let {v 1, v 2,..., v n } be a basis of V. Then {T (v 1 ), T (v 2 ),..., T (v n )} spans R(T ) by Theorem 2.2. One of the corollaries to the Replacement Theorem implies that {x 1 = T (v 1 ), x 2 = T (v 2 ),..., x n = T (v n )} is a basis for V. We define U : V V by, i, U(x i ) = v i. By Theorem 2.6, this is a welldefined linear transformation. Let x V, x = a 1 v 1 + a 2 v a n v n for some scalars a 1, a 2,..., a n. UT (x) = UT (a 1 v 1 + a 2 v a n v n ) = U(a 1 T (v 1 ) + a 2 T (v 2 ) + + a n T (v n ) = U(a 1 x 1 + a 2 x a n x n ) = a 1 U(x 1 ) + a 2 U(x 2 ) + + a n U(x n ) = a 1 v 1 + a 2 v a n v n = x Let x V, x = b 1 x 1 + b 2 x b n x n for some scalars b 1, b 2,..., b n. T U(x) = T U(b 1 x 1 + b 2 x b n x n ) = T (b 1 U(x 1 ) + b 2 T (x 2 ) + + b n T (x n ) = T (b 1 v 1 + b 2 v b n v n ) = b 1 T (v 1 ) + b 2 T (v 2 ) + + b n T (v n ) = b 1 x 1 + b 2 x b n x n = x Defn 8. (p. 100) Let A be an n n matrix. Then A is invertible if there exits an n n matrix B such that AB = BA = I. Fact 6. If A is invertible, then the inverse of A is unique. Proof. A M n (F ) and AB = BA = I n. Suppose AC = CA = I n. Then we have B = BI n = BAC = I n C = C.
8 8 Defn 9. We denote the inverse of A by A 1. Lemma 1. (p. 101) Let T be an invertible linear transformation from V to W. Then V is finitedimensional if and only if W is finitedimensional. In this case, dim(v ) = dim(w ). Proof. ( ) T : V W is invertible. Then T T 1 = I W and T 1 T = I V. Assume dim V = n and {v 1, v 2,..., v n } is a basis of V. We know T is onto since, if y W then for x = T 1 (y), we have T (x) = y. We know R(T ) = W since T is onto and {T (v 1 ), T (v 2 ),..., T (v n )} spans W, by Theorem 2.2. So W is finite dimensional and dim W n = dim V. ( ) Assume dim W = m. T 1 : W V is invertible. Applying the same argument as in the last case, we obtain that V is finite dimensional and dim V dim W. We see that when either of V or W is finite dimensional then so is the other and so dim W dim V and dim V dim W imply that dim V = dim W. Theorem Let V and W be finitedimensional vector spaces with ordered bases β and γ, respectively. Let T : V W be linear. Then T is invertible if and only if [T ] γ β is invertible. Furthermore, [T 1 ] β γ = ([T ] γ β ) 1. Proof. Assume T is invertible. Let dim V = n and dim W = m. By the lemma, n = m. Let β = { 1, 2,..., n } and γ = {w 1, w 2,..., w n }. We have T 1 such that T T 1 = I W and T 1 T = I V. We will show that [T 1 ] β γ is the inverse of [T ] γ β. We will show the following 2 matrix equations: (6) (7) [T ] γ β [T 1 ] β γ = I n [T 1 ] β γ[t ] γ β = I n To prove (1) we will take the approach of showing that the i th column of [T ] γ β [T 1 ] β γ is e i, the i th characteristic vector. Notice that for any matrix A, Ae i is the i th column of A. Consider [T ] γ β [T 1 ] β γe i. By repeated use of Fact 2a, we have: [T ] γ β [T 1 ] β γe i = [T ] γ β [T 1 ] β γ[w i ] γ = [T ] γ β [T 1 (w i )] β = [T T 1 (w i )] γ = [w i ] γ = e i To prove (7) the proof is very similar. We will show that the i th column of [T 1 ] β γ[t ] γ β is e i, the i th.
9 9 Consider [T 1 ] β γ[t ] γ β e i. By repeated use of Fact 2a, we have: [T 1 ] β γ[t ] γ β e i = [T 1 ] β γ[t ] γ β [v i] β = [T 1 ] β γ[t (v i )] γ = [T 1 (T (v i ))] β = [v i ] β = e i Thus we have ( ) and the Furthermore part of the statement. Now we assume A = [T ] γ β is invertible. Call its inverse A 1. We know that it is square, thus n = m. Let β = { 1, 2,..., n } and γ = {w 1, w 2,..., w n }. Notice that the i th column of A is A i = (t 1,i, t 2,i,..., t n,i ) where T (v i ) = t 1,i w 1 + t 2,i w t n,i w n To show T is invertible, we will define a function U and prove that it is the inverse of T. Let the i th column of A 1 be C i = (c 1,i, c 2,i,..., c n,i ). We define U : W V by for all i [n], U(w i ) = c 1,i 1 +c 2,i c n,i n. Since we have defined U for the basis vectors γ, we know from Theorem 2.6 that U is a linear transformation. We wish to show that T U is the identity transformation in W and UT is the identity transformation in V. So, we show (1) T U(x) = x, x W and (2) UT (x) = x, x V. Starting with (1.). First lets see why, i [n], T (U(w i )) = w i. T (U(w i )) = T (c 1,i 1 +c 2,i c n,i n ) = c 1,i (t 1,1 w 1 + t 2,1 w t n,1 w n ) +c 2,i (t 1,2 w 1 + t 2,2 w t n,2 w n ) Gathering coefficients of w 1, w 2,..., w n, we have: + + c n,i (t 1,n w 1 + t 2,n w t n,n w n ) T (U(w i )) = (c 1,i t 1,1 + c 2,i t 1,2 + + c n,i t 1,n )w 1 +(c 1,i t 2,1 + c 2,i t 2,2 + + c n,i t 2,n )w (c 1,i t n,1 + c 2,i t n,2 + + c n,i t n,n )w n We see that the coefficient of w j in the above expression is Row j of A dot Column i of A 1, which is always equal to 1 if and only if i = j. Thus we have that T (U(w i )) = w i. Now we see that for x W then x = b 1 w 1 +b 2 w 2 + +b n w n for some scalars b 1, b 2,..., b n. T U(x) = T U(b 1 w 1 + b 2 w b n w n ) = T (b 1 U(w 1 ) + b 2 U(w 2 ) + + b n U(w n )) = b 1 T (U(w 1 )) + b 2 T (U(w 2 )) + + b n T (U(w n )) = b 1 w 1 + b 2 w b n w n = x
10 10 Similarly, UT (x) = x for all x in V. Cor 1. Let V be a finitedimensional vector space with an ordered basis β, and let T : V V be linear. Then T is invertible if and only if [T ] β is invertible. Furthermore, [T 1 ] β = ([T ] β ) 1. Proof. Clear. Defn 10. Let A M m n (F ). Then the function L A : F n F m where for x F n, L A (x) = Ax and is called leftmultiplication by A. Fact 6. (a) L A is a linear transformation and for β = {e 1, e 2,..., e n }, [L A ] β = A. Proof. We showed in class. Cor 2. Let A be an n n matrix. Furthermore, (L A ) 1 = L A 1. Then A is invertible if and only if L A is invertible. Proof. Let V = F n. Then L A : V V. Apply Corollary 1 with β = {e 1, e 2,..., e n } and [L A ] β = A, we have: L A is invertible if and only if A is invertible. The furthermore part of this corollary says that (L A ) 1 is left multiplication by A 1. We will show L A L A 1 = L A 1L A = I V. So we must show L A L A 1(x) = x, x V and L A 1L A (x) = x, x V. We have L A L A 1(x) = AA 1 x = x and L A 1L A (x) = A 1 Ax = x. Defn 11. Let V and W be vector spaces. We say that V is isomorphic to W if there exists a linear transformation T : V W that is invertible. Such a linear transformation is called an isomorphism from V onto W. Theorem Let V and W be finitedimensional vector spaces (over the same field). Then V is isomorphic to W if and only if dim(v ) = dim(w ). Proof. ( ) Let T : V W be an isomorphism. Then T is invertible and dim V = dim W by the Lemma. ( ) Assume dim V = dim W. Let β = {v 1,..., v k } and γ = {w 1, w 2,..., w k } be bases for V and W, respectively. Define T : V W where v i w i for all i. Then [T ] γ β = I which is invertible and Theorem 2.18 implies that T is invertible and thus, an isomorphism. Fact 7. Let P M n (F ) be invertible. W is a subspace of F n implies L P (W ) is a subspace of F n and dim(l P (W )) = dim(w ). Proof. Let x, y L P (W ), a F, then there exist x and y such that P x = x and P y = y. So P (ax + y ) = ap x + P y = ax + y. So ax + y L P (W ). Also, we know 0 W and L P (0) = 0 since L P is linear, so we have that 0 L P (W ). Therefore, L P (W ) is a subspace. Let {x 1, x 2,..., x k } be a basis of W. a 1 P (x 1 ) + a 2 P (x 2 ) + a k P (x k ) = 0 P (a 1 x 1 + a 2 x 2 + a k x k ) = 0 a 1 x 1 + a 2 x 2 + a k x k = 0 a 1 = a 2 = = a k = 0 {P (x 1 ), P (x 2 ),..., P (x k )} is linearly independent
11 To show that {P (x 1 ), P (x 2 ),..., P (x k )} spans L P (W ), we let z L P (W ). Then there is some w W, such that L P (W ) = z. If w = a 1 x 1 + a 2 x a k x k, we have z = P w = P (a 1 x 1 + a 2 x a k x k ) = a 1 P (x 1 ) + a 2 P (x 2 ) + + a k P (x k ). And with have that {P (x 1 ), P (x 2 ),..., P (x k )} is a basis of L P (W ) and dim(w ) = dim(l P (W )). 11 So far, we did not define the rank of a matrix. We have: Defn 12. Let A M m n (F ). Then, Rank(A)= Rank(L A ). And the range of a matrix, R(A) is the same as R(L A ). Fact 8. Let S M n (F ). If S is invertible, then R(S) = F n. Proof. L S : F n F n. We know S is invertible implies L S is invertible. We already know that, since L S is onto, R(L S ) = F n. By the definition of R(S), we have R(S) = F n. Fact 9. S, T M n (F ), S invertible and T invertible imply ST is invertible and its inverse is T 1 S 1. Proof. This is the oldest proof in the book. Theorem Let V and W be finitedimensional vector spaces over F of dimensions n and m, respectively, and let β and γ be ordered bases for V and W, respectively. Then the function Φ : L(V, W ) M m n (F ) defined by Φ(T ) = [T ] γ β for T L(V, W ), is an isomorphism. Proof. To show Φ is an isomorphism, we must show that it is invertible. Let β = {v 1, v 2,..., v n } and γ = {w 1, w 2,..., w n }. We define U : M m n (F ) L(V, W ) as follows. Let A M m n (F ) and have i th column: a 1,i a 2,i. a n,i So U maps matrices to transformations. U(A) is a transformation in L(V, W ). We describe the action of the linear transformation U(A) on v i and realize that this uniquely defines a linear transformation in L(V, W ). i [n], U(A)(v i ) = a 1,i w 1 + a 2,i w a n,i w n. Then A = [U(A)] γ β. Thus, ΦU(A) = A. To verify that U(Φ)(T ) = T, we see what the action is on v i. U(Φ(T ))(v i ) = U([T ] γ β )(v i) = t 1,i w 1 + t 2,i w t n,i w n = T (v i ) Cor 1. Let V and W be finitedimensional vector spaces over F of dimensions n and m, respectively. Then L(V, W ) is finitedimensional of dimension mn.
12 12 Proof. This by Theorem Theorem Let β and γ be two ordered bases for a finitedimensional vector space V, and let Q = [I V ] γ β. Then (a) Q is invertible. (b) For any v V, [v] γ = Q[v] β. Proof. We see that Let β = {v 1,..., v n } and γ = {w 1,..., w n } We claim the inverse of Q is [I V ] β γ. We see that the i th column of [I V ] γ β [I V ] β γ is [I V ] γ β [I V ] β γe i. We have [I V ] γ β [I V ] β γe i = [I V ] γ β [I V ] β γ[w i ] γ = [I V ] γ β [I V (w i )] β = [I V (I V (w i ))] γ = [w i ] γ = e i and [I V ] β γ[i V ] γ β e i = [I V ] β γ[i V ] γ β [v i] β = [I V ] β γ[i V (v i )] γ = [I V (I V (v i ))] β = [v i ] β = e i We know, [v] γ = [I V (v)] γ = [I V ] γ β [v] β = Q[v] β. Defn 13. (p. 112) The matrix Q = [I V ] γ β defined in Theorem 2.22 is called a change of coordinate matrix. Because of part (b) of the theorem, we say that Q changes β coordinates into γcoordinates. Notice that Q 1 = [I V ] β γ. Defn 14. (p. 112) A linear transformation T : V V is called a linear operator. Theorem Let T be a linear operator on a finitedimensional vector space V, and let β and γ be ordered bases for V. Suppose that Q is the change of coordinate matrix that changes βcoordinates into γcoordinates. Then Proof. The statement is short for: [T ] β = Q 1 [T ] γ Q. [T ] β β = Q 1 [T ] γ γq.
13 13 Let β = {v 1, v 2,..., v n }. As usual, we look at the i th column of each side. We have The i th column of [T ] β β. Q 1 [T ] γ γqe i = [I V ] β γ[t ] γ γ[i V ] γ β [v i] β = [I V ] β γ[t ] γ γ[i V (v i )] γ = [I V ] β γ[t ] γ γ[v i ] γ = [I V ] β γ[t (v i )] γ = [I V (T (v i ))] β = [T (v i )] β = [T ] β β [v i] β = [T ] β β e i Cor 1. Let A M n n (F ), and let γ be an ordered basis for F n. Then [L A ] γ = Q 1 AQ, where Q is the n n matrix whose j th column is the j th vector of γ. Proof. Let β be the standard ordered basis for F n. Then A = [L A ] β. So by the theorem, [L A ] γ = Q 1 [L A ] β Q. Defn 15. (p. 115) Let A and B be in M n (F ). We say that B is similar to A if there exists an invertible matrix Q such that B = Q 1 AQ. Defn 16. (p. 119) For a vector space V over F, we define the dual space of V to be the vector space L(V, F ), denoted by V. Let β = {x 1, x 2,..., x n } be an ordered basis for V. For each i [n], we define f i (x) = a i where a i is the i th coordinate of [x] β. Then f i is in V called the i th coordinate function with respect to the basis β. Theorem Suppose that V is a finitedimensional vector space with the ordered basis β = {x 1, x 2,..., x n }. Let f i (1 i n) be the ith coordinate function with respect to β and let β = {f 1, f 2,..., f n }. Then β is an ordered basis for V, and, for any f V, we have f = f(x i )f i. Proof. Let f V I. Since dim V = n, we need only show that f = f(x i )f i, from which it follows that β generates V, and hence is a basis by a Corollary to the Replacement Theorem. Let g = f(x i )f i. For 1 j n, we have ( ) g(x j ) = f(x i )f i (x j ) = f(x i )f i (x j ) = f(x i )δ i,j = f(x j ).
14 14 Therefore, f = g.
Math 333  Practice Exam 2 with Some Solutions
Math 333  Practice Exam 2 with Some Solutions (Note that the exam will NOT be this long) Definitions (0 points) Let T : V W be a transformation Let A be a square matrix (a) Define T is linear (b) Define
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationSection 6.1  Inner Products and Norms
Section 6.1  Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More informationProblems for Advanced Linear Algebra Fall 2012
Problems for Advanced Linear Algebra Fall 2012 Class will be structured around students presenting complete solutions to the problems in this handout. Please only agree to come to the board when you are
More informationMatrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
More informationSec 4.1 Vector Spaces and Subspaces
Sec 4. Vector Spaces and Subspaces Motivation Let S be the set of all solutions to the differential equation y + y =. Let T be the set of all 2 3 matrices with real entries. These two sets share many common
More informationLecture Note on Linear Algebra 15. Dimension and Rank
Lecture Note on Linear Algebra 15. Dimension and Rank WeiShi Zheng, wszheng@ieee.org, 211 November 1, 211 1 What Do You Learn from This Note We still observe the unit vectors we have introduced in Chapter
More information2.1: MATRIX OPERATIONS
.: MATRIX OPERATIONS What are diagonal entries and the main diagonal of a matrix? What is a diagonal matrix? When are matrices equal? Scalar Multiplication 45 Matrix Addition Theorem (pg 0) Let A, B, and
More informationMA106 Linear Algebra lecture notes
MA106 Linear Algebra lecture notes Lecturers: Martin Bright and Daan Krammer Warwick, January 2011 Contents 1 Number systems and fields 3 1.1 Axioms for number systems......................... 3 2 Vector
More informationMethods for Finding Bases
Methods for Finding Bases Bases for the subspaces of a matrix Rowreduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 19967 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationUNIT 2 MATRICES  I 2.0 INTRODUCTION. Structure
UNIT 2 MATRICES  I Matrices  I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress
More informationMATH 240 Fall, Chapter 1: Linear Equations and Matrices
MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS
More informationUsing determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible:
Cramer s Rule and the Adjugate Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible: Theorem [Cramer s Rule] If A is an invertible
More information( ) which must be a vector
MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are
More informationMATH 304 Linear Algebra Lecture 4: Matrix multiplication. Diagonal matrices. Inverse matrix.
MATH 304 Linear Algebra Lecture 4: Matrix multiplication. Diagonal matrices. Inverse matrix. Matrices Definition. An mbyn matrix is a rectangular array of numbers that has m rows and n columns: a 11
More informationLinear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)
MAT067 University of California, Davis Winter 2007 Linear Maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007) As we have discussed in the lecture on What is Linear Algebra? one of
More informationLecture Notes: Matrix Inverse. 1 Inverse Definition
Lecture Notes: Matrix Inverse Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Inverse Definition We use I to represent identity matrices,
More informationMA 242 LINEAR ALGEBRA C1, Solutions to Second Midterm Exam
MA 4 LINEAR ALGEBRA C, Solutions to Second Midterm Exam Prof. Nikola Popovic, November 9, 6, 9:3am  :5am Problem (5 points). Let the matrix A be given by 5 6 5 4 5 (a) Find the inverse A of A, if it exists.
More informationLinear Transformations
a Calculus III Summer 2013, Session II Tuesday, July 23, 2013 Agenda a 1. Linear transformations 2. 3. a linear transformation linear transformations a In the m n linear system Ax = 0, Motivation we can
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationInverses and powers: Rules of Matrix Arithmetic
Contents 1 Inverses and powers: Rules of Matrix Arithmetic 1.1 What about division of matrices? 1.2 Properties of the Inverse of a Matrix 1.2.1 Theorem (Uniqueness of Inverse) 1.2.2 Inverse Test 1.2.3
More informationMATH 110 Spring 2015 Homework 6 Solutions
MATH 110 Spring 2015 Homework 6 Solutions Section 2.6 2.6.4 Let α denote the standard basis for V = R 3. Let α = {e 1, e 2, e 3 } denote the dual basis of α for V. We would first like to show that β =
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More informationMATH1231 Algebra, 2015 Chapter 7: Linear maps
MATH1231 Algebra, 2015 Chapter 7: Linear maps A/Prof. Daniel Chan School of Mathematics and Statistics University of New South Wales danielc@unsw.edu.au Daniel Chan (UNSW) MATH1231 Algebra 1 / 43 Chapter
More informationLecture 6. Inverse of Matrix
Lecture 6 Inverse of Matrix Recall that any linear system can be written as a matrix equation In one dimension case, ie, A is 1 1, then can be easily solved as A x b Ax b x b A 1 A b A 1 b provided that
More information160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
More informationVector Spaces II: Finite Dimensional Linear Algebra 1
John Nachbar September 2, 2014 Vector Spaces II: Finite Dimensional Linear Algebra 1 1 Definitions and Basic Theorems. For basic properties and notation for R N, see the notes Vector Spaces I. Definition
More informationMATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix.
MATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix. Inverse matrix Definition. Let A be an n n matrix. The inverse of A is an n n matrix, denoted
More informationMath 315: Linear Algebra Solutions to Midterm Exam I
Math 35: Linear Algebra s to Midterm Exam I # Consider the following two systems of linear equations (I) ax + by = k cx + dy = l (II) ax + by = 0 cx + dy = 0 (a) Prove: If x = x, y = y and x = x 2, y =
More information1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
More informationDiagonal, Symmetric and Triangular Matrices
Contents 1 Diagonal, Symmetric Triangular Matrices 2 Diagonal Matrices 2.1 Products, Powers Inverses of Diagonal Matrices 2.1.1 Theorem (Powers of Matrices) 2.2 Multiplying Matrices on the Left Right by
More information4.1 VECTOR SPACES AND SUBSPACES
4.1 VECTOR SPACES AND SUBSPACES What is a vector space? (pg 229) A vector space is a nonempty set, V, of vectors together with two operations; addition and scalar multiplication which satisfies the following
More information1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
More information4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns
L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows
More information7  Linear Transformations
7  Linear Transformations Mathematics has as its objects of study sets with various structures. These sets include sets of numbers (such as the integers, rationals, reals, and complexes) whose structure
More informationMath 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
More informationB such that AB = I and BA = I. (We say B is an inverse of A.) Definition A square matrix A is invertible (or nonsingular) if matrix
Matrix inverses Recall... Definition A square matrix A is invertible (or nonsingular) if matrix B such that AB = and BA =. (We say B is an inverse of A.) Remark Not all square matrices are invertible.
More information4.6 Null Space, Column Space, Row Space
NULL SPACE, COLUMN SPACE, ROW SPACE Null Space, Column Space, Row Space In applications of linear algebra, subspaces of R n typically arise in one of two situations: ) as the set of solutions of a linear
More informationNOTES on LINEAR ALGEBRA 1
School of Economics, Management and Statistics University of Bologna Academic Year 205/6 NOTES on LINEAR ALGEBRA for the students of Stats and Maths This is a modified version of the notes by Prof Laura
More informationChapter 7. Matrices. Definition. An m n matrix is an array of numbers set out in m rows and n columns. Examples. ( 1 1 5 2 0 6
Chapter 7 Matrices Definition An m n matrix is an array of numbers set out in m rows and n columns Examples (i ( 1 1 5 2 0 6 has 2 rows and 3 columns and so it is a 2 3 matrix (ii 1 0 7 1 2 3 3 1 is a
More informationMatrices, transposes, and inverses
Matrices, transposes, and inverses Math 40, Introduction to Linear Algebra Wednesday, February, 202 Matrixvector multiplication: two views st perspective: A x is linear combination of columns of A 2 4
More informationMATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all ndimensional column
More information3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More information4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More information1 0 5 3 3 A = 0 0 0 1 3 0 0 0 0 0 0 0 0 0 0
Solutions: Assignment 4.. Find the redundant column vectors of the given matrix A by inspection. Then find a basis of the image of A and a basis of the kernel of A. 5 A The second and third columns are
More informationSection 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj
Section 5. l j v j = [ u u j u m ] l jj = l jj u j + + l mj u m. l mj Section 5. 5.. Not orthogonal, the column vectors fail to be perpendicular to each other. 5..2 his matrix is orthogonal. Check that
More informationDeterminants. Dr. Doreen De Leon Math 152, Fall 2015
Determinants Dr. Doreen De Leon Math 52, Fall 205 Determinant of a Matrix Elementary Matrices We will first discuss matrices that can be used to produce an elementary row operation on a given matrix A.
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More information1.5 Elementary Matrices and a Method for Finding the Inverse
.5 Elementary Matrices and a Method for Finding the Inverse Definition A n n matrix is called an elementary matrix if it can be obtained from I n by performing a single elementary row operation Reminder:
More informationOrthogonal Projections
Orthogonal Projections and Reflections (with exercises) by D. Klain Version.. Corrections and comments are welcome! Orthogonal Projections Let X,..., X k be a family of linearly independent (column) vectors
More informationDefinition 12 An alternating bilinear form on a vector space V is a map B : V V F such that
4 Exterior algebra 4.1 Lines and 2vectors The time has come now to develop some new linear algebra in order to handle the space of lines in a projective space P (V ). In the projective plane we have seen
More informationMatrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc.
2 Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Theorem 8: Let A be a square matrix. Then the following statements are equivalent. That is, for a given A, the statements are either all true
More informationLinear Dependence Tests
Linear Dependence Tests The book omits a few key tests for checking the linear dependence of vectors. These short notes discuss these tests, as well as the reasoning behind them. Our first test checks
More informationAu = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
More informationMatrices: 2.3 The Inverse of Matrices
September 4 Goals Define inverse of a matrix. Point out that not every matrix A has an inverse. Discuss uniqueness of inverse of a matrix A. Discuss methods of computing inverses, particularly by row operations.
More informationLecture 11. Shuanglin Shao. October 2nd and 7th, 2013
Lecture 11 Shuanglin Shao October 2nd and 7th, 2013 Matrix determinants: addition. Determinants: multiplication. Adjoint of a matrix. Cramer s rule to solve a linear system. Recall that from the previous
More informationMatrix Calculations: Kernels & Images, Matrix Multiplication
Matrix Calculations: Kernels & Images, Matrix Multiplication A. Kissinger (and H. Geuvers) Institute for Computing and Information Sciences Intelligent Systems Version: spring 2016 A. Kissinger Version:
More information17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function
17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function, : V V R, which is symmetric, that is u, v = v, u. bilinear, that is linear (in both factors):
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 34 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationSolutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus ndimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More informationVector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a nonempty
More informationNON SINGULAR MATRICES. DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that
NON SINGULAR MATRICES DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that AB = I n = BA. Any matrix B with the above property is called
More informationFacts About Eigenvalues
Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v
More informationUniversity of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
More informationTopic 1: Matrices and Systems of Linear Equations.
Topic 1: Matrices and Systems of Linear Equations Let us start with a review of some linear algebra concepts we have already learned, such as matrices, determinants, etc Also, we shall review the method
More informationRieszFredhölm Theory
RieszFredhölm Theory T. Muthukumar tmk@iitk.ac.in Contents 1 Introduction 1 2 Integral Operators 1 3 Compact Operators 7 4 Fredhölm Alternative 14 Appendices 18 A AscoliArzelá Result 18 B Normed Spaces
More information5.3 ORTHOGONAL TRANSFORMATIONS AND ORTHOGONAL MATRICES
5.3 ORTHOGONAL TRANSFORMATIONS AND ORTHOGONAL MATRICES Definition 5.3. Orthogonal transformations and orthogonal matrices A linear transformation T from R n to R n is called orthogonal if it preserves
More informationPractice Math 110 Final. Instructions: Work all of problems 1 through 5, and work any 5 of problems 10 through 16.
Practice Math 110 Final Instructions: Work all of problems 1 through 5, and work any 5 of problems 10 through 16. 1. Let A = 3 1 1 3 3 2. 6 6 5 a. Use Gauss elimination to reduce A to an upper triangular
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationName: Section Registered In:
Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are
More information5. Linear algebra I: dimension
5. Linear algebra I: dimension 5.1 Some simple results 5.2 Bases and dimension 5.3 Homomorphisms and dimension 1. Some simple results Several observations should be made. Once stated explicitly, the proofs
More informationa 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More informationLinear Codes. In the V[n,q] setting, the terms word and vector are interchangeable.
Linear Codes Linear Codes In the V[n,q] setting, an important class of codes are the linear codes, these codes are the ones whose code words form a subvector space of V[n,q]. If the subspace of V[n,q]
More information10. Graph Matrices Incidence Matrix
10 Graph Matrices Since a graph is completely determined by specifying either its adjacency structure or its incidence structure, these specifications provide far more efficient ways of representing a
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An ndimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0534405967. Systems of Linear Equations Definition. An ndimensional vector is a row or a column
More informationMAT188H1S Lec0101 Burbulla
Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u
More informationTHE DIMENSION OF A VECTOR SPACE
THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field
More informationGROUP ALGEBRAS. ANDREI YAFAEV
GROUP ALGEBRAS. ANDREI YAFAEV We will associate a certain algebra to a finite group and prove that it is semisimple. Then we will apply Wedderburn s theory to its study. Definition 0.1. Let G be a finite
More informationSummary of week 8 (Lectures 22, 23 and 24)
WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry
More informationNotes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 201112 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More informationLinear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
More informationPROVING STATEMENTS IN LINEAR ALGEBRA
Mathematics V2010y Linear Algebra Spring 2007 PROVING STATEMENTS IN LINEAR ALGEBRA Linear algebra is different from calculus: you cannot understand it properly without some simple proofs. Knowing statements
More information2.1 Functions. 2.1 J.A.Beachy 1. from A Study Guide for Beginner s by J.A.Beachy, a supplement to Abstract Algebra by Beachy / Blair
2.1 J.A.Beachy 1 2.1 Functions from A Study Guide for Beginner s by J.A.Beachy, a supplement to Abstract Algebra by Beachy / Blair 21. The Vertical Line Test from calculus says that a curve in the xyplane
More information1 Orthogonal projections and the approximation
Math 1512 Fall 2010 Notes on least squares approximation Given n data points (x 1, y 1 ),..., (x n, y n ), we would like to find the line L, with an equation of the form y = mx + b, which is the best fit
More informationRow and column operations
Row and column operations It is often very useful to apply row and column operations to a matrix. Let us list what operations we re going to be using. 3 We ll illustrate these using the example matrix
More informationSolving Linear Systems, Continued and The Inverse of a Matrix
, Continued and The of a Matrix Calculus III Summer 2013, Session II Monday, July 15, 2013 Agenda 1. The rank of a matrix 2. The inverse of a square matrix Gaussian Gaussian solves a linear system by reducing
More informationChapter 6. Linear Transformation. 6.1 Intro. to Linear Transformation
Chapter 6 Linear Transformation 6 Intro to Linear Transformation Homework: Textbook, 6 Ex, 5, 9,, 5,, 7, 9,5, 55, 57, 6(a,b), 6; page 7 In this section, we discuss linear transformations 89 9 CHAPTER
More informationMATH36001 Background Material 2015
MATH3600 Background Material 205 Matrix Algebra Matrices and Vectors An ordered array of mn elements a ij (i =,, m; j =,, n) written in the form a a 2 a n A = a 2 a 22 a 2n a m a m2 a mn is said to be
More informationMatrix Inverse and Determinants
DM554 Linear and Integer Programming Lecture 5 and Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1 2 3 4 and Cramer s rule 2 Outline 1 2 3 4 and
More informationNotes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 918/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
More informationSolutions to Assignment 4
Solutions to Assignment 4 Math 412, Winter 2003 3.1.18 Define a new addition and multiplication on Z y a a + 1 and a a + a, where the operations on the righthand side off the equal signs are ordinary
More informationWeek 5 Integral Polyhedra
Week 5 Integral Polyhedra We have seen some examples 1 of linear programming formulation that are integral, meaning that every basic feasible solution is an integral vector. This week we develop a theory
More information