1 Sets and Set Notation.


 Maurice Richard
 1 years ago
 Views:
Transcription
1 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most often name sets using capital letters, like A, B, X, Y, etc., while the elements of a set will usually be given lowercase letters, like x, y, z, v, etc. Two sets X and Y are called equal if X and Y consist of exactly the same elements. In this case we write X = Y. Example (Examples of Sets). () Let X be the collection of all integers greater than or equal to 5 and strictly less than. Then X is a set, and we may write: X = {5, 6, 7, 8, 9} The above notation is an example of a set being described explicitly, i.e. just by listing out all of its elements. The set brackets { } indicate that we are talking about a set and not a number, sequence, or other mathematical object. (2) Let E be the set of all even natural numbers. We may write: E = {, 2, 4, 6, 8,...} This is an example of an explicity described set with infinitely many elements. The ellipsis (...) in the above notation is used somewhat informally, but in this case its meaning, that we should continue counting forever, is clear from the context. (3) Let Y be the collection of all real numbers greater than or equal to 5 and strictly less than. Recalling notation from previous math courses, we may write: Y = [5, ) This is an example of using interval notation to describe a set. Note that the set Y obviously consists of infinitely many elements, but that there is no obvious way to write down the elements of Y explicitly like we did the set E in Example (2). Even though [5, ) is a set, we don t need to use the set brackets in this case, as interval notation has a wellestablished meaning which we have used in many other math courses. (4) Now and for the remainder of the course, let the symbol denote the empty set, that is, the unique set which consists of no elements. Written explicitly, = { }. (5) Now and for the remainder of the course, let the symbol N denote the set of all natural numbers, i.e. N = {,, 2, 3,...}. (6) Now and for the remainder of the course, let the symbol R denote the set of all real numbers. We may think of R geometrically as being the collection of all the points on the number line.
2 2 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES (7) Let R 2 denote the set of all ordered pairs of real numbers. That is, let R 2 be the set which consists of all pairs (x, y) where x and y are both real numbers. We may think of R 2 geometrically as the set of all points on the Cartesian coordinate plane. If (x,[ y) is ] an element of R 2, it will often be convenient for us to write the pair as the column x vector. For our purposes the two notations will be interchangeable. It is important to y [ ] [ ] x y note here that the order matters when we talk about pairs, so in general we have. y x (8) Let R 3 be the set of all ordered triples of real numbers, i.e. R 3 is the set of all triples (x, y, z) such that x, y, and z are all real numbers. R 3 may be visualized geometrically as the set of all points in 3dimensional Euclidean coordinate space. We will also write elements (x, y, z) of R 3 be using the column vector notation x y z. (9) Lastly and most generally, let n be any natural number. We will let R n be the set of all ordered ntuples of real numbers, i.e. the set of all ntuples (x, x 2,..., x n ) for which each x coordinate x i, i n, is a real number. We will also use the column vector notation x 2... x n in this context. Definition 2 (Set Notation). If A is a set and x is an element of A, then we write: x A. If B is a set such that every element of B is an element of A (i.e. if x B then x A), then we call B a subset of A and we write: B A. In order to distinguish particular subsets we wish to talk about, we will frequently use setbuilder notation, which for convenience we will describe informally using examples, rather than give a formal definition. For an example, suppose we wish to formally describe the set E of all even positive integers (See Example (2)). Then we may write E = {x N : x is evenly divisible by 2}. The above notation should be read as The set of all x in N such that x is evenly divisible by 2, which clearly and precisely defines our set E. For another example, we could write Y = {x R : 5 x < }, which reads The set of all x in R such that 5 is less than or equal to x and x is strictly less than. The student should easily verify that Y = [5, ) from Example (3). In general, given a set A and a precise mathematical sentence P (x) about a variable x, the setbuilder notation should be read as follows. { x A : P (x)} The set of all elements x in A such that sentence P (x) is true for the element x.
3 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 3 2 Vector Spaces and Subspaces. Definition 3. A (real) vector space is a nonempty set V, whose elements are called vectors, together with an operation +, called addition, and an operation, called scalar multiplication, which satisfy the following ten axioms: Addition Axioms. () If u V and v V, then u + v V. (Closure under addition.) (2) u + v = v + u for all u, v V. (Commutative property of addition.) (3) ( u + v) + w = u + ( v + w) for all u, v, w V. (Associative property of addition.) (4) There exists a vector V which satisfies u + = u for all u V. (Existence of an additive identity.) (5) For every u V, there exists a vector u V such that u + ( u) =. (Existence of additive inverses.) Scalar multiplication axioms. (6) If u V and c R, then c u V. (Closure under scalar multiplication.) (7) c ( u + v) = c u + c v for all c R, u, v V. (First distributive property of multiplication over addition.) (8) (c + d) u = c u + d u for all c, d R, u V. (Second distributive property of multiplication over addition.) (9) c (d u) = (c d) u for all c, d R, u V. (Associative property of scalar multiplication.) () u = u for every u V. We use the arrow above notation to help differentiate vectors ( u, v, etc.), which may or may not be real numbers, from scalars, which are always real numbers. When no confusion will arise, we will often drop the symbol in scalar multiplcation and simply write c u instead of c u, c( u + v) instead of c ( u + v), etc. Example 2. Let V be an arbitrary vector space. () Prove that + u = u for every u V. (2) Prove that the zero vector is unique. That is, prove that if w V has the property that u + w = u for every u V, then we must have w =. (3) Prove that for every u V, the additive inverse u is unique. That is, prove that if w V has the property that u + w =, then we must have w = u. Proof. () By Axiom (2), the commutativity of addition, we have + u = u +. Hence by Axiom (4), we have + u = u + = u.
4 4 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES (2) Suppose w V has the property that u + w = u for every u V. Then in particular, we have + w =. But + w = w by part () above; so w = + w =. (3) Let u V, and suppose w V has the property that u + w =. Let u be the additive inverse of u guaranteed by Axiom (5). Adding u to both sides of the equality above, and applying Axioms (2) and (3) (commutativity and associativity), we get u + ( u + w) = u + ( u + u) + w = u ( u + ( u)) + w = u + w = u. Now by part () above, we have w = + w = u. Example 3 (Examples of Vector Spaces). () The real number line R is a vector space, where both + and are interpreted in the usual way. In this case the axioms are the familiar properties of real numbers which we learn {[ in ] elementary} school. x (2) Consider the plane R 2 = : x, y R. Define an operation + on R y 2 by the rule [ ] [ ] [ ] x z x + z + = for all x, y, z, w R, y w y + w and a scalar multiplication by the rule [ x c y ] = [ cx cy ] for every c R, x, y R. Then R 2 becomes a vector space. (Verify that each of the axioms holds.) (3) Consider the following purely geometric description. Let V be the set of all arrows in twodimensional space, with two arrows being regarded as equal if they have the same length and point in the same direction. Define an addition + on V as follows: if u and v are two arrows in V, then lay them endtoend, so the base of v lies at the tip of u. Then define u + v to be the arrow which shares its base with u and its tip with v. (A picture helps here.) Define a scalar multiplication by letting c u be the arrow which points in the same direction as u, but whose length is c times the length of u. Is V a vector space? (What is the relationship of V with R 2?) (4) In general if n N, n, then R n is a vector space, where the addition and scalar multiplication are coordinatewise a la part (2) above. (5) Let n N, and let P n denote the set of all polynomials of degree at most n. That is, P n consists of all polynomials of the form p(x) = a + a x + a 2 x a n x n where the coefficients a, a,..., a n are real numbers, and x is an abstract variable. Define + and as follows: Suppose c R, and p, q P n, so p(x) = a + a x a n x n and q(x) = b + b x b n x n for some coefficients a,..., a n, b,..., b n R. Then (p + q)(x) = p(x) + q(x) = (a + b ) + (a + b )x (a n + b n )x n
5 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 5 and Then P n is a vector space. (cp)(x) = cp(x) = ca + ca x ca n x n. Definition 4. Let V be a vector space, and let W V. If W is also a vector space using the same operations + and inherited from V, then we call W a vector subspace, or just subspace, of V. Example 4. For each of the following, prove or disprove your answer. {[ ] } x () Let V = R y 2 : x, y, so V is the first quadrant of the Cartesian plane. Is V a vector subspace of R 2? (2) Let V be the set of all points on the graph of y = 5x. Is V a vector subspace of R 2? (3) Let V be the set of all points on the graph of y = 5x +. Is V a vector subspace of R 2? (4) Is R a vector subspace of R 2? (5) Is { } a vector subspace of R 2? (This space is called the trivial space.) Proof. (2) We claim that the set V of all points on the graph of y = 5x is indeed a subspace of R 2. To prove this, observe that since V is a subset of R 2 with the same addition and scalar multiplication operations, it is immediate that Axioms (2), (3), (7), (8), (9), and () hold (Verify this in your brain!) So we need only check Axioms (), (4), (5), and (6). [ ] [ ] x x2 If, V, then by the definition of V we have y y y = 5x and y 2 = 5x 2. It follows 2 [ ] x + x that y +y 2 = 5x +5x 2 = 5(x +x 2 ). Hence the sum 2 satisfies the defining condition y + y [ ] [ ] [ ] 2 x + x of V, and so 2 x x2 = + V. So V is closed under addition and Axiom () y + y 2 y y 2 is satisfied. [ ] Check that the additive identity is in V, so Axiom (3) is satisfied. [ ] [ ] [ ] x x cx If V and c R, then c = V since cy y y cy = c(5x ) = 5(cx ), so V is closed under scalar multiplication. Hence Axiom (6)[ is satisfied. ] Moreover, if we take x c = in the previous equalities, we see that each vector V has an additive inverse y [ ] x V. So Axiom (5) is satisfied. Thus V meets all the criteria to be a vector space, as y we claimed. (3) On the other hand, if we take V to be the set of all points on the graph of y = 5x +, then V is not a vector subspace of R 2. To see this, it suffices to check, for instance, that V fails Axiom (), i.e. V is not closed under addition. To show that V is not [ closed ] under addition, [ ] it suffices to exhibit two vectors in V whose sum is not in V. So let u = and let v =. Both u and v are in V. (Why?) But their sum [ ] 6 u + v = is not a solution of the equation y = 5x +, and hence not in V. So V fails to 7 satisfy Axiom (), and cannot be a vector space. We will not prove the following fact, but the reader should think about why it must be true.
6 6 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Fact. Let V be a vector space, and let W V. Then W, together with the addition and scalar multiplication inherited from V, satisfies Axioms (2), (3), (7), (8), (9), () in the definition of a vector space. Theorem. Let V be a vector space and let W V. Then W is a subspace of V if and only if the following three properties hold: () W. (2) W is closed under addition. That is, for each u, v W, we have u + v W. (3) W is closed under scalar multiplication. That is, for each c R and u W, we have c u W. Proof. One direction of the proof is trivial: if W is a vector subspace of V, then W satisfies the three conditions above because it satisfies Axioms (4), (), and (6) respectively in the definition of a vector space. Conversely, suppose W V, and W satisfies the three conditions above. The three conditions imply that W satisfies Axioms (4), (), and (6), respectively, while our previous fact implies that W also satisfies Axioms (2), (3), (7), (8), (9), and (). So the only axiom left to check is (5). To see that (5) is satisfied, let u W. Since W V, the vector u is in V, and hence has an additive inverse u V. We must show that in fact u W. Note that since W is closed under scalar multiplication, the vector u is in W. But on our homework problem #(d), we show that in fact u = u. So u = u V as we hoped, and the proof is complete. Example 5. What do the vector subspaces of R 2 look like? What about R 3? 3 Linear Combinations and Spanning Sets. Definition 5. Let V be a vector space. Let v, v 2,..., v n V, and let c, c 2,..., c n R. We say that a vector u defined by u = c v + c 2 v c n v n is a linear combination of the vectors v,..., v n with weights c,..., c n. Notice that since addition is associative in any vector space V, we may omit parentheses from the sum above. Also notice that the zero vector is always a linear combination of any collection of vectors v,..., v n, since we may always take c = c 2 =... = c n =. Example 6. Let v = [ ] [ 2 and let v 2 = () Which points in R 2 are linear combinations of v and v 2, using integer weights? (2) Which points in R 2 are linear combinations of v and v 2, using any weights? 2 Example 7. Let a = 2 and a 2 = 5 be vectors in R () Let b = 7 4. May b be written as a linear combination of a and a 2? 3 6 (2) Let b = 3. May b be written as a linear combination of a and a 2? 5 Partial Solution. which ]. () We wish to answer the question: Do there exist real numbers c and c 2 for c a + c 2 a 2 = b?
7 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 7 Written differently, are there c, c 2 R for which c c = c + 2c 2 2c + 5c 2 5c + 6c 2 = Recall that vectors in R 3 are equal if and only if each pair of entries are equal. So to identify a c and c 2 which make the above equality true, we wish to solve the following system of equations in two variables: 7 4 3? c + 2c 2 = 7 2c + 5c 2 = 4 5c + 6c 2 = 3 This can be done manually using elementary algebra techniques. We should get c = 3 and c 2 = 2. Since a solution exists, b = 3 a + 2 a 2 is indeed a linear combination of a and a 2. Definition 6. Let V be a vector space and v,..., v n V. We denote the set of all possible linear combinations of v,..., v n by Span{ v,..., v n }, and we call this set the subset of V spanned by v,..., v n. We also call Span{ v,..., v n } the subset of V generated by v,..., v n. {[ ]} 3 Example 8. () Find Span in R 2. (2) Find Span, in R3. Theorem 2. Let V be a vector space and let W V. Then W is a subspace of V if and only if W is closed under linear combinations, i.e., for every v,..., v n W and every c,..., c n R, we have c v c n v n W. Corollary. Let V be a vector space and v,..., v n V. Then Span{ v,..., v n } is a subspace of V. 4 Matrix Row Reductions and Echelon Forms. As our previous few examples should indicate, our techniques for solving systems of linear equations are extremely relevant to our understanding of linear combinations and spanning sets. The student should recall the basics of solving systems from a previous math course, but in this section we will develop a new notationally convenient method for finding solutions to such systems. Definition 7. Recall that a linear equation in n variables is an equation that can be written in the form a x + a 2 x a n x n = b where a,..., a n, b R and x,..., x n are variables. A system of linear equations is any finite collection of linear equations involving the same variables x,..., x n. A solution to the system is an ntuple (s, s 2,..., s n ) R n, that makes each equation in the system true if we substitute s,..., s n for x,..., x n respectively. The set of all possible solutions to a system is called the solution set. Two systems are called equivalent if they have the same solution set. It is only possible for a system of linear equations to have no solutions, exactly one solution, or infinitely many solutions. (Geometrically one may think of parallel lines, intersecting lines, and coincident lines, respectively.
8 8 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES A system of linear equations is called consistent if it has at least one solution; if the system s solution set is, then the system is inconsistent. Example 9. Solve the following system of three equations in three variables. x 2x 2 + x 3 = 2x 2 8x 2 = 8 4x + 5x 2 + 9x 3 = 9 Solution. Our goal here will be to solve the system using the elimination technique, but with a stripped down notation which preserves only the necessary information and makes computations by hand go quite a bit faster. First let us introduce a little terminology. Given the system we are working with, the matrix of coefficients is the 3x3 matrix below The augmented matrix of the system is the 3x4 matrix below Note that the straight black line between the third and fourth columns in the matrix above is not necessary. Its sole purpose is to remind us where the equals sign was in our original system, and may be omitted if the student prefers. To solve our system of equations, we will perform operations on the augmented matrix which encode the process of elimination. For instance, if we were using elimination on the given system, we might add 4 times the first equation to the third equation, in order to eliminate the x variable from the third equation. Let s do the analogous operation to our matrix instead. Add 4 times the top row to the bottom row: Notice that the new matrix we obtained above can also be interpreted as the augmented matrix of a system of linear equations, which is the new system we would have obtained by just using the elimination technique. In particular, the first augmented matrix and the new augmented matrix represent systems which are equivalent to one another. (However, the two matrices are obviously not equal to one another, which is why we will stick to the arrow notation instead of using equals signs!) Now we will continue the process, replacing our augmented matrix with a new, simpler augmented matrix, in such a way that the linear systems the matrices represent are all equivalent to one another. Multiply the second row by 2 :
9 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 9 Add 3 times the second row to the third row: Add 4 times the third row to the second row: Add times the third row to the first row: Add 2 times the second row to the first row: Now let s stop and examine what we ve done. following system: x = 29 x 2 = 6 x 3 = 3 So the system is solved, and has a unique solution (29, 6, 3). The last augmented matrix above represents the Definition 8. Given an augmented matrix, the three elementary row operations are the following. () (Replacement) Replace one row by the sum of itself and a multiple of another row. (2) (Interchange) Interchange two rows. (3) (Scaling) Multiply all entries in a row by a nonzero constant. These are the legal moves when solving systems of equations using augmented matrices. Two matrices are called row equivalent if one can be transformed into the other using a finite sequence of elementary row operations. Fact 2. If two augmented matrices are row equivalent, then the two linear systems they represent are equivalent, i.e. they have the same solution set. Definition 9. A matrix is said to be in echelon form, or row echelon form, if it has the following three properties: () All rows which have nonzero entries are above all rows which have only zero entries. (2) Each leading nonzero entry of a row is in a column to the right of the leading nonzero entry of the row above it. (3) All entries in a column below a leading nonzero entry are zeros.
10 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES A matrix is said to be in reduced echelon form or reduced row echelon form if it is in echelon form, and also satisfies the following two properties: (4) Each nonzero leading entry is. (5) Each leading is the only nonzero entry in its column. Every matrix may be transformed, via elementary row operations, into a matrix in reduced row echelon form. This process is called row reduction. Example. Use augmented matrices and row reductions to find solution sets for the following systems of equations. () (2) 2x 6x 3 = 8 x 2 + 2x 3 = 3 3x + 6x 2 2x 3 = 4 x 5x 2 + 4x 3 = 3 2x 7x 2 + 3x 3 = 2 2x + x 2 + 7x 3 = 5 Functions and Function Notation Definition. Let X and Y be sets. A function f from X to Y is just a rule by which we associate to each element x X, an element f(x) Y. The input set X is called the domain of f, while the set Y of possible outputs is called the codomain of f. To denote the domain and codomain when we define a function, we write f : X Y. We regard the terms map, mapping, and transformation as all being synonymous with function. Example. () For the familiar quadratic function f(x) = x 2, we would write f : R R, since f takes real numbers for input and returns real numbers for output. Notice that the codomain R here is distinct from the range, i.e. the set of all actual outputs of the function, which the reader should know is just the set [, ). This is an ok part of the notation. (2) On the other hand, for the familiar hyperbolic function f(x) = x, we would NOT write f : R R; this is because R but is not part of the domain of f. 6 Linear Transformations Definition. Let V and W be vector spaces. A function T : V W is a linear transformation if () T ( v + w) = T ( v) + T ( w) for every v, w V. (2) T (c v) = ct ( v) for every v V and c R. Note that the transformations which are linear are those which respect the addition and scalar multiplication operations of the vector spaces involved. Example 2. Are the following maps linear transformations? [ ] x () T : R R, T (x) =. 3x [ ] x (2) T : R R 2, T (x) = x 2. [ ] 2 (3) T : R 2 R 2, T ( v) = A v, where A =.
11 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Fact 3. Every matrix transformation T : R n R m is a linear transformation. Fact 4. If T : V W is a linear transformation, then T preserves linear combinations. In other words, for every collection of vectors v,..., v n V and every choice of weights c,..., c n R, we have T (c v c n v n ) = c T ( v ) c n T ( v n ). Definition 2. Let V = R n be ndimensional Euclidean space. Define vectors e,..., e n V by: e =... ; e 2 =... ;...; e n = These vectors e,..., e n are called the standard basis for R n. Note that any vector v =.... v v 2... v n may be written as a linear combination of the standard basis vectors in an obvious way: v = v e + v 2 e v n e n. 5 Example 3. Suppose T : R 2 R 3 is a linear transformation such that T ( e ) = 7 and T ( e 2 ) = Find an explicit formula for T. Theorem 3. Let T : R n R m (n, m N) be a linear transformation. Then there exists a unique m n matrix A such that T ( v) = A v for every v R n. In fact, we have A = T ( e ) T ( e 2 )... T ( e n ). Proof. Notice that for any v = v v 2... Rn, we have v = v e v n e n. Since T is a linear transformation, it respects linear combinations, and hence v n T ( v) = v T ( e ) v n T ( e n ) = T ( e ) T ( e 2 )... T ( e n ) = A v, where A is as we defined in the statement of the theorem. Example 4. Let T : R 2 R 2 be the linear transformation which scales up every vector by a factor of 3, i.e. T ( v) = 3 v for every v R 2. Find a matrix representation for T. Example 5. Let T : R 2 R 2 be the linear transformation which rotates the plane about the origin at some fixed angle θ. Is T a linear transformation? If so, find its matrix representation. v v 2... v n
12 2 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 7 The Null Space of a Linear Transformation Definition 3. Let V and W be vector spaces, and let T : V W be a linear transformation. Define the following set: Nul T = { v V : T ( v) = }. Then this set Nul T is called the null space or kernel of the map T. Example 6. Let T : R 3 R 3 be the matrix transformation T ( v) = A v, where A = () Let u = (2) Let w = Is u Nul T?. Is w Nul T? Theorem 4. Let V, W be vector spaces and T : V W a linear transformation. Then Nul T is a vector subspace of V. Proof. We will show that Nul T is closed under taking linear combinations, and hence the result will follow from Theorem 2. Let v,..., v n Nul T and c,..., c n R be arbitrary. We must show that c v c n v n Nul T, i.e. that T (c v c n v n ) =. To see this, simply observe that since T is a linear transformation, we have. T (c v c n v n ) = c T ( v c n T ( v n ). But since v,..., v n Nul T, we have T ( v ) =... = T ( v n ) =. So in fact we have T (c v c n v n ) = c c n =. It follows that c v +...+c n v n Nul T, and Nul T is closed under linear combinations. This completes the proof. Example 7. For the following matrix transformations, give an explicit description of Nul T by finding a spanning set. x [ ] x () T : R 4 R 2, T x x 3 = x x 3 for every x, x 2, x 3, x 4 R. x 4 x 4 (2) T : R 5 R 3, T x x 2 x 3 x 4 x 5 = x x 2 x 3 x 4 x 5 for every x, x 2, x 3, x 4, x 5 R. Definition 4. Let V, W be sets. A mapping T : V W is called onetoone if the following holds: for every v, w V, if v w, then T ( v) T ( w).
13 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 3 Equivalently, T is onetoone if the following holds: for every v, w V, if T ( v) = T ( w), then v = w). A map is onetoone only if it sends distinct elements in V to distinct elements in W, i.e. no two elements in V are mapped to the same place in W. Example 8. Are the following linear transformations onetoone? [ ] x () T : R R 2, T (x) =. 3x (2) T : R 2 R 2, T ([ x x 2 ]) = [ ] [ x x 2 ]. Theorem 5. A linear transformation T : V W is onetoone if and only if Nul T = { }. Proof. ( ) First suppose T is onetoone, and let v Nul T. We will show v =. To see this, note that T ( v) = since v Nul T. But we also have Nul T since Nul T is a subspace of V, so T ( ) = = T ( v). Since T is onetoone, we must have = v, and hence Nul T is the trivial subspace Nul T = { }. ( ) Conversely, suppose T is not onetoone; we will show Nul T is nontrivial. Since T is not onetoone, there exist two distinct vectors v, w V, v w, such that T ( v) = T ( w). Set u = v w. Since v w and additive inverses are unique, we have u, and we also have T ( u) = T ( v w) = T ( v) T ( w) = T ( v) T ( v) =. So u Nul T. Since u is not the zero vector, Nul T { }. Definition 5. Let V, W be sets. A mapping T : V W is called onto if the following holds: for every w W, there exists a vector v V such that T ( v) = w. Example 9. Are the following linear transformations onto? [ ] x () T : R R 2, T (x) =. 3x ([ ]) (2) T : R 2 x R, T = x x 2 x. ([ 2 ]) [ ] (3) T : R 2 R 2 x x, T = 2. x 2 2x + x 2 Definition 6. Let V, W be vector spaces and T : V W a linear transformation. If T is both onetoone and onto, then T is called an isomorphism. In this case the domain V and codomain W are called isomorphic as vector spaces, or just isomorphic. It means that V and W are indistinguishable from one another in terms of their vector space structure. Example 2. Prove that the mapping T : R 2 R 2, where T rotates the plane by a fixed angle θ, is an isomorphism. Example 2. Let W be the graph of the line y = 3x, a vector subspace of R 2. isomorphic to W. Prove that R is 8 The Range Space of a Linear Transformation and the Column Space of a Matrix Definition 7. Let V, W be vector spaces and T : V W a linear transformation. Define the following set: Ran T = { w W : there exists a vector v V such that T ( v) = w}.
14 4 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Then Ran T is called the range of T. (This definition should coincide with the student s knowledge of the range of a function from previous courses.) Theorem 6. Let V, W be vector spaces and T : V W a linear transformation. Then Ran T is a vector subspace of W. Proof. Again we will show that Ran T is closed under linear combinations, and appeal to Theorem 2. To that end, let w,..., w n Ran T and let c,..., c n R all be arbitrary. We must show c w c n w n Ran T. To see this, note that since w,..., w n Ran T, there exist vectors v,..., v n V such that T ( v ) = w,..., T ( v n ) = w n. Set u = c v c n v n. Since V is a vector space, V is closed under taking linear combinations and hence u V. Moreover, we have T ( u) = T (c v c n v n ) = c T ( v ) c n T ( v n ) = c w c n w n. So the vector c w c n w n is the image of u under T, and hence c w c n w n Ran T. This shows Ran T is closed under linear combinations and hence a vector subspace of W. Theorem 7. Let V, W be vector spaces and T : V W a linear transformation. Then T is onto if and only if Ran T = W. Proof. Obvious if you think about it! Corollary 2. Let V, W be vector spaces and T : V W a linear transformation. Then the following statements are all equivalent: () T is an isomorphism. (2) T is onetoone and onto. (3) Nul T = { } and Ran T = W. Definition 8. Let m, n N, and let A be an m n matrix (so A induces a linear transformation from R n into R m ). Write A = w n w..., where each of w,..., w n is a column vector in R m. Define the following set: Col A = Span{ w,..., w n }. Then Col A is called the column space of the matrix A. Col A is exactly the set of all vectors in W which may be written as a linear combination of the columns of the matrix A. Of course Col A is also a vector subspace of R m, by Corollary.
15 Example 22. Let A = LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 5 2. () Determine whether or not the vector 3 5 is in Col A. (2) What does Col A look like in R 3? 6a b Example 23. Let W = a + b : a, b R. Find a matrix A so that W = Col A. 7a Theorem 8. Let T : R n R m (n, m N) be a linear transformation. Let A be the m n matrix representation of A guaranteed by Theorem 3. Then Ran T = Col A. Proof. Recall from Theorem 3 that the matrix A is given by A = T ( e ) T ( e 2 )... T ( e n ), where e,..., e n are the standard basis vectors in R n. So its columns are just T ( e ),..., T ( e n ). The first thing we will show is that Ran T Col A. So suppose w Ran T. Then there exists some v vector v R n for which T ( v) = w. Write v =... = v e v n e n for some real numbers v,..., v n R. Then, since T is a linear transformation, we have v n w = T ( v) = T (v e v n e n ) = v T ( e v n T ( e n ). So the above equality displays w as a linear combination of the columns of A, with weights v,..., v n. This proves w Col A. Since w was taken arbitrarily out of Ran T, we must have Ran T Col A. The next thing we will show is that Col A Ran T. So let w Col A. Then w can be written as a linear combination of the columns of A, so there exist some weights c,..., c n R for which w = c T ( e ) c n T ( e n ). Set v = c e c n e n. So v R n, and we claim that T ( v) = w. To see this, just compute (again using the fact that T is linear): T ( v) = T (c e +...c n e n ) = c T ( e ) c n T ( e n ) = w.
16 6 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES This shows w Ran T, and again since w was taken arbitrarily out of Col A, we have shown that Col A Ran T. Since Ran T Col A and Col A Ran T, we must have Ran T = Col A. This completes the proof. Example 24. Let T : R 2 [ R] 2 be the linear transformation defined by T ( e ) = 2 e + 4 e 2 and 2 T ( e 2 ) = e + 2 e 2. Let w =. Is w Ran T? 9 Linear Independence and Bases. Definition 9. Let V be a vector space and let v,..., v n V. Consider the equation: c v c n v n = where c,..., c n are interpreted as real variables. Notice that c =... = c n = is always a solution to the equation, which is called the trivial solution, but that there may be others depending on our choice of v,..., v n. If there exists a nonzero solution (c,..., c n ) to the equation, i.e. a solution where c k for at least one c k ( k n) then the vectors v,..., v n are called linearly dependent. Otherwise if the trivial solution is the only solution, then the vectors v,..., v n are called linearly independent. Example 25. Let v = 2 3, v 2 = 4 5 6, and v 3 = () Are the vectors v, v 2, v 3 linearly independent? (2) If possible, find a dependence relation, i.e. a nontrivial linear combination of v, v 2, v 3 which sums to. Fact 5. Let v,..., v n be column vectors in R m. Define an m n matrix A by: A = 2. v n v..., and define a linear transformation T : R n R m by the rule T ( v) = A v for every v R n. Then the following statements are all equivalent: () The vectors v,..., v n are linearly independent. (2) Nul T = { }. (3) T is onetoone. [ 2 Example 26. Let v = ] [ 4 and v 2 = ]. Are v and v 2 linearly independent? Fact 6. Let V be a vector space and v,..., v n V. Then v,..., v n are linearly dependent if and only if there exists some k {,..., n} such that v k Span{ v,..., v k, v k+,..., v n }, i.e. v k can be written as a linear combination of the other vectors. [ ] 2 Example 27. Let v = R 2. () Describe all vectors v 2 R 2 for which v, v 2 are linearly independent.
17 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 7 (2) Describe all pairs of vectors v 2, v 3 R 2 for which v, v 2, v 2 are linearly independent. Example 28. Let v = and v 2 = 3. Describe all vectors v 3 R 3 for which v, v 2, v 3 are 3 linearly independent. Definition 2. Let V be a vector space and let b,..., b n V. The set { b,..., b n } V is called a basis for V if () b,..., b n are linearly independent, and (2) V = Span{ b,..., b n }. 3 Example 29. Let v = 6, v 2 = [ ] 2 Example 3. Let v = and v 3 2 = () Is { v, v 2 } a basis for R 2? [ ] (2) What if v 2 =? [ 5, and v 3 = Fact 7. The standard basis { e,..., e n } is a basis for R n. Fact 8. The set {, x, x 2,..., x n } is a basis for P n. ] Is { v, v 2, v 3 } a basis for R 3? Fact 9. Let V be any vector space and let B = { b,..., b n } be a basis for V. () If you remove any one element b k from B, then the resulting set no longer spans V. This is because b k is linearly independent from the other vectors in B, and hence cannot be written as a linear combination of them by Fact 6. In this sense a basis is a minimal spanning set. (2) If you add any one element v, not already in B, to B, then the result set is no longer linearly independent. This is because B spans V, and hence there are no vectors in V which are independent from those in B by Fact 6. In this sense a basis is a maximal linearly independent set. Theorem 9 (Unique Representation Theorem). Let V be a vector space and let B = { b,..., b n } be a basis for V. Then every vector v V may be written as a linear combination v = c b c n bn in one and only one way, i.e. v has a unique representation with respect to B. Proof. Since { b,..., b n } is a basis for V, in particular the set spans V, so we have v Span{ b,..., b n }. It follows that there exist some scalars c,..., c n R n for which v = c b c n bn. So we need only check that the representation above is unique. To that end, suppose d,..., d n R is another set of scalars for which v = d b d n bn. We will show that in fact d = c,..., d n = c n, and hence there is really only one choice of scalars to begin with. To see this, observe the following equalities:
18 8 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES (d c ) b (d n c n ) b n = [d b d n bn ] [c b c n bn ] = v v =. So we have written as a linear combination of b,..., b n. But the collection is a basis and hence linearly independent, so the coefficients must all be zero, i.e. d c =... = d n c n =. The conclusion now follows. Dimension Theorem (Steinitz Exchange Lemma). Let V be a vector space and let B = { b,..., b n } be a basis for V. Let v V. By Theorem 9 there are unique scalars c,..., c n for which v = c b c n bn. If there is k {,..., n} for which c k, then exchanging v for b k yields another basis for the space, i.e. the collection { b,..., b k, v, b k+,..., b n } is a basis for V. Proof. Call the new collection ˆB = { b,..., b k, v, b k+,..., b n }. We must show that ˆB is both linearly independent and that it spans V. First we check that ˆB is linearly independent. Consider the equation d b d k b k + d k v + d k+ b k d nbn =. Let s substitute the representation for v in the above: d b d k b k + d k (c b c nbn ) + d k+ b k d nbn =. Now collecting like terms we get: (d + d k c k ) b (d k + d k c k ) b k + d k c k bk + (d k+ + d k c k+ ) b k (d n + d k c n ) b n =. Now since b,..., b n are linearly independent, all the coefficients in the above equation must be zero. In particular, we have d k c k =. But c k by our hypothesis, so we may divide through by c k and get d k =. Now substituting back in for d k we have: d b +...d k b k + d k+ b k d nbn. Now using the linear independence of b,..., b n one more time, we conclude that d =...d k = d k = d k+ =... = d n =. So the only solution is the trivial solution, and hence ˆB is linearly independent. It remains only to check that V = Span ˆB. To see this, let w V be arbitrary. By Theorem 9, write w as a linear combination of the vectors in B: w = a b a n bn
19 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES 9 for some scalars a,..., a n R. Now notice that since v = c b +...c n bn and since c k, we may solve for the vector b k as follows: b k = c c b k... c k b k + c k v c k+ b k+... cn c bn k. c k Note that for the above to make sense, it is crucial that c k. Now, if we substitute the above expression for b k in the equation w = a b a nbn and then collect like terms, we will see w written as a linear combination of the vectors b,..., b k, b k+,..., b n and the vector v. This implies that w Span ˆB. Since w was arbitrary in V, we have V = Span ˆB. So ˆB is indeed a basis for V and the theorem is proved. Corollary 3. Let V be a vector space which has a finite basis. Then every basis of V is the same size. Proof. Let B = { b,..., b n } be a basis of minimal size. Let D be any other basis for V, consisting of arbitrarily many elements of V. We will show that in fact D has exactly n elements. Let d D be arbitrary. Since d V = Span{ b,..., b n }, it is possible to write d as a nontrivial linear combination d = c b c n bn, i.e. c k for at least one k {,.., n}. By reordering the terms of the basis B if necessary, we may assume without loss of generality that c. Define a new set B = { d, b 2,..., b n }; by the Steinitz Exchange Lemma, B is also a basis for V. Now we go one step further. Since B was a basis of minimal size, we know D has at least n many elements. So let d 2 D be distinct from d. We know d 2 V = Span B, so there exist constants c,..., c n for which d 2 = c d + c 2 b c n bn. Notice that we must have c k for some k {2,..., n}, since if c were the only nonzero coefficient, then would be a scalar multiple of, which is not the case since they are linearly independent. By reordering the terms b 2,..., b n if necessary, we may assume without loss of generality that c 2. Define B 2 = { d, d 2, b 3,..., b 4 }. By the Steinitz Exchange Lemma, B 2 is also a basis for V. Continue this process. At each step k {,..., n}, we will have obtained a new basis B k = { d,..., d k, b k+,..., b n } for V. Since D has at least n many elements, there is a d k+ V which is not equal to any of d,..., d k. Write d k+ = c d c kdk + c k+ b k c nbn for some constants c,..., c n, and observe that one of the constants c k+,..., c n must be nonzero since d k+ is independent from d,..., d k. Then reorder the terms b k+,..., b n and use the Steinitz Exchange Lemma to replace b k+ with d k+ to obtain a new basis. After finitely many steps we end up with the basis B n = { d,..., d n } D. Since B n spans V, it is not possible for D to contain any more elements, since nothing in V is linearly independent from the vectors in B n. So in fact D = B n and the proof is complete. Definition 2. Let V be a vector space, and suppose V has a finite basis B = { b,..., b n } consisting of n vectors. Then we say that the dimension of V is n, or that V is ndimensional. We also write that dim V = n. Notice that every finitedimensional space V has a unique dimension by Corollary 3. If V has no finite basis, we say that V is infinitedimensional. Example 3. Determine the dimension of the following vector spaces. () R. (2) R 2. (3) R 3. (4) R n. c k
20 2 LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES (5) P n. (6) The trivial space { }. (7) Let V be the set of all infinite sequences (a, a 2, a 3,...) of real numbers which eventualy end in an infinite string of s. (Verify that V is a vector space using coordinatewise addition and scalar multiplication.) (8) The space P of all polynomials. (Verify that P is a vector space using standard polynomial addition and scalar multiplication.) (9) Let V be the set of all continuous functions f : R R. (Verify that V is a vector space using function addition and scalar multiplication. () Any line through the origin in R 3. () Any line through the origin in R 2. (2) Any plane through the origin in R 3. Corollary 4. Suppose V is a vector space with dim V = n. Then no linearly independent set in V consists of more than n elements. Proof. This follows from the proof of Corollary 3, since when considering D we only used the fact that D was linearly independent, not spanning. Corollary 5. Let V be finitedimensional vector space. expanded to make a basis for V. Any linearly independent set in V may be Proof. If a linearly independent set { d,..., d n } already spans V then it is already a basis. If it doesn t span, then we can find a linearly independent vector d n+ / Span{ d,..., d n } by Fact 6 and consider { d,..., d n, d n+ }. If it spans V, then we re done. Otherwise, continue adding vectors. By the previous corollary, we know the process stops in finitely many steps. Corollary 6. Let V be a finitedimensional vector space. Any spanning set in V may be shrunk to make a basis for V. Proof. Let S be a spanning set for V, i.e. V = Span S. Without loss of generality we may assume that / S, for if is in S, then we can toss it out and the set S will still span V. If S is empty then V = Span S = { } and S is already a basis, and we are done. If S is not empty, then choose s S. If V = Span{ s }, we are done and { s } is a basis. Otherwise, observe that if every vector in S were in Span{ s }, then we would have Span S = Span{ s }; since S spans V and { s } does not, this is impossible. So there is some s 2 S with s 2 / Span{ s }, i.e. s and s 2 are linearly independent. If { s, s 2 } spans V, we are done. Otherwise we can find a third linearly independent vector s 3, and so on. The process stops in finitely many steps. Corollary 7. Let V be an ndimensional vector space. Then a set of n vectors in V is linearly independent if and only if it spans V. Coordinate Systems Definition 22. Let V be a vector space and B = { b,..., b n } be a basis for V. Let v v. By the Unique Representation Theorem, there are unique weights c,..., c n R for which v = c b c n bn. Call
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 19967 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationDecember 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in twodimensional space (1) 2x y = 3 describes a line in twodimensional space The coefficients of x and y in the equation
More informationNotes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 918/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
More informationMATH 240 Fall, Chapter 1: Linear Equations and Matrices
MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An ndimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0534405967. Systems of Linear Equations Definition. An ndimensional vector is a row or a column
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationDeterminants. Dr. Doreen De Leon Math 152, Fall 2015
Determinants Dr. Doreen De Leon Math 52, Fall 205 Determinant of a Matrix Elementary Matrices We will first discuss matrices that can be used to produce an elementary row operation on a given matrix A.
More informationUNIT 2 MATRICES  I 2.0 INTRODUCTION. Structure
UNIT 2 MATRICES  I Matrices  I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress
More information( ) which must be a vector
MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are
More informationSolutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
More informationMath 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationLinear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
More information1 Eigenvalues and Eigenvectors
Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x
More informationMATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix.
MATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix. Inverse matrix Definition. Let A be an n n matrix. The inverse of A is an n n matrix, denoted
More informationSec 4.1 Vector Spaces and Subspaces
Sec 4. Vector Spaces and Subspaces Motivation Let S be the set of all solutions to the differential equation y + y =. Let T be the set of all 2 3 matrices with real entries. These two sets share many common
More informationThe Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 34 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationSolving Systems of Linear Equations
LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how
More informationChapter 1  Matrices & Determinants
Chapter 1  Matrices & Determinants Arthur Cayley (August 16, 1821  January 26, 1895) was a British Mathematician and Founder of the Modern British School of Pure Mathematics. As a child, Cayley enjoyed
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationMath 315: Linear Algebra Solutions to Midterm Exam I
Math 35: Linear Algebra s to Midterm Exam I # Consider the following two systems of linear equations (I) ax + by = k cx + dy = l (II) ax + by = 0 cx + dy = 0 (a) Prove: If x = x, y = y and x = x 2, y =
More informationMATH1231 Algebra, 2015 Chapter 7: Linear maps
MATH1231 Algebra, 2015 Chapter 7: Linear maps A/Prof. Daniel Chan School of Mathematics and Statistics University of New South Wales danielc@unsw.edu.au Daniel Chan (UNSW) MATH1231 Algebra 1 / 43 Chapter
More informationLEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. FiniteDimensional
More informationMATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all ndimensional column
More informationName: Section Registered In:
Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are
More informationLecture 6. Inverse of Matrix
Lecture 6 Inverse of Matrix Recall that any linear system can be written as a matrix equation In one dimension case, ie, A is 1 1, then can be easily solved as A x b Ax b x b A 1 A b A 1 b provided that
More informationDETERMINANTS. b 2. x 2
DETERMINANTS 1 Systems of two equations in two unknowns A system of two equations in two unknowns has the form a 11 x 1 + a 12 x 2 = b 1 a 21 x 1 + a 22 x 2 = b 2 This can be written more concisely in
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationa 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus ndimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More informationMATH10212 Linear Algebra B Homework 7
MATH22 Linear Algebra B Homework 7 Students are strongly advised to acquire a copy of the Textbook: D C Lay, Linear Algebra and its Applications Pearson, 26 (or other editions) Normally, homework assignments
More information3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More informationby the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the yaxis We observe that
More informationT ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
More informationChapter 8. Matrices II: inverses. 8.1 What is an inverse?
Chapter 8 Matrices II: inverses We have learnt how to add subtract and multiply matrices but we have not defined division. The reason is that in general it cannot always be defined. In this chapter, we
More informationSection 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
More informationRecall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
More information(a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular.
Theorem.7.: (Properties of Triangular Matrices) (a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular. (b) The product
More informationINTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL
SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics
More informationUniversity of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
More information2.1: MATRIX OPERATIONS
.: MATRIX OPERATIONS What are diagonal entries and the main diagonal of a matrix? What is a diagonal matrix? When are matrices equal? Scalar Multiplication 45 Matrix Addition Theorem (pg 0) Let A, B, and
More information2.1: Determinants by Cofactor Expansion. Math 214 Chapter 2 Notes and Homework. Evaluate a Determinant by Expanding by Cofactors
2.1: Determinants by Cofactor Expansion Math 214 Chapter 2 Notes and Homework Determinants The minor M ij of the entry a ij is the determinant of the submatrix obtained from deleting the i th row and the
More informationMA 242 LINEAR ALGEBRA C1, Solutions to Second Midterm Exam
MA 4 LINEAR ALGEBRA C, Solutions to Second Midterm Exam Prof. Nikola Popovic, November 9, 6, 9:3am  :5am Problem (5 points). Let the matrix A be given by 5 6 5 4 5 (a) Find the inverse A of A, if it exists.
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More information1 Determinants. Definition 1
Determinants The determinant of a square matrix is a value in R assigned to the matrix, it characterizes matrices which are invertible (det 0) and is related to the volume of a parallelpiped described
More informationNOTES on LINEAR ALGEBRA 1
School of Economics, Management and Statistics University of Bologna Academic Year 205/6 NOTES on LINEAR ALGEBRA for the students of Stats and Maths This is a modified version of the notes by Prof Laura
More informationMatrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
More informationAu = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
More informationLecture notes on linear algebra
Lecture notes on linear algebra David Lerner Department of Mathematics University of Kansas These are notes of a course given in Fall, 2007 and 2008 to the Honors sections of our elementary linear algebra
More informationMathematics Review for MS Finance Students
Mathematics Review for MS Finance Students Anthony M. Marino Department of Finance and Business Economics Marshall School of Business Lecture 1: Introductory Material Sets The Real Number System Functions,
More information4. MATRICES Matrices
4. MATRICES 170 4. Matrices 4.1. Definitions. Definition 4.1.1. A matrix is a rectangular array of numbers. A matrix with m rows and n columns is said to have dimension m n and may be represented as follows:
More informationTopic 1: Matrices and Systems of Linear Equations.
Topic 1: Matrices and Systems of Linear Equations Let us start with a review of some linear algebra concepts we have already learned, such as matrices, determinants, etc Also, we shall review the method
More informationMATH2210 Notebook 1 Fall Semester 2016/2017. 1 MATH2210 Notebook 1 3. 1.1 Solving Systems of Linear Equations... 3
MATH0 Notebook Fall Semester 06/07 prepared by Professor Jenny Baglivo c Copyright 009 07 by Jenny A. Baglivo. All Rights Reserved. Contents MATH0 Notebook 3. Solving Systems of Linear Equations........................
More informationLINEAR ALGEBRA. September 23, 2010
LINEAR ALGEBRA September 3, 00 Contents 0. LUdecomposition.................................... 0. Inverses and Transposes................................. 0.3 Column Spaces and NullSpaces.............................
More informationMatrix Inverse and Determinants
DM554 Linear and Integer Programming Lecture 5 and Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1 2 3 4 and Cramer s rule 2 Outline 1 2 3 4 and
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationMathematics Notes for Class 12 chapter 3. Matrices
1 P a g e Mathematics Notes for Class 12 chapter 3. Matrices A matrix is a rectangular arrangement of numbers (real or complex) which may be represented as matrix is enclosed by [ ] or ( ) or Compact form
More information4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
More information4.6 Null Space, Column Space, Row Space
NULL SPACE, COLUMN SPACE, ROW SPACE Null Space, Column Space, Row Space In applications of linear algebra, subspaces of R n typically arise in one of two situations: ) as the set of solutions of a linear
More information1. LINEAR EQUATIONS. A linear equation in n unknowns x 1, x 2,, x n is an equation of the form
1. LINEAR EQUATIONS A linear equation in n unknowns x 1, x 2,, x n is an equation of the form a 1 x 1 + a 2 x 2 + + a n x n = b, where a 1, a 2,..., a n, b are given real numbers. For example, with x and
More informationDiagonal, Symmetric and Triangular Matrices
Contents 1 Diagonal, Symmetric Triangular Matrices 2 Diagonal Matrices 2.1 Products, Powers Inverses of Diagonal Matrices 2.1.1 Theorem (Powers of Matrices) 2.2 Multiplying Matrices on the Left Right by
More informationCofactor Expansion: Cramer s Rule
Cofactor Expansion: Cramer s Rule MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Introduction Today we will focus on developing: an efficient method for calculating
More information26. Determinants I. 1. Prehistory
26. Determinants I 26.1 Prehistory 26.2 Definitions 26.3 Uniqueness and other properties 26.4 Existence Both as a careful review of a more pedestrian viewpoint, and as a transition to a coordinateindependent
More informationx1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.
Cross product 1 Chapter 7 Cross product We are getting ready to study integration in several variables. Until now we have been doing only differential calculus. One outcome of this study will be our ability
More informationUsing determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible:
Cramer s Rule and the Adjugate Using determinants, it is possible to express the solution to a system of equations whose coefficient matrix is invertible: Theorem [Cramer s Rule] If A is an invertible
More informationChapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm.
Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm. We begin by defining the ring of polynomials with coefficients in a ring R. After some preliminary results, we specialize
More informationFacts About Eigenvalues
Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v
More informationGENERATING SETS KEITH CONRAD
GENERATING SETS KEITH CONRAD 1 Introduction In R n, every vector can be written as a unique linear combination of the standard basis e 1,, e n A notion weaker than a basis is a spanning set: a set of vectors
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationChapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
More information2.5 Elementary Row Operations and the Determinant
2.5 Elementary Row Operations and the Determinant Recall: Let A be a 2 2 matrtix : A = a b. The determinant of A, denoted by det(a) c d or A, is the number ad bc. So for example if A = 2 4, det(a) = 2(5)
More informationMatrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.
Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that
More informationOrthogonal Projections
Orthogonal Projections and Reflections (with exercises) by D. Klain Version.. Corrections and comments are welcome! Orthogonal Projections Let X,..., X k be a family of linearly independent (column) vectors
More informationMAT188H1S Lec0101 Burbulla
Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u
More informationSolution of Linear Systems
Chapter 3 Solution of Linear Systems In this chapter we study algorithms for possibly the most commonly occurring problem in scientific computing, the solution of linear systems of equations. We start
More informationVector Spaces II: Finite Dimensional Linear Algebra 1
John Nachbar September 2, 2014 Vector Spaces II: Finite Dimensional Linear Algebra 1 1 Definitions and Basic Theorems. For basic properties and notation for R N, see the notes Vector Spaces I. Definition
More informationSolution to Homework 2
Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if
More informationThe Cayley Hamilton Theorem
The Cayley Hamilton Theorem Attila Máté Brooklyn College of the City University of New York March 23, 2016 Contents 1 Introduction 1 1.1 A multivariate polynomial zero on all integers is identically zero............
More informationMAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =
MAT 200, Midterm Exam Solution. (0 points total) a. (5 points) Compute the determinant of the matrix 2 2 0 A = 0 3 0 3 0 Answer: det A = 3. The most efficient way is to develop the determinant along the
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationMathematics of Cryptography
CHAPTER 2 Mathematics of Cryptography Part I: Modular Arithmetic, Congruence, and Matrices Objectives This chapter is intended to prepare the reader for the next few chapters in cryptography. The chapter
More informationDeterminants LECTURE Calculating the Area of a Parallelogram. Definition Let A be a 2 2 matrix. A = The determinant of A is the number
LECTURE 13 Determinants 1. Calculating the Area of a Parallelogram Definition 13.1. Let A be a matrix. [ a c b d ] The determinant of A is the number det A) = ad bc Now consider the parallelogram formed
More informationDeterminants, Areas and Volumes
Determinants, Areas and Volumes Theodore Voronov Part 2 Areas and Volumes The area of a twodimensional object such as a region of the plane and the volume of a threedimensional object such as a solid
More informationMAT 242 Test 2 SOLUTIONS, FORM T
MAT 242 Test 2 SOLUTIONS, FORM T 5 3 5 3 3 3 3. Let v =, v 5 2 =, v 3 =, and v 5 4 =. 3 3 7 3 a. [ points] The set { v, v 2, v 3, v 4 } is linearly dependent. Find a nontrivial linear combination of these
More informationChapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
More informationInverses and powers: Rules of Matrix Arithmetic
Contents 1 Inverses and powers: Rules of Matrix Arithmetic 1.1 What about division of matrices? 1.2 Properties of the Inverse of a Matrix 1.2.1 Theorem (Uniqueness of Inverse) 1.2.2 Inverse Test 1.2.3
More informationSYSTEMS OF EQUATIONS AND MATRICES WITH THE TI89. by Joseph Collison
SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI89 by Joseph Collison Copyright 2000 by Joseph Collison All rights reserved Reproduction or translation of any part of this work beyond that permitted by Sections
More informationMATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
More informationFactorization Theorems
Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization
More information13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in threespace, we write a vector in terms
More informationThe basic unit in matrix algebra is a matrix, generally expressed as: a 11 a 12. a 13 A = a 21 a 22 a 23
(copyright by Scott M Lynch, February 2003) Brief Matrix Algebra Review (Soc 504) Matrix algebra is a form of mathematics that allows compact notation for, and mathematical manipulation of, highdimensional
More informationUsing row reduction to calculate the inverse and the determinant of a square matrix
Using row reduction to calculate the inverse and the determinant of a square matrix Notes for MATH 0290 Honors by Prof. Anna Vainchtein 1 Inverse of a square matrix An n n square matrix A is called invertible
More informationMath 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = 36 + 41i.
Math 5A HW4 Solutions September 5, 202 University of California, Los Angeles Problem 4..3b Calculate the determinant, 5 2i 6 + 4i 3 + i 7i Solution: The textbook s instructions give us, (5 2i)7i (6 + 4i)(
More informationPractice Math 110 Final. Instructions: Work all of problems 1 through 5, and work any 5 of problems 10 through 16.
Practice Math 110 Final Instructions: Work all of problems 1 through 5, and work any 5 of problems 10 through 16. 1. Let A = 3 1 1 3 3 2. 6 6 5 a. Use Gauss elimination to reduce A to an upper triangular
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More information