# MA106 Linear Algebra lecture notes

Size: px
Start display at page:

Transcription

1 MA106 Linear Algebra lecture notes Lecturers: Martin Bright and Daan Krammer Warwick, January 2011 Contents 1 Number systems and fields Axioms for number systems Vector spaces Examples of vector spaces Linear independence, spanning and bases of vector spaces Linear dependence and independence Spanning vectors Bases of vector spaces Existence of a basis Subspaces 10 5 Linear transformations Definition and examples Kernels and images Rank and nullity Operations on linear maps Matrices 18 7 Linear transformations and matrices Setting up the correspondence The correspondence between operations on linear maps and matrices Linear equations and the inverse image problem Elementary operations and the rank of a matrix Gauss transformations Elementary row operations The augmented matrix Row reducing a matrix Elementary column operations The rank of a matrix The rank criterion The inverse of a linear transformation and of a matrix Definitions Matrix inversion by row reduction Elementary matrices Application to linear equations

2 10 The determinant of a matrix Definition of the determinant The effect of matrix operations on the determinant The determinant of a product Minors and cofactors The inverse of a matrix using determinants Cramer s rule for solving simultaneous equations Change of basis and equivalent matrices Similar matrices, eigenvectors and eigenvalues Similar matrices Eigenvectors and eigenvalues The scalar product symmetric and orthogonal matrices

3 1 Number systems and fields We introduce the number systems most commonly used in mathematics. 1. The natural numbers N = {1, 2, 3, 4,...}. In N, addition is possible but not subtraction; e.g. 2 3 N. 2. The integers Z = {..., 2, 1, 0, 1, 2, 3,...}. In Z, addition, subtraction and multiplication are always possible, but not division; e.g. 2/3 Z. 3. The rational numbers Q = {p/q p, q Z, q 0}. In Q, addition, subtraction, multiplication and division (except by zero) are all possible. However, 2 Q. 4. The real numbers R. These are the numbers which can be expressed as decimals. The rational numbers are those with finite or recurring decimals. In R, addition, subtraction, multiplication and division (except by zero) are still possible, and all positive numbers have square roots, but 1 R. 5. The complex numbers C = {x + iy x, y R}, where i 2 = 1. In C, addition, subtraction, multiplication and division (except by zero) are still possible, and all, numbers have square roots. In fact all polynomial equations with coefficients in C have solutions in C. 1.1 Axioms for number systems Laws governing the way numbers combine together are called axioms. Any particular axiom might be true in some number systems but not in others. Axioms for addition. Let S be a number system. A1. α + β = β + α for all α, β S. A2. (α + β) + γ = α + (β + γ) for all α, β, γ S. A3. There is a number 0 S such that α + 0 = 0 + α = α for all α S. A4. For each number α S there exists a number α S such that α + ( α) = ( α) + α = 0. These axioms may or may not be satisfied by a given number system S. For example, in N, A1 and A2 hold but A3 and A4 do not hold. A1 A4 all hold in Z, Q, R and C. Axioms for multiplication. M1. α.β = β.α for all α, β S. M2. (α.β).γ = α.(β.γ) for all α, β, γ S. M3. There is a number 1 S such that α.1 = 1.α = α for all α S. M4. For each number α S with α 0, there exists a number α 1 S such that α.α 1 = α 1.α = 1. In N and Z, M1 M3 hold but M4 does not hold. M1 M4 all hold in Q, R and C. 3

4 Axiom relating addition and multiplication. D. (α + β).γ = α.γ + β.γ for all α, β, γ S. Definition. A set S on which addition and multiplication are defined is called a field if it satisfies each of the axioms A1, A2, A3, A4, M1, M2, M3, M4, D, and if, in addition, 1 0. Roughly speaking, S is a field if addition, subtraction, multiplication and division (except by zero) are all possible in S. We shall always use the letter K for a general field. Example. N and Z are not fields, but Q, R and C are all fields. There are many other fields, including some finite fields. For example, for each prime number p, there is a field F p = {0, 1, 2,..., p 1} with p elements, where addition and multiplication are carried out modulo p. Thus, in F 7, we have = 2, 5 4 = 6 and 5 1 = 3 because 5 3 = 1. The smallest such field F 2 has just two elements 0 and 1, where 1+1 = 0. This field is extremely important in Computer Science since an element of F 2 represents a bit of information. Various other familiar properties of numbers, such as 0α = 0, ( α)β = (αβ) = α( β), ( α)( β) = αβ, ( 1)α = α, for all α, β S, can be proved from the axioms. Why would we want to do this, when we can see they re true anyway? The point is that, when we meet a new number system, it is enough to check whether the axioms hold; if they do, then all these properties follow automatically. However, occasionally you need to be careful. For example, in F 2 we have = 0, and so it is not possible to divide by 2 in this field. 2 Vector spaces Definition. A vector space over a field K is a set V which has two basic operations, addition and scalar multiplication, satisfying certain requirements. Thus for every pair u, v V, u + v V is defined, and for every α K, αv V is defined. For V to be called a vector space, the following axioms must be satisfied for all α, β K and all u, v V. (i) Vector addition satisfies axioms A1, A2, A3 and A4. (ii) α(u + v) = αu + αv; (iii) (α + β)v = αv + βv; (iv) (αβ)v = α(βv); (v) 1v = v. Elements of the field K will be called scalars. Note that we will use boldface letters like v to denote vectors. The zero vector in V will be written as 0 V, or usually just 0. This is different from the zero scalar 0 = 0 K K. For nearly all results in this course, there is no loss in assuming that K is the field R of real numbers. So you may assume this if you find it helpful to do so. Just occasionally, we will need to assume K = C the field of complex numbers. However, it is important to note that nearly all arguments in Linear Algebra use only the axioms for a field and so are valid for any field, which is why shall use a general field K for most of the course. 4

5 2.1 Examples of vector spaces 1. K n = {(α 1, α 2,..., α n ) α i K}. This is the space of row vectors. Addition and scalar multiplication are defined by the obvious rules: (α 1, α 2,..., α n ) + (β 1, β 2,..., β n ) = (α 1 + β 1, α 2 + β 2,..., α n + β n ); The most familiar examples are λ(α 1, α 2,..., α n ) = (λα 1, λα 2,..., λα n ). R 2 = {(x, y) x, y R} and R 3 = {(x, y, z) x, y, z R}, which we can think of geometrically as the points in ordinary 2- and 3-dimensional space, equipped with a coordinate system. Vectors in R 2 and R 3 can also be thought of as directed lines joining the origin to the points with coordinates (x, y) or (x, y, z). 0 =(0,0) (x, y) Addition of vectors is then given by the parallelogram law. v 1 + v 2 v 2 v 1 0 Note that K 1 is essentially the same as K itself. 2. Let K[x] be the set of polynomials in an indeterminate x with coefficients in the field K. That is, K[x] = {α 0 + α 1 x + + α n x n n 0, α i K}. Then K[x] is a vector space over K. 3. Let K[x] n be the set of polynomials over K of degree at most n, for some n 0. Then K[x] n is also a vector space over K; in fact it is a subspace of K[x]. Note that the polynomials of degree exactly n do not form a vector space. (Why not?) 4. Let K = R and let V be the set of n-times differentiable functions f : R R which are solutions of the differential equation d n f λ 0 dx n + λ d n 1 f 1 dx n λ df n 1 dx + λ nf = 0. for fixed λ 0, λ 1,..., λ n R. Then V is a vector space over R, for if f(x) and g(x) are both solutions of this equation, then so are f(x) + g(x) and αf(x) for all α R. 5. The previous example is a space of functions. There are many such examples that are important in Analysis. For example, the set C k ((0, 1), R), consisting of all functions f : (0, 1) R such that the kth derivative f (k) exists and is continuous, is a vector space over R with the usual pointwise definitions of addition and scalar multiplication of functions. 5

6 6. Any n bits of information can be thought of as a vector in F n 2. Facing such a variety of vector spaces, a mathematician wants to derive useful methods of handling all these vector spaces. If work out techniques for dealing with a single example, say R 3, how can we be certain that our methods will also work for R 8 or even C 8? That is why we use the axiomatic approach to developing mathematics. We must use only arguments based on the vector space axioms. We have to avoid making any other assumptions. This ensures that everything we prove is valid for all vector spaces, not just the familiar ones like R 3. We shall be assuming the following additional simple properties of vectors and scalars from now on. They can all be deduced from the axioms (and it is a useful exercise to do so). (i) α0 = 0 for all α K (ii) 0v = 0 for all v V (iii) (αv) = ( α)v = α( v), for all α K and v V. (iv) if αv = 0 then α = 0 or v = 0. 3 Linear independence, spanning and bases of vector spaces 3.1 Linear dependence and independence Definition. Let V be a vector space over the field K. The vectors v 1, v 2,... v n are said to be linearly dependent if there exist scalars α 1, α 2,..., α n K, not all zero, such that α 1 v 1 + α 2 v α n v n = 0. v 1, v 2,... v n are said to be linearly independent if they are not linearly dependent. In other words, they are linearly independent if the only scalars α 1, α 2,..., α n K that satisfy the above equation are α 1 = 0, α 2 = 0,..., α n = 0. Definition. Vectors of the form α 1 v 1 + α 2 v α n v n for α 1, α 2,..., α n K are called linear combinations of v 1, v 2,... v n. Example. Let V = R 2, v 1 = (1, 3), v 2 = (2, 5). Then α 1 v 1 + α 2 v 2 = (α 1 + 2α 2, 3α 1 + 5α 2 ), which is equal to 0 = (0, 0) if and only if α 1 + 2α 2 = 0 and 3α 1 + 5α 2 = 0. Thus we have a pair of simultaneous equations in α 1, α 2 and the only solution is α 1 = α 2 = 0, so v 1, v 2 are linearly independent. Example. Let V = Q 2, v 1 = (1, 3), v 2 = (2, 6). This time the equations are α 1 + 2α 2 = 0 and 3α 1 + 6α 2 = 0, and there are non-zero solutions, such as α 1 = 2, α 2 = 1, and so v 1, v 2 are linearly dependent. Lemma 3.1. v 1, v 2,..., v n V are linearly dependent if and only if either v 1 = 0 or, for some r, v r is a linear combination of v 1,..., v r 1. Proof. If v 1 = 0 then by putting α 1 = 1 and α i = 0 for i > 1 we get α 1 v 1 + +α n v n = 0, so v 1, v 2,..., v n V are linearly dependent. If v r is a linear combination of v 1,..., v r 1, then v r = α 1 v α r 1 v r 1 for some α 1,..., α r 1 K and so we get α 1 v α r 1 v r 1 1 v r = 0 and again v 1, v 2,..., v n V are linearly dependent. Conversely, suppose that v 1, v 2,..., v n V are linearly dependent, and α i are scalars, not all zero, satisfying α 1 v 1 + α 2 v α n v n = 0. Let r be maximal with α r 0; then α 1 v 1 + α 2 v α r v r = 0. If r = 1 then α 1 v 1 = 0 which, by (iv) above, is only possible if v 1 = 0. Otherwise, we get v r = α 1 α r v 1 α r 1 α r v r 1. In other words, v r is a linear combination of v 1,..., v r 1. 6

7 3.2 Spanning vectors Definition. The vectors v 1,..., v n in V span V if every vector v V is a linear combination α 1 v 1 + α 2 v α n v n of v 1,..., v n. 3.3 Bases of vector spaces Definition. The vectors v 1,..., v n in V form a basis of V if they are linearly independent and span V. Proposition 3.2. The vectors v 1,..., v n form a basis of V if and only if every v V can be written uniquely as v = α 1 v 1 + α 2 v α n v n ; that is, the coefficients α 1,..., α n are uniquely determined by the vector v. Proof. Suppose that v 1,..., v n form a basis of V. Then they span V, so certainly every v V can be written as v = α 1 v 1 + α 2 v α n v n. Suppose that we also had v = β 1 v 1 + β 2 v β n v n for some other scalars β i K. Then we have 0 = v v = (α 1 β 1 )v 1 + (α 2 β 2 )v (α n β n )v n and so (α 1 β 1 ) = (α 2 β 2 ) = = (α n β n ) = 0 by linear independence of v 1,..., v n. Hence α i = β i for all i, which means that the α i are uniquely determined. Conversely, suppose that every v V can be written uniquely as v = α 1 v 1 + α 2 v α n v n. Then v 1,..., v n certainly span V. If α 1 v 1 + α 2 v α n v n = 0, then α 1 v 1 + α 2 v α n v n = 0v 1 + 0v v n and so the uniqueness assumption implies that α 1 = α 2 = = α n = 0, and v 1,..., v n are linearly independent. Hence they form a basis of V. Definition. The scalars α 1,..., α n in the statement of the proposition are called the coordinates of v with respect to the basis v 1,..., v n. With respect to a different basis, v will have different coordinates. Indeed, a basis for a vector space can be thought of as a choice of a system of coordinates. Examples Here are some examples of bases of vector spaces. 1. (1, 0) and (0, 1) form a basis of K 2. This follows from Proposition 3.2, because each element (α 1, α 2 ) K 2 can be written as α 1 (1, 0) + α 2 (0, 1), and this expression is clearly unique. 2. More generally, (1, 0, 0), (0, 1, 0), (0, 0, 1) form a basis of K 3, (1, 0, 0, 0), (0, 1, 0, 0), (0, 0, 1, 0), (0, 0, 0, 1) form a basis of K 4 and so on. This is called the standard basis of K n for n N. (To be precise, the standard basis of K n is v 1,..., v n, where v i is the vector with a 1 in the ith position and a 0 in all other positions.) 3. There are many other bases of K n. For example (1, 0), (1, 1) form a basis of K 2, because (α 1, α 2 ) = (α 1 α 2 )(1, 0) + α 2 (1, 1), and this expression is unique. In fact, any two non-zero vectors such that one is not a scalar multiple of the other form a basis for K As we have defined a basis, it has to consist of a finite number of vectors. Not every vector space has a finite basis. For example, let K[x] be the space of polynomials in x with coefficients in K. Let p 1 (x), p 2 (x),..., p n (x) be any finite collection of polynomials in K[x]. Then, if d is the maximum degree of p 1 (x), p 2 (x),..., p n (x), any linear combination of p 1 (x), p 2 (x),..., p n (x) has degree at most d, and so p 1 (x), p 2 (x),..., p n (x) cannot span K[x]. On the other hand, it is possible (with a little care) to define what it means for an infinite set of vectors to be a basis of a vector space; in fact the infinite sequence of vectors 1, x, x 2, x 3,..., x n,... is a basis of K[x]. 7

8 A vector space with a finite basis is called finite-dimensional. In fact, nearly all of this course will be about finite-dimensional spaces, but it is important to remember that these are not the only examples. The spaces of functions mentioned in Example 5. of Section 2 typically have uncountably infinite dimension. Theorem 3.3 (The basis theorem). Suppose that v 1,..., v m and w 1,..., w n are both bases of the vector space V. Then m = n. In other words, all finite bases of V contain the same number of vectors. The proof of this theorem is quite tricky and uses the concept of sifting which we introduce after the next lemma. Definition. The number n of vectors in a basis of the finite-dimensional vector space V is called the dimension of V and we write dim(v ) = n. Thus, as we might expect, K n has dimension n. K[x] is infinite-dimensional, but the space K[x] n of polynomials of degree at most n has basis 1, x, x 2,..., x n, so its dimension is n + 1 (not n). Note that the dimension of V depends on the field K. Thus the complex numbers C can be considered as a vector space of dimension 1 over C, with one possible basis being the single element 1; a vector space of dimension 2 over R, with one possible basis given by the two elements 1, i; a vector space of infinite dimension over Q. The first step towards proving the basis theorem is to be able to remove unnecessary vectors from a spanning set of vectors. Lemma 3.4. Suppose that the vectors v 1, v 2,..., v n, w span V and that w is a linear combination of v 1,..., v n. Then v 1,..., v n span V. Proof. Since v 1, v 2,..., v n, w span V, any vector v V can be written as v = α 1 v α n v n + βw, But w is a linear combination of v 1,..., v n, so w = γ 1 v γ n v n for some scalars γ i, and hence v = (α 1 + βγ 1 )v (α n + βγ n )v n is a linear combination of v 1,..., v n, which therefore span V. There is an important process, which we shall call sifting, which can be applied to any sequence of vectors v 1, v 2,..., v n in a vector space V, as follows. We consider each vector v i in turn. If it is zero, or a linear combination of the preceding vectors v 1,..., v i 1, then we remove it from the list. Example. Let us sift the following sequence of vectors in R 3. v 1 = (0, 0, 0) v 2 = (1, 1, 1) v 3 = (2, 2, 2) v 4 = (1, 0, 0) v 5 = (3, 2, 2) v 6 = (0, 0, 0) v 7 = (1, 1, 0) v 8 = (0, 0, 1) v 1 = 0, so we remove it. v 2 is non-zero so it stays. v 3 = 2v 2 so it is removed. v 4 is clearly not a linear combination of v 2, so it stays. We have to decide next whether v 5 is a linear combination of v 2, v 4. If so, then (3, 2, 2) = α 1 (1, 1, 1) + α 2 (1, 0, 0), which (fairly obviously) has the solution α 1 = 2, α 2 = 1, so remove v 5. Then v 6 = 0 so that is removed too. Next we try v 7 = (1, 1, 0) = α 1 (1, 1, 1) + α 2 (1, 0, 0), and looking at the three components, this reduces to the three equations 1 = α 1 + α 2 ; 1 = α 1 ; 0 = α 1. 8

9 The second and third of these equations contradict each other, and so there is no solution. Hence v 7 is not a linear combination of v 2, v 4, and it stays. Finally, we need to try leading to the three equations v 8 = (0, 0, 1) = α 1 (1, 1, 1) + α 2 (1, 0, 0) + α 3 (1, 1, 0) 0 = α 1 + α 2 + α 3 0 = α 1 + α 3 ; 1 = α 1 and solving these in the normal way, we find a solution α 1 = 1, α 2 = 0, α 3 = 1. Thus we delete v 8 and we are left with just v 2, v 4, v 7. Of course, the vectors that are removed during the sifting process depends very much on the order of the list of vectors. For example, if v 8 had come at the beginning of the list rather than at the end, then we would have kept it. The idea of sifting allows us to prove the following theorem, stating that every finite sequence of vectors which spans a vector space V actually contains a basis for V. Theorem 3.5. Suppose that the vectors v 1,..., v r span the vector space V. Then there is a subsequence of v 1,..., v r which forms a basis of V. Proof. We sift the vectors v 1,..., v r. The vectors that we remove are linear combinations of the preceding vectors, and so by Lemma 3.4, the remaining vectors still span V. After sifting, no vector is zero or a linear combination of the preceding vectors (or it would have been removed), so by Lemma 3.1, the remaining vectors are linearly independent. Hence they form a basis of V. The theorem tells us that any vector space with a finite spanning set is finitedimensional, and indeed the spanning set contains a basis. We now prove the dual result: any linearly independent set is contained in a basis. Theorem 3.6. Let V be a vector space over K which has a finite spanning set, and suppose that the vectors v 1,..., v r are linearly independent in V. Then we can extend the sequence to a basis v 1,..., v n of V, where n r. Proof. Suppose that w 1,..., w q is a spanning set for V. We sift the combined sequence v 1,..., v r, w 1,..., w q. Since w 1,..., w q span V, the whole sequence spans V. Sifting results in a basis for V as in the proof of Theorem 3.5. Since v 1,..., v r are linearly independent, none of them can be a linear combination of the preceding vectors, and hence none of the v i are deleted in the sifting process. Thus the resulting basis contains v 1,..., v r. Example. The vectors v 1 = (1, 2, 0, 2), v 2 = (0, 1, 0, 2) are linearly independent in R 4. Let us extend them to a basis of R 4. The easiest thing is to append the standard basis of R 4, giving the combined list of vectors v 1 = (1, 2, 0, 2), v 2 = (0, 1, 0, 2), w 1 = (1, 0, 0, 0), w 2 = (0, 1, 0, 0), w 3 = (0, 0, 1, 0), w 4 = (0, 0, 0, 1), which we shall sift. We find that (1, 0, 0, 0) = α 1 (1, 2, 0, 2)+α 2 (0, 1, 0, 2) has no solution, so w 1 stays. However, w 2 = v 1 v 2 w 1 so w 2 is deleted. It is clear that w 3 is not a linear combination of v 1, v 2, w 1, because all of those have a 0 in their third component. Hence w 3 remains. Now we have four linearly independent vectors, so must have a basis at this stage, and we can stop the sifting early. The resulting basis is v 1 = (1, 2, 0, 2), v 2 = (0, 1, 0, 2), w 1 = (1, 0, 0, 0), w 3 = (0, 0, 1, 0). We are now ready to prove Theorem 3.3. Since bases of V are both linearly independent and span V, the following proposition implies that any two bases contain the same number of vectors. 9

10 Proposition 3.7 (The exchange lemma). Suppose that vectors v 1,..., v n span V and that vectors w 1,..., w m V are linearly independent. Then m n. Proof. The idea is to place the w i one by one in front of the sequence v 1,..., v n, sifting each time. Since v 1,..., v n span V, w 1, v 1,..., v n are linearly dependent, so when we sift, at least one v j is deleted. We then place w 2 in front of the resulting sequence and sift again. Then we put w 3 in from of the result, and sift again, and carry on doing this for each w i in turn. Since w 1,..., w m are linearly independent none of them are ever deleted. Each time we place a vector in front of a sequence which spans V, and so the extended sequence is linearly dependent, and hence at least one v j gets eliminated each time. But in total, we append m vectors w i, and each time at least one v j is eliminated, so we must have m n. Corollary 3.8. Let V be a vector space of dimension n over K. Then any n vectors which span V form a basis of V, and no n 1 vectors can span V. Proof. After sifting a spanning sequence as in the proof of Theorem 3.5, the remaining vectors form a basis, so by Theorem 3.3, there must be precisely n = dim(v ) vectors remaining. The result is now clear. Corollary 3.9. Let V be a vector space of dimension n over K. Then any n linearly independent vectors form a basis of V and no n + 1 vectors can be linearly independent. Proof. By Theorem 3.6 any linearly independent set is contained in a basis but by Theorem 3.3, there must be precisely n = dim(v ) vectors in the extended set. The result is now clear. 3.4 Existence of a basis Although we have studied bases quite carefully in the previous section, we have not addressed the following fundamental question. Let V be a vector space. Does it contain a basis? Theorem 3.5 gives a partial answer that is good for many practical purposes. Let us formulate it as a corollary. Corollary If a non-trivial vector space V is spanned by a finite number of vectors, then it has a basis. In fact, if we define the idea of an infinite basis carefully, then it can be proved that any vector space has a basis. That result will not be proved in this course. Its proof, which necessarily deals with infinite sets, requires a subtle result in axiomatic set theory called Zorn s lemma. 4 Subspaces Let V be a vector space over the field K. Certain subsets of V have the nice property of being closed under addition and scalar multiplication; that is, adding or taking scalar multiples of vectors in the subset gives vectors which are again in the subset. We call such a subset a subspace: Definition. A subspace of V is a non-empty subset W V such that (i) W is closed under addition: u, v W u + v W ; (ii) W is closed under scalar multiplication: v W, α K αv W. These two conditions can be replaced with a single condition u, v W, α, β K αu + βv W. A subspace W is itself a vector space over K under the operations of vector addition and scalar multiplication in V. Notice that all vector space axioms of W hold automatically. (They are inherited from V.) 10

11 Example. The subset of R 2 given by W = {(α, β) R 2 β = 2α}, that is, the subset consisting of all row vectors whose second entry is twice their first entry, is a subspace of R 2. You can check that adding two vectors of this form always gives another vector of this form; and multiplying a vector of this form by a scalar always gives another vector of this form. For any vector space V, V is always a subspace of itself. Subspaces other than V are sometimes called proper subspaces. We also always have a subspace {0} consisting of the zero vector alone. This is called the trivial subspace, and its dimension is 0, because it has no linearly independent sets of vectors at all. Intersecting two subspaces gives a third subspace: Proposition 4.1. If W 1 and W 2 are subspaces of V then so is W 1 W 2. Proof. Let u, v W 1 W 2 and α K. Then u + v W 1 (because W 1 is a subspace) and u + v W 2 (because W 2 is a subspace). Hence u + v W 1 W 2. Similarly, we get αv W 1 W 2, so W 1 W 2 is a subspace of V. Warning! It is not necessarily true that W 1 W 2 is a subspace, as the following example shows. Example. Let V = R 2, let W 1 = {(α, 0) α R} and W 2 = {(0, α) α R}. Then W 1, W 2 are subspaces of V, but W 1 W 2 is not a subspace, because (1, 0), (0, 1) W 1 W 2, but (1, 0) + (0, 1) = (1, 1) W 1 W 2. Note that any subspace of V that contains W 1 and W 2 has to contain all vectors of the form u + v for u W 1, v W 2. This motivates the following definition. Definition. Let W 1, W 2 be subspaces of the vector space V. Then W 1 + W 2 is defined to be the set of vectors v V such that v = w 1 + w 2 for some w 1 W 1, w 2 W 2. Or, if you prefer, W 1 + W 2 = {w 1 + w 2 w 1 W 1, w 2 W 2 }. Do not confuse W 1 + W 2 with W 1 W 2. Proposition 4.2. If W 1, W 2 are subspaces of V then so is W 1 + W 2. In fact, it is the smallest subspace that contains both W 1 and W 2. Proof. Let u, v W 1 + W 2. Then u = u 1 + u 2 for some u 1 W 1, u 2 W 2 and v = v 1 +v 2 for some v 1 W 1, v 2 W 2. Then u+v = (u 1 +v 1 )+(u 2 +v 2 ) W 1 +W 2. Similarly, if α K then αv = αv 1 + αv 2 W 1 + W 2. Thus W 1 + W 2 is a subspace of V. Any subspace of V that contains both W 1 and W 2 must contain W 1 + W 2, so it is the smallest such subspace. Theorem 4.3. Let V be a finite-dimensional vector space, and let W 1, W 2 be subspaces of V. Then dim(w 1 + W 2 ) = dim(w 1 ) + dim(w 2 ) dim(w 1 W 2 ). Proof. First note that any subspace W of V is finite-dimensional. This follows from Corollary 3.9, because a largest linearly independent subset of W contains at most dim(v ) vectors, and such a subset must be a basis of W. Let dim(w 1 W 2 ) = r and let e 1,..., e r be a basis of W 1 W 2. Then e 1,..., e r is a linearly independent set of vectors, so by Theorem 3.6 it can be extended to a basis e 1,..., e r,f 1,..., f s of W 1 where dim(w 1 ) = r + s, and it can also be extended to a basis e 1,..., e r, g 1,..., g t of W 2, where dim(w 2 ) = r + t. To prove the theorem, we need to show that dim(w 1 + W 2 ) = r + s + t, and to do this, we shall show that e 1,..., e r, f 1,..., f s, g 1,..., g t is a basis of W 1 + W 2. Certainly they all lie in W 1 + W 2. 11

12 First we show that they span W 1 + W 2. Any v W 1 + W 2 is equal to w 1 + w 2 for some w 1 W 1, w 2 W 2. So we can write for some scalars α i, β j K, and w 1 = α 1 e α r e r + β 1 f β s f s for some scalars γ i, δ j K. Then w 2 = γ 1 e γ r e r + δ 1 g δ t g t v = (α 1 + γ 1 )e (α r + γ r )e r + β 1 f β s f s + δ 1 g δ t g t and so e 1,..., e r, f 1,..., f s, g 1,..., g t span W 1 + W 2. Finally we have to show that e 1,..., e r, f 1,..., f s, g 1,..., g t are linearly independent. Suppose that α 1 e α r e r + β 1 f β s f s + δ 1 g δ t g t = 0 for some scalars α i, β j, δ k K. Then α 1 e α r e r + β 1 f β s f s = δ 1 g 1 δ t g t ( ) The left-hand side of this equation lies in W 1 and the right-hand side of this equation lies in W 2. Since the two sides are equal, both must in fact lie in W 1 W 2. Since e 1,..., e r is a basis of W 1 W 2, we can write for some γ i K, and so δ 1 g 1 δ t g t = γ 1 e γ r e r γ 1 e γ r e r + δ 1 g δ t g t = 0. But, e 1,..., e r, g 1,..., g t form a basis of W 2, so they are linearly independent, and hence γ i = 0 for 1 i r and δ i = 0 for 1 i t. But now, from the equation ( ) above, we get α 1 e α r e r + β 1 f β s f s = 0. Now e 1,..., e r, f 1,..., f s form a basis of W 1, so they are linearly independent, and hence α i = 0 for 1 i r and β i = 0 for 1 i s. Thus e 1,..., e r, f 1,..., f s, g 1,..., g t are linearly independent, which completes the proof that they form a basis of W 1 + W 2. Hence dim(w 1 + W 2 ) = r + s + t = (r + s) + (r + t) r = dim(w 1 ) + dim(w 2 ) dim(w 1 W 2 ). Another way to form subspaces is to take linear combinations of some given vectors: Proposition 4.4. Let v 1,..., v n be vectors in the vector space V. Then the set of all linear combinations α 1 v 1 + α 2 v α n v n of v 1,..., v n forms a subspace of V. The proof of this is completely routine and will be omitted. The subspace in this proposition is known as the subspace spanned by v 1,..., v n. Definition. Two subspaces W 1, W 2 of V are called complementary if W 1 W 2 = {0} and W 1 + W 2 = V. Proposition 4.5. Let W 1, W 2 be subspaces of V. Then W 1, W 2 are complementary subspaces if and only if each vector in v V can be written uniquely as v = w 1 + w 2 with w 1 W 1 and w 2 W 2. 12

13 Proof. Suppose first that W 1, W 2 are complementary subspaces and let v V. Then W 1 + W 2 = V, so we can find w 1 W 1 and w 2 W 2 with v = w 1 + w 2. If we also had v = w 1 + w 2 with w 1 W 1, w 2 W 2, then we would have w 1 w 1 = w 2 w 2. The left-hand side lies in W 1 and the right-hand side lies in W 2, and so both sides (being equal) must lie in W 1 W 2 = {0}. Hence both sides are zero, which means w 1 = w 1 and w 2 = w 2, so the expression is unique. Conversely, suppose that every v V can be written uniquely as v = w 1 + w 2 with w 1 W 1 and w 2 W 2. Then certainly W 1 + W 2 = V. If v was a non-zero vector in W 1 W 2, then in fact v would have two distinct expressions as w 1 + w 2 with w 1 W 1 and w 2 W 2, one with w 1 = v, w 2 = 0 and the other with w 1 = 0, w 2 = v. Hence W 1 W 2 = {0}, and W 1 and W 2 are complementary. Examples We give some examples of complementary subspaces. 1. As in the previous example, let V = R 2, W 1 = {(α, 0) α R} and W 2 = {(0, α) α R}. Then W 1 and W 2 are complementary subspaces. 2. Let V = R 3, W 1 = {(α, 0, 0) α R} and W 2 = {(0, α, β) α, β R}. Then W 1 and W 2 are complementary subspaces. 3. Let V = R 2, W 1 = {(α, α) α R} and W 2 = {( α, α) α R}. Then W 1 and W 2 are complementary subspaces. 5 Linear transformations When you study sets, the notion of function is extremely important. There is little to say about a single isolated set, while functions allow you to link different sets. Similarly, in Linear Algebra, a single isolated vector space is not the end of the story. We have to connect different vector spaces by functions. However, a function having little regard to the vector space operations may be of little value. 5.1 Definition and examples Often in mathematics, it is as important to study special classes of functions as it is to study special classes of objects. Often these are functions which preserve certain properties or structures. For example, continuous functions preserve which points are close to which other points. In linear algebra, the functions which preserve the vector space structure are called linear transformations. Definition. Let U, V be two vector spaces over the same field K. A linear transformation or linear map T from U to V is a function T : U V such that (i) T (u 1 + u 2 ) = T (u 1 ) + T (u 2 ) for all u 1, u 2 U; (ii) T (αu) = αt (u) for all α K and u U. Notice that the two conditions for linearity are equivalent to a single condition T (αu 1 + βu 2 ) = αt (u 1 ) + βt (u 2 ) for all u 1, u 2 U, α, β K. First let us state a couple of easy consequences of the definition: Lemma 5.1. Let T : U V be a linear map. Then (i) T (0 U ) = 0 V ; (ii) T ( u) = T (u) for all u U. Proof. For (i), the definition of linear map gives T (0 U ) = T (0 U + 0 U ) = T (0 U ) + T (0 U ), and therefore T (0 U ) = 0 V. For (ii), just put α = 1 in the definition of linear map. 13

14 Examples Many familiar geometrical transformations, such as projections, rotations, reflections and magnifications are linear maps, and the first three examples below are of this kind. Note, however, that a nontrivial translation is not a linear map, because it does not satisfy T (0 U ) = 0 V. 1. Let U = R 3, V = R 2 and define T : U V by T ((α, β, γ)) = (α, β). Then T is a linear map. This type of map is known as a projection, because of the geometrical interpretation. z v y 0 T(v) x (Note: In future we shall just write T (α, β, γ) instead of T ((α, β, γ)).) 2. Let U = V = R 2. We interpret v in R 2 as a directed line vector from 0 to v (see the examples in Section 2), and let T (v) be the vector obtained by rotating v through an angle θ anti-clockwise about the origin. 0 T (v) θ v It is easy to see geometrically that T (u 1 + u 2 ) = T (u 1 ) + T (u 2 ) and T (αu) = αt (u) (because everything is simply rotated about the origin), and so T is a linear map. By considering the unit vectors, we have T (1, 0) = (cos θ, sin θ) and T (0, 1) = ( sin θ, cos θ), and hence T (α, β) = αt (1, 0) + βt (0, 1) = (α cos θ β sin θ, α sin θ + β cos θ). (Exercise: Show this directly.) 3. Let U = V = R 2 again. Now let T (v) be the vector resulting from reflecting v through a line through the origin that makes an angle θ/2 with the x-axis. T (v) v θ/2 0 This is again a linear map. We find that T (1, 0) = (cos θ, sin θ) and T (0, 1) = (sin θ, cos θ), and so T (α, β) = αt (1, 0) + βt (0, 1) = (α cos θ + β sin θ, α sin θ β cos θ). 4. Let U = V = R[x], the set of polynomials over R, and let T be differentiation; i.e. T (p(x)) = p (x) for p R[x]. This is easily seen to be a linear map. 14

15 5. Let U = K[x], the set of polynomials over K. Every α K gives rise to two linear maps, shift S α : U U, S α (f(x)) = f(x α) and evaluation E α : U K, E α (f(x)) = f(α). The next two examples seem dull but are important! 6. For any vector space V, we define the identity map I V : V V by I V (v) = v for all v V. This is a linear map. 7. For any vector spaces U, V over the field K, we define the zero map 0 U,V : U V by 0 U,V (u) = 0 V for all u U. This is also a linear map. One of the most useful properties of linear maps is that, if we know how a linear map U V acts on a basis of U, then we know how it acts on the whole of U. Proposition 5.2 (Linear maps are uniquely determined by their action on a basis). Let U, V be vector spaces over K, let u 1,..., u n be a basis of U and let v 1,..., v n be any sequence of n vectors in V. Then there is a unique linear map T : U V with T (u i ) = v i for 1 i n. Proof. Let u U. Then, since u 1,..., u n is a basis of U, by Proposition 3.2, there exist uniquely determined α 1,..., α n K with u = α 1 u α n u n. Hence, if T exists at all, then we must have T (u) = T (α 1 u α n u n ) = α 1 v α n v n, and so T is uniquely determined. On the other hand, it is routine to check that the map T : U V defined by the above equation is indeed a linear map, so T does exist and is unique. 5.2 Kernels and images To any linear map U V, we can associate a subspace of U and a subspace of V. Definition. Let T : U V be a linear map. The image of T, written as im(t ), is the set of vectors v V such that v = T (u) for some u U. Definition. The kernel of T, written as ker(t ), is the set of vectors u U such that T (u) = 0 V. If you prefer: im(t ) = {T (u) u U}; ker(t ) = {u U T (u) = 0 V }. Examples Let us consider the examples 1 7 above. In example 1, ker(t ) = {(0, 0, γ) γ R}, and im(t ) = R 2. In example 2 and 3, ker(t ) = {0} and im(t ) = R 2. In example 4, ker(t ) is the set of all constant polynomials (i.e. those of degree 0), and im(t ) = R[x]. In example 5, ker(s α ) = {0}, and im(s α ) = K[x], while ker(e α ) is the set of all polynomials divisible by x α, and im(e α ) = K. In example 6, ker(i V ) = {0} and im(t ) = V. In example 7, ker(0 U,V ) = U and im(0 U,V ) = {0}. Proposition 5.3. Let T : U V be a linear map. Then (i) im(t ) is a subspace of V ; (ii) ker(t ) is a subspace of U. 15

16 Proof. For i, we must show that im(t ) is closed under addition and scalar multiplication. Let v 1, v 2 im(t ). Then v 1 = T (u 1 ), v 2 = T (u 2 ) for some u 1, u 2 U. Then and v 1 + v 2 = T (u 1 ) + T (u 2 ) = T (u 1 + u 2 ) im(t ) αv 1 = αt (u 1 ) = T (αu 1 ) im(t ), so im(t ) is a subspace of V. Let us now prove ii. Similarly,we must show that ker(t ) is closed under addition and scalar multiplication. Let u 1, u 2 ker(t ). Then and T (u 1 + u 2 ) = T (0 U + 0 U ) = T (0 U ) = 0 V T (αu 1 ) = αt (u 1 ) = α0 V = 0 V, so u 1 + u 2, αu 1 ker(t ) and ker(t ) is a subspace of U. 5.3 Rank and nullity The dimensions of the kernel and image of a linear map contain important information about it, and are related to each other. Definition. let T : U V be a linear map. (i) dim(im(t )) is called the rank of T ; (ii) dim(ker(t )) is called the nullity of T. Theorem 5.4 (The rank-nullity theorem). Let U, V be vector spaces over K with U finite-dimensional, and let T : U V be a linear map. Then rank(t ) + nullity(t ) = dim(u). Proof. Since U is finite-dimensional and ker(t ) is a subspace of U, ker(t ) is finitedimensional. Let nullity(t ) = s and let e 1,..., e s be a basis of ker(t ). By Theorem 3.6, we can extend e 1,..., e s to a basis e 1,..., e s, f 1,..., f r of U. Then dim(u) = s + r, so to prove the theorem we have to prove that dim(im(t )) = r. Clearly T (e 1 ),..., T (e s ), T (f 1 ),..., T (f r ) span im(t ), and since T (e 1 ) = = T (e s ) = 0 V this implies that T (f 1 ),..., T (f r ) span im(t ). We shall show that T (f 1 ),..., T (f r ) are linearly independent. Suppose that, for some scalars α i, we have α 1 T (f 1 ) + + α r T (f r ) = 0 V. Then T (α 1 f α r f r ) = 0 V, so α 1 f α r f r ker(t ). But e 1,..., e s is a basis of ker(t ), so there exist scalars β i with α 1 f α r f r = β 1 e β s e s = α 1 f α r f r β 1 e 1 β s e s = 0 U. But we know that e 1,..., e s, f 1,..., f r form a basis of U, so they are linearly independent, and hence α 1 = = α r = β 1 = = β s = 0, and we have proved that T (f 1 ),..., T (f r ) are linearly independent. Since T (f 1 ),..., T (f r ) both span im(t ) and are linearly independent, they form a basis of im(u), and hence dim(im(t )) = r, which completes the proof. 16

17 Examples Once again, we consider examples 1 7 above. Since we only want to deal with finite-dimensional spaces, we restrict to an (n + 1)-dimensional space K[x] n in examples 4 and 5, that is, we consider T : R[x] n R[x] n, S α : K[x] n K[x] n, and E α : K[x] n K correspondingly. Let n = dim(u) = dim(v ) in 5 and 6. Example rank(t ) nullity(t ) dim(u) n 1 n S α n n E α 1 n n n 0 n 7 0 n n Corollary 5.5. Let T : U V be a linear map, and suppose that dim(u) = dim(v ) = n. Then the following properties of T are equivalent: (i) T is surjective; (ii) rank(t ) = n; (iii) nullity(t ) = 0; (iv) T is injective; (v) T is bijective; Proof. That T is surjective means precisely that im(t ) = V, so (i) (ii). But if rank(t ) = n, then dim(im(t )) = dim(v ) so (by Corollary 3.9) a basis of im(t ) is a basis of V, and hence im(t ) = V. Thus (ii) (i). That (ii) (iii) follows directly from Theorem 5.4. Now nullity(t ) = 0 means that ker(t ) = {0} so clearly (iv) (iii). On the other hand, if ker(t ) = {0} and T (u 1 ) = T (u 2 ) then T (u 1 u 2 ) = 0, so u 1 u 2 ker(t ) = {0}, which implies u 1 = u 2 and T is injective. Thus (iii) (iv). (In fact, this argument shows that (iii) (iv) is true for any linear map T.) Finally, (v) is equivalent to (i) and (iv), which we have shown are equivalent to each other. Definition. If the conditions in the above corollary are met, then T is called a nonsingular linear map. Otherwise, T is called singular. Notice that the terms singular and non-singular are only used for linear maps T : U V for which U and V have the same dimension. 5.4 Operations on linear maps We can define the operations of addition, scalar multiplication and composition on linear maps. Let T 1 : U V and T 2 : U V be two linear maps, and let α K be a scalar. Definition (Addition of linear maps). We define a map T 1 + T 2 : U V by the rule (T 1 + T 2 )(u) = T 1 (u) + T 2 (u) for u U. Definition (Scalar multiplication of linear maps). We define a map by the rule (αt 1 )(u) = αt 1 (u) for u U. αt 1 : U V Now let T 1 : U V and T 2 : V W be two linear maps. 17

18 Definition (Composition of linear maps). We define a map by (T 2 T 1 )(u) = T 2 (T 1 (u)) for u U. T 2 T 1 : U W In particular, we define T 2 = T T and T i+1 = T i T for i > 2. It is routine to check that T 1 + T 2, αt 1 and T 2 T 1 are themselves all linear maps (but you should do it!). Furthermore, for fixed vector spaces U and V over K, the operations of addition and scalar multiplication on the set Hom K (U, V ) of all linear maps from U to V makes Hom K (U, V ) into a vector space over K. Given a vector space U over a field K, the vector space U = Hom K (U, K) plays a special role. It is often called the dual space or the space of covectors of U. One can think of coordinates as elements of U. Indeed, let e i be a basis of U. Every x U can be uniquely written as x = α 1 e α n e n, α i K. The elements α i depend on x as well as on a choice of the basis, so for each i one can write the coordinate function e i : U K, e i (x) = α i. It is routine to check that e i is a linear map, and indeed the functions e i form a basis of the dual space U. 6 Matrices The material in this section will be familiar to many of you already, at least when K is the field of real numbers. Definition. Let K be a field and m, n N. An m n matrix A over K is an m n rectangular array of numbers (i.e. scalars) in K. The entry in row i and column j is often written α ij. (We use the corresponding Greek letter.) We write A = (α ij ) to make things clear. For example, we could take 2 1 π 0 K = R, m = 3, n = 4, A = (α ij ) = 3 3/2 0 6, and then α 13 = π, α 33 = 10 10, α 34 = 0, and so on. Having defined what matrices are, we want to be able add them, multiply them by scalars, and multiply them by each other. You probably already know how to do this, but we will define these operations anyway. Definition (Addition of matrices). Let A = (α ij ) and B = (β ij ) be two m n matrices over K. We define A + B to be the m n matrix C = (γ ij ), where γ ij = α ij + β ij for all i, j. Example. ( ) ( ) 2 3 = 1 4 ( ) Definition (Scalar multiplication of matrices). Let A = (α ij ) be an m n matrix over K and let β K be a scalar. We define the scalar multiple βa to be the m n matrix C = (γ ij ), where γ ij = βα ij for all i, j. 18

19 Definition (Multiplication of matrices). Let A = (α ij ) be an l m matrix over K and let B = (β ij ) be an m n matrix over K. The product AB is an l n matrix C = (γ ij ) where, for 1 i l and 1 j n, γ ij = m α ik β kj = α i1 β 1j + α i2 β 2j + + α im β mj. k=1 It is essential that the number m of columns of A is equal to the number of rows of B; otherwise AB makes no sense. If you are familiar with scalar products of vectors, note also that γ ij is the scalar product of the ith row of A with the jth column of B. Example. Let A = ( ) , B = Then ( ) ( ) AB = =, BA = ( ) Let C =. Then AC and CA are not defined ( ) ( ) Let D =. Then AD is not defined, but DA = Proposition 6.1. Matrices satisfy the following laws whenever the sums and products involved are defined: (i) A + B = B + A; (ii) (A + B)C = AC + BC; (iii) C(A + B) = CA + CB; (iv) (λa)b = λ(ab) = A(λB); (v) (AB)C = A(BC). Proof. These are all routine checks that the entries of the left-hand sides are equal to the corresponding entries on the right-hand side. Let us do (v) as an example. Let A, B and C be l m, m n and n p matrices, respectively. Then AB = D = (δ ij ) is an l n matrix with δ ij = m s=1 α isβ sj, and BC = E = (ε ij ) is an m p matrix with ε ij = n t=1 β itγ tj. Then (AB)C = DC and A(BC) = AE are both l p matrices, and we have to show that their coefficients are equal. The (i, j)-coefficient of DC is n n m m n m δ it γ tj = ( α is β st )γ tj = α is ( β st γ tj ) = α is ε sj t=1 t=1 s=1 s=1 which is the (i, j)-coefficient of AE. Hence (AB)C = A(BC). t=1 There are some useful matrices to which we give names. Definition. The m n zero matrix 0 mn over any field K has all of its entries equal to 0. Definition. The n n identity matrix I n = (α ij ) over any field K has α ii = 1 for 1 i n, but α ij = 0 when i j. s=1 19

20 Example. I 1 = (1), I 2 = ( ) , I = Note that I n A = A for any n m matrix A and AI n = A for any m n matrix A. The set of all m n matrices over K will be denoted by K m,n. Note that K m,n is itself a vector space over K using the operations of addition and scalar multiplication defined above, and it has dimension mn. (This should be obvious is it?) A 1 n matrix is called a row vector. We will regard K 1,n as being the same as K n. A n 1 matrix is called a column vector. We will denote the the space K n,1 of all column vectors by K n,1. In matrix calculations, we will use K n,1 more often than K n. 7 Linear transformations and matrices We shall see in this section that, for fixed choice of bases, there is a very natural one-one correspondence between linear maps and matrices, such that the operations on linear maps and matrices defined in Chapters 5 and 6 also correspond to each other. This is perhaps the most important idea in linear algebra, because it enables us to deduce properties of matrices from those of linear maps, and vice-versa. It also explains why we multiply matrices in the way we do. 7.1 Setting up the correspondence Let T : U V be a linear map, where dim(u) = n, dim(v ) = m. Suppose that we are given a basis e 1,..., e n of U and a basis f 1,..., f m of V. Now, for 1 j n, the vector T (e j ) lies in V, so T (e j ) can be written uniquely as a linear combination of f 1,..., f m. Let T (e 1 ) = α 11 f 1 + α 21 f α m1 f m T (e 2 ) = α 12 f 1 + α 22 f α m2 f m T (e n ) = α 1n f 1 + α 2n f α mn f m where the coefficients α ij K (for 1 i m, 1 j n) are uniquely determined. Putting it more compactly, we define scalars α ij by T (e j ) = m α ij f i for 1 j n. i=1 The coefficients α ij form an m n matrix α 11 α 12 α 1n α 21 α 22 α 2n A = α m1 α m2 α mn over K. Then A is called the matrix of the linear map T with respect to the chosen bases of U and V. In general, different choice of bases gives different matrices. We shall address this issue later in the course, in Section 11. Notice the role of individual columns in A The jth column of A consists of the coordinates of T (e j ) with respect to the basis vf 1,..., f m of V. Theorem 7.1. Let U, V be vector spaces over K of dimensions n, m, respectively. Then, for a given choice of bases of U and V, there is a one-one correspondence between the set Hom K (U, V ) of linear maps U V and the set K m,n of m n matrices over K. 20

21 Proof. As we saw above, any linear map T : U V determines an m n matrix A over K. Conversely, let A = (α ij ) be an m n matrix over K. Then, by Proposition 5.2, there is just one linear T : U V with T (e j ) = m i=1 α ijf i for 1 j n, so we have a one-one correspondence. Examples Once again, we consider our examples from Section T : R 3 R 2, T (α, β, γ) = (α, β). Usually, we choose the standard bases of K m and K n, which in this case are e 1 = (1, 0, 0), e 2 = (0, 1, 0), e 3 = (0, 0, 1) and f 1 = (1, 0), f 2 = (0, 1). We have T (e 1 ) = f 1, T (e 2 ) = f 2 ), T (e 3 ) = 0, and the matrix is ( ) But suppose we chose different bases, say e 1 = (1, 1, 1), e 2 = (0, 1, 1), e 3 = (1, 0, 1), and f 1 = (0, 1), f 2 = (1, 0). Then we have T (e 1 ) = (1, 1) = f 1 +f 2, T (e 2 ) = (0, 1) = f 1, T (e 3 ) = (1, 0) = f 2, and the matrix is ( ) T : R 2 R 2, T is a rotation through θ anti-clockwise about the origin. We saw that T (1, 0) = (cos θ, sin θ) and T (0, 1) = ( sin θ, cos θ), so the matrix using the standard bases is ( ) cos θ sin θ. sin θ cos θ 3. T : R 2 R 2, T is a reflection through the line through the origin making an angle θ/2 with the x-axis. We saw that T (1, 0) = (cos θ, sin θ) and T (0, 1) = (sin θ, cos θ), so the matrix using the standard bases is ( ) cos θ sin θ. sin θ cos θ 4. This time we take the differentiation map T from R[x] n to R[x] n 1. Then, with respect to the bases 1, x, x 2,..., x n and 1, x, x 2,..., x n 1 of R[x] n and R[x] n 1, respectively, the matrix of T is n n 5. Let S α : K[x] n K[x] n be the shift. With respect to the basis 1, x, x 2,..., x n of K[x] n, we calculate S α (x n ) = (x α) n. The binomial formula gives the matrix of S α, 1 α α 2 ( 1) n α n 0 1 2α ( 1) n 1 nα n ( 1) n 2 n(n 1) 2 α n nα In the same basis of K[x] n and the basis 1 of K, E α (x n ) = α n. The matrix of E α is ( 1 α α2 α n 1 α n). 21

### Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

### T ( a i x i ) = a i T (x i ).

Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)

### MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

### NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

### Inner Product Spaces

Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

### 1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

### 1 Introduction to Matrices

1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

### MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

### Section 4.4 Inner Product Spaces

Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer

### Matrix Representations of Linear Transformations and Changes of Coordinates

Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under

### a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

### BANACH AND HILBERT SPACE REVIEW

BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but

### Vector and Matrix Norms

Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

### Linear Algebra I. Ronald van Luijk, 2012

Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.

### Similarity and Diagonalization. Similar Matrices

MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

### Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

### 4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

### MATH1231 Algebra, 2015 Chapter 7: Linear maps

MATH1231 Algebra, 2015 Chapter 7: Linear maps A/Prof. Daniel Chan School of Mathematics and Statistics University of New South Wales danielc@unsw.edu.au Daniel Chan (UNSW) MATH1231 Algebra 1 / 43 Chapter

### Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

### MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

### Systems of Linear Equations

Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

### THE DIMENSION OF A VECTOR SPACE

THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field

### LINEAR ALGEBRA W W L CHEN

LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,

### 5. Linear algebra I: dimension

5. Linear algebra I: dimension 5.1 Some simple results 5.2 Bases and dimension 5.3 Homomorphisms and dimension 1. Some simple results Several observations should be made. Once stated explicitly, the proofs

### by the matrix A results in a vector which is a reflection of the given

Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

### Notes on Determinant

ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

### LEARNING OBJECTIVES FOR THIS CHAPTER

CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional

### Orthogonal Diagonalization of Symmetric Matrices

MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

### Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

### 1 Sets and Set Notation.

LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most

### 13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms

### Solutions to Math 51 First Exam January 29, 2015

Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not

### v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with

### Math 333 - Practice Exam 2 with Some Solutions

Math 333 - Practice Exam 2 with Some Solutions (Note that the exam will NOT be this long) Definitions (0 points) Let T : V W be a transformation Let A be a square matrix (a) Define T is linear (b) Define

### Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

MAT067 University of California, Davis Winter 2007 Linear Maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007) As we have discussed in the lecture on What is Linear Algebra? one of

### Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone

### I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called

### Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry

### 3. Prime and maximal ideals. 3.1. Definitions and Examples.

COMMUTATIVE ALGEBRA 5 3.1. Definitions and Examples. 3. Prime and maximal ideals Definition. An ideal P in a ring A is called prime if P A and if for every pair x, y of elements in A\P we have xy P. Equivalently,

### Chapter 20. Vector Spaces and Bases

Chapter 20. Vector Spaces and Bases In this course, we have proceeded step-by-step through low-dimensional Linear Algebra. We have looked at lines, planes, hyperplanes, and have seen that there is no limit

### Name: Section Registered In:

Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are

### Methods for Finding Bases

Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,

### MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column

### Introduction to Matrices for Engineers

Introduction to Matrices for Engineers C.T.J. Dodson, School of Mathematics, Manchester Universit 1 What is a Matrix? A matrix is a rectangular arra of elements, usuall numbers, e.g. 1 0-8 4 0-1 1 0 11

### MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar

### Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal

### Unified Lecture # 4 Vectors

Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,

### 3. INNER PRODUCT SPACES

. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.

### Math 4310 Handout - Quotient Vector Spaces

Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable

### Solution to Homework 2

Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if

### Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

### FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES

FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied

PYTHAGOREAN TRIPLES KEITH CONRAD 1. Introduction A Pythagorean triple is a triple of positive integers (a, b, c) where a + b = c. Examples include (3, 4, 5), (5, 1, 13), and (8, 15, 17). Below is an ancient

### MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α

### Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

### Lecture L3 - Vectors, Matrices and Coordinate Transformations

S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between

### Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that

### 160 CHAPTER 4. VECTOR SPACES

160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results

### 4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:

### ( ) which must be a vector

MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are

### MAT188H1S Lec0101 Burbulla

Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u

### University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients

### PROJECTIVE GEOMETRY. b3 course 2003. Nigel Hitchin

PROJECTIVE GEOMETRY b3 course 2003 Nigel Hitchin hitchin@maths.ox.ac.uk 1 1 Introduction This is a course on projective geometry. Probably your idea of geometry in the past has been based on triangles

### Notes on Symmetric Matrices

CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.

### 3. Mathematical Induction

3. MATHEMATICAL INDUCTION 83 3. Mathematical Induction 3.1. First Principle of Mathematical Induction. Let P (n) be a predicate with domain of discourse (over) the natural numbers N = {0, 1,,...}. If (1)

### Lecture 2 Matrix Operations

Lecture 2 Matrix Operations transpose, sum & difference, scalar multiplication matrix multiplication, matrix-vector product matrix inverse 2 1 Matrix transpose transpose of m n matrix A, denoted A T or

### MAT 242 Test 2 SOLUTIONS, FORM T

MAT 242 Test 2 SOLUTIONS, FORM T 5 3 5 3 3 3 3. Let v =, v 5 2 =, v 3 =, and v 5 4 =. 3 3 7 3 a. [ points] The set { v, v 2, v 3, v 4 } is linearly dependent. Find a nontrivial linear combination of these

### 1 Homework 1. [p 0 q i+j +... + p i 1 q j+1 ] + [p i q j ] + [p i+1 q j 1 +... + p i+j q 0 ]

1 Homework 1 (1) Prove the ideal (3,x) is a maximal ideal in Z[x]. SOLUTION: Suppose we expand this ideal by including another generator polynomial, P / (3, x). Write P = n + x Q with n an integer not

### Metric Spaces. Chapter 7. 7.1. Metrics

Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some

### Math 312 Homework 1 Solutions

Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please

### December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation

### Lecture Notes 2: Matrices as Systems of Linear Equations

2: Matrices as Systems of Linear Equations 33A Linear Algebra, Puck Rombach Last updated: April 13, 2016 Systems of Linear Equations Systems of linear equations can represent many things You have probably

### Linear Algebra Review. Vectors

Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length

### Introduction to Matrix Algebra

Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

### Linear Algebra. A vector space (over R) is an ordered quadruple. such that V is a set; 0 V ; and the following eight axioms hold:

Linear Algebra A vector space (over R) is an ordered quadruple (V, 0, α, µ) such that V is a set; 0 V ; and the following eight axioms hold: α : V V V and µ : R V V ; (i) α(α(u, v), w) = α(u, α(v, w)),

### Notes on Linear Algebra. Peter J. Cameron

Notes on Linear Algebra Peter J. Cameron ii Preface Linear algebra has two aspects. Abstractly, it is the study of vector spaces over fields, and their linear maps and bilinear forms. Concretely, it is

### Linear Algebra Notes

Linear Algebra Notes Chapter 19 KERNEL AND IMAGE OF A MATRIX Take an n m matrix a 11 a 12 a 1m a 21 a 22 a 2m a n1 a n2 a nm and think of it as a function A : R m R n The kernel of A is defined as Note

### MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =

MAT 200, Midterm Exam Solution. (0 points total) a. (5 points) Compute the determinant of the matrix 2 2 0 A = 0 3 0 3 0 Answer: det A = 3. The most efficient way is to develop the determinant along the

### Finite dimensional C -algebras

Finite dimensional C -algebras S. Sundar September 14, 2012 Throughout H, K stand for finite dimensional Hilbert spaces. 1 Spectral theorem for self-adjoint opertors Let A B(H) and let {ξ 1, ξ 2,, ξ n

### MATH 551 - APPLIED MATRIX THEORY

MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

### Math 115A - Week 1 Textbook sections: 1.1-1.6 Topics covered: What is a vector? What is a vector space? Span, linear dependence, linear independence

Math 115A - Week 1 Textbook sections: 1.1-1.6 Topics covered: What is Linear algebra? Overview of course What is a vector? What is a vector space? Examples of vector spaces Vector subspaces Span, linear

### Solving Systems of Linear Equations

LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

### Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,

### 9 MATRICES AND TRANSFORMATIONS

9 MATRICES AND TRANSFORMATIONS Chapter 9 Matrices and Transformations Objectives After studying this chapter you should be able to handle matrix (and vector) algebra with confidence, and understand the

### Chapter 6. Orthogonality

6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

### Quotient Rings and Field Extensions

Chapter 5 Quotient Rings and Field Extensions In this chapter we describe a method for producing field extension of a given field. If F is a field, then a field extension is a field K that contains F.

### Geometric Transformations

Geometric Transformations Definitions Def: f is a mapping (function) of a set A into a set B if for every element a of A there exists a unique element b of B that is paired with a; this pairing is denoted

### Math 67: Modern Linear Algebra (UC Davis, Fall 2011) Summary of lectures Prof. Dan Romik

Math 67: Modern Linear Algebra (UC Davis, Fall 2011) Summary of lectures Prof. Dan Romik [Version of November 30, 2011 this document will be updated occasionally with new material] Lecture 1 (9/23/11)

### MATH10040 Chapter 2: Prime and relatively prime numbers

MATH10040 Chapter 2: Prime and relatively prime numbers Recall the basic definition: 1. Prime numbers Definition 1.1. Recall that a positive integer is said to be prime if it has precisely two positive

### 1 Solving LPs: The Simplex Algorithm of George Dantzig

Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

### MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform

MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish

### 16.3 Fredholm Operators

Lectures 16 and 17 16.3 Fredholm Operators A nice way to think about compact operators is to show that set of compact operators is the closure of the set of finite rank operator in operator norm. In this

### Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

### Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

### LS.6 Solution Matrices

LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions

### Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points

Introduction to Algebraic Geometry Bézout s Theorem and Inflection Points 1. The resultant. Let K be a field. Then the polynomial ring K[x] is a unique factorisation domain (UFD). Another example of a

### 4. How many integers between 2004 and 4002 are perfect squares?

5 is 0% of what number? What is the value of + 3 4 + 99 00? (alternating signs) 3 A frog is at the bottom of a well 0 feet deep It climbs up 3 feet every day, but slides back feet each night If it started

### SOLUTIONS TO EXERCISES FOR. MATHEMATICS 205A Part 3. Spaces with special properties

SOLUTIONS TO EXERCISES FOR MATHEMATICS 205A Part 3 Fall 2008 III. Spaces with special properties III.1 : Compact spaces I Problems from Munkres, 26, pp. 170 172 3. Show that a finite union of compact subspaces