Section 4.4 Inner Product Spaces


 Delilah Lucinda Cain
 2 years ago
 Views:
Transcription
1 Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer consider vector spaces V over arbitrary fields F; rather, we restrict F to be the field of real or complex numbers. In the first case V is called a real vector space, in the second, a complex vector space. We all have had some experience with real vector spaces in fact both analytic geometry and the subject matter of vector analysis deal with these. What concepts used there can we carry over to a more abstract setting? To begin with we had in these concrete examples the idea of length; secondly we had the idea of perpendicularity, or, more generally, that of angle. These became special cases of the notion of a dot product (often called a scalar or inner product). Let us recall some properties of dot product as it pertained to the special case of the threedimensional real vectors. Given the vectors v = (x 1, x 2, x 3 ) and w = (y 1, y 2, y 3 ), where the x s and y s are real numbers, the dot product of v and w, denoted by v w, was defined as v w = x 1 y 1 + x 2 y 2 + x 3 y 3 Note that the length of v is given by v v and the angle θ between v and w is determined by cos θ = v w v v w w What formal properties does this dot product enjoy? We list a few: (1) v v and v v = if and only if v = ; (2) v w = w v; (3) u (αv + βw) = α(u v) + β(u w); for any vectors u, v, w and real numbers α, β. Everything that has been said can be carried over to complex vector spaces. However, to get geometrically reasonable definitions we must make some modifications. If we simply define v w = x 1 y 1 + x 2 y 2 + x 3 y 3 for v = (x 1, x 2, x 3 ) and w = (y 1, y 2, y 3 ) where the x s and y s are complex numbers, then it is quite possible that v v = with v ; this is illustrated by the vector v = (1, i, ). In fact, v v need not even be real. If, as in the real case, we should want v v to represent somehow the length of v, we should like that this length be real and that a nonzero vector should not have zero length. We can achieve this much by altering the definition of dot product slightly. If α denotes the complex conjugate of the complex number α, returning to the v and w of the paragraph above let us define v w = x 1 y 1 + x 2 y 2 + x 3 y 3. For real vectors this new definition coincides with the old one; on the other hand, for arbitrary complex vectors v, not only is v v real, it is in fact positive. Thus we have the possibility of introducing, in a natural way, a nonnegative length. However, we do lose something; for instance it is no longer true that v w = w v. In fact the exact relationship between these is v w = w v. Let us list a few properties of this dot product: 1
2 (1) v w = w v; (2) v v, and v v = if and only if v = ; (3) (αu + βv) w = α(u w) + β(v w); (4) u (αv + βw) = α(u v) + β(u w); for all complex numbers α, β and all complex vectors u, v, w. We reiterate that in what follows F is either the field of real or complex numbers. DEFINITION. The vector space V over F is said to be an inner product space if there is defined for any two vectors u, v V an element (u, v) in F such that: (1) (u, v) = (v, u); (2) (u, u) and (u, u) = if and only if u = ; (3) (αu + βv, w) = α(u, w) + β(v, w); for any u, v, w V and α, β F. A few observations about properties (1), (2), and (3) are in order. A function satisfying them is called an inner product. If F is the field of complex numbers, property (1) implies that (u, u) is real, and so property (2) makes sense. Using (1) and (3) we see that (u, αv + βw) = (αv + βw, u) = α(v, u) + β(w, u) = α(v, u) + β(w, u) = α(u, v) + β(u, w). We pause to look at some examples of inner product spaces. EXAMPLE 4.4.1: In F (n) define, for u = (α 1,..., α n ) and v = (β 1,..., β n ), This defines an inner product on F (n). (u, v) = α 1 β 1 + α 2 β α n β n. EXAMPLE 4.4.2: In F (2) define for u = (α 1, α 2 ) and v = (β 1, β 2 ), (u, v) = 2α 1 β 1 + α 1 β 2 + α 2 β 1 + α 2 β 2. It is easy to verify that this defines an inner product on F (2). EXAMPLE 4.4.3: Let V be the set of all continuous complexvalued functions on the closed unit interval [, 1]. If f(t), g(t) V, define (f(t), g(t)) = f(t)g(t)dt. We leave it to the reader to verify that this defines an inner product on V. For the remainder of this section V will denote an inner product space. DEFINITION: If v V then the length of v (or norm of v), written as v, is defined by v = (v, v). 2
3 LEMMA 4.4.1: If u, v V and α, β F then (αu + βv, αu + βv) = αα(u, u) + αβ(u, v) + αβ(v, u) + ββ(v, v). Proof: By property (3) defining an inner product space, But and (αu + βv, αu + βv) = α(u, αu + βv) + β(v, αu + βv). (u, αu + βv) = α(u, u) + β(u, v) (v, αu + βv) = α(v, u) + β(v, v). Substituting these in the expression for (αu + βv, αu + βv) we get the desired result. COROLLARY: αu = α u. Proof: We have αu 2 = (αu, αu) = αα(u, u) by Lemma (with v = ). Since αα = α 2 and (u, u) = u 2, taking square roots yields αu = α u. We digress for a moment, and prove a very elementary and familiar result about real quadratic equations. LEMMA 4.4.2: If a, b, c are real numbers such that a > and aλ 2 + 2bλ + c for all real numbers λ, then b 2 ac. Proof: Completing the squares, aλ 2 + 2bλ + c = 1 a (aλ + b)2 + ) (c b2. a Since it is greater than or equal to for all λ, in particular this must be true for λ = b a. Thus c b2 a, and since a > we get b2 ac. We now proceed to an extremely important inequality, usually known as the Schwarz inequality THEOREM 4.4.1: If u, v V then (u, v) u v. Proof: If u = then both (u, v) = and u v =, so that the result is true there. Suppose, for the moment, that (u, v) is real and u. By Lemma 4.4.1, for any real number λ, (λu + v, λu + v) = λ 2 (u, u) + 2(u, v)λ + (v, v). Let a = (u, u), b = (u, v), and c = (v, v); for these the hypothesis of Lemma is satisfied, so that b 2 ac. That is, (u, v) 2 (u, u)(v, v); from this it is immediate that (u, v) u v. 3
4 If α = (u, v) is not real, then it certainly is not, so that u/α is meaningful. Now, ( u ) α, v = 1 1 (u, v) = (u, v) = 1, α (u, v) and so it is certainly real. By the case of the Schwarz inequality discussed in the paragraph above, ( u ) 1 = α, v u v ; α since u = 1 α α u, we get 1 u v α whence α u v. Putting in that α = (u, v) we obtain (u, v) u v, the desired result. Specific cases of the Schwarz inequality are themselves of great interest. We point out two of them. (1) If V = F (n) with (u, v) = α 1 β α n β n, where u = (α 1,..., α n ) and v = (β 1,..., β n ) then Theorem implies that α 1 β α n β n 2 ( α α n 2 )( β β n 2 ). (2) If V is the set of all continuous, complexvalued functions on [, 1] with inner product defined by then Theorem implies that (f(t), g(t)) = f(t)g(t)dt 2 f(t)g(t)dt, f(t) 2 dt g(t) 2 dt. The concept of perpendicularity is an extremely useful and important one in geometry. We introduce its analog in general inner product spaces. DEFINITION: If u, v V then u is said to be orthogonal to v if (u, v) =. Note that if u is orthogonal to v then v is orthogonal to u, for (v, u) = (u, v) = =. DEFINITION. If W is a subspace of V, the orthogonal complement of W, W, is defined by W = {x V (x, w) = for all w W }. LEMMA 4.4.3: W is a subspace of V. 4
5 Proof: If a, b W then for all α, β F and all w W, since a, b W. (αa + βb, w) = α(a, w) + β(b, w) = Note that W W = (), for if w W W it must be selforthogonal, that is (w, w) =. The defining properties of an inner product space rule out this possibility unless w =. One of our goals is to show that V = W + W. Once this is done, the remark made above will become of some interest, for it will imply that V is the direct sum of W and W. DEFINITION. The set of vectors {v i } in V is an orthonormal set if (1) each v i is of length 1 (i.e., (v i, v i ) = 1) (2) for i j, (v i, v j ) =. LEMMA 4.4.4: If {v i } is an orthonormal set, then the vectors in {v i } are linearly independent. If w = α 1 v α n v n, then α i = (w, v i ) for i = 1, 2,..., n. Proof: Suppose that Therefore α 1 v 1 + α 2 v α n v n =. = (α 1 v α n v n, v i ) = α 1 (v 1, v i ) α n (v n, v i ). Since (v j, v i ) = for j i while (v i, v i ) = 1, this equation reduces to α i =. Thus the v j s are linearly independent. If w = α 1 v α n v n then computing as above yields (w, v i ) = α i. Similar in spirit and in proof to Lemma is LEMMA 4.4.5: If {v 1,..., v n } is an orthonormal set in V and if w V, then u = w (w, v 1 )v 1 (w, v 2 )v 2... (w, v i )v i... (w, v n )v n is orthogonal to each of v 1, v 2,...,v n. Proof: Computing (u, v i ) for any i n, using the orthonormality of v 1,...,v n yields the result. The construction carried out in the proof of the next theorem is one which appears and reappears in many parts of mathematics. It is a basic procedure and is known as the Gram Schmidt orthogonalization process. Although we shall be working in a finitedimensional inner product space, the GramSchmidt process works equally well in infinitedimensional situations. THEOREM 4.4.2: Let V be a finitedimensional inner product space; then V has an orthonormal set as a basis. Proof: Let V be of dimension n over F and let v 1,...,v n be a basis of V. From this basis we shall construct an orthonormal set of n vectors; by Lemma this set is linearly independent so must form a basis of V. We proceed with the construction. We seek n vectors w 1,...,w n each of length 1 such that for i j, (w i, w j ) =. In fact we shall finally produce them in the following form: w 1 will 5
6 be a multiple of v 1, w 2 will be in the linear span of w 1 and v 2, w 3 in the linear span of w 1, w 2, and v 3, and more generally, w i in the linear span of w 1, w 2,...,w i, v i. Let w 1 = v 1 v 1, then ( ) v1 (w 1, w 1 ) = v 1, v 1 = 1 v 1 v 1 2(v 1, v 1 ) = 1, whence w 1 = 1. We now ask: for what value of α is αw 1 +v 2 orthogonal to w 1? All we need is that (αw 1 + v 2, w 1 ) =, that is α(w 1, w 1 ) + (v 2, w 1 ) =. Since (w 1, w 1 ) = 1, α = (v 2, w 1 ) will do the trick. Let u 2 = (v 2, w 1 )w 1 + v 2 ; u 2 is orthogonal to w 1 ; since v 1 and v 2 are linearly independent, w 1 and v 2 must be linearly independent, and so u 2. Let w 2 = (u 2 / u 2 ); then {w 1, w 2 } is an orthonormal set. We continue. Let u 3 = (v 3, w 1 )w 1 (v 3, w 2 )w 2 + v 3. A simple check verifies that (u 3, w 1 ) = (u 3, w 2 ) =. Since w 1, w 2, and v 3 are linearly independent (for w 1, w 2 are in the linear span of v 1 and v 2 ), u 3. Let w 3 = (u 3 / u 3 ); then {w 1, w 2, w 3 } is an orthonormal set. The road ahead is now clear. Suppose that we have constructed w 1, w 2,...,w i, in the linear span of v 1,..., v i, which form an orthonormal set. How do we construct the next one, w i+1? Merely put u i+1 = (v i+1, w 1 )w 1 (v i+1, w 2 )w 2... (v i+1, w i )w i + v i+1. That u i+1 and that it is orthogonal to each of w 1,...,w i we leave to the reader. Put w i+1 = (u i+1 / u i+1 ). In this way, given r linearly independent elements in V, we can construct an orthonormal set having r elements. If particular, when dim V = n, from any basis of V we can construct an orthonormal set having n elements. This provides us with the required basis for V. We illustrate the construction used in the last proof in a concrete case. Let F be the real field and let V be the set of polynomials, in a variable x, over F of degree 2 or less. In V we define an inner product by: if p(x), q(x) V, then (p(x), q(x)) = p(x)q(x)dx. Let us start with the basis v 1 = 1, v 2 = x, v 3 = x 2 of V. Following the construction used, w 1 = v 1 v 1 = 1 = 1, u 2 = (v 2, w 1 )w 1 + v 2, 1 2 1dx which after the computations reduces to u 2 = x, and so w 2 = u 2 u 2 = 6 x x 2 dx = 3 2 x;
7 finally, and so u 3 = (v 3, w 1 )w 1 (v 3, w 2 )w 2 + v 3 = 3 + x2, w 3 = u 3 u 3 = 3 + x2 ( ) x2 dx = 1 4 ( + 3x2 ). We mentioned the next theorem earlier as one of our goals. We are now able to prove it. THEOREM 4.4.3: If V is a finitedimensional inner product space and if W is a subspace of V, then V = W + W. More particularly, V is the direct sum of W and W. Proof: Because of the highly geometric nature of the result, and because it is so basic, we give several proofs. The first will make use of Theorem and some of the earlier lemmas. The second will be motivated geometrically. First Proof: As a subspace of the inner product space V, W is itself an inner product space (its inner product being that of V restricted to W). Thus we can find an orthonormal set w 1,...,w r in W which is a basis of W. If v V by Lemma 4.4.5, v = v (v, w 1 )w 1 (v, w 2 )w 2... (v, w r )w r is orthogonal to each of w 1,..., w r and so is orthogonal to W. Thus v W, and since v = v + ((v, w 1 )w (v, w r )w r ) it follows that v W +W. Therefore V = W +W. Since W W = (), this sum is direct. Second Proof: In this proof we shall assume that F is the field of real numbers. The proof works, in almost the same way, for the complex numbers; however, it entails a few extra details which might tend to obscure the essential ideas used. Let v V ; suppose that we could find a vector w W such that v w v w for all w W. We claim that then (v w, w) = for all w W, that is, v w W. If w W, then w + w W, in consequence of which However, the righthand side is (v w, v w ) (v (w + w), v (w + w)). (w, w) + (v w, v w ) 2(v w, w), leading to 2(v w, w) (w, w) for all w W. If m is any positive integer, since w/m W we have that 2 ( m (v w, w) = 2 v w, w ) ( w m m m), w = 1 m2(w, w), and so 2(v w, w) (1/m)(w, w) for any positive integer m. However (1/m)(w, w) as m, whence 2(v w, w). Similarly, w W, and so 2(v w, w) = 2(v w, w), yielding (v w, w) = for all w W. Thus v w W ; hence v w + W W + W. 7
8 To finish the second proof we must prove the existence of a w W such that v w v w for all w W. We indicate sketchily two ways of proving the existence of such a w. Let u 1,..., u k be a basis of W; thus any w W is of the form w = λ 1 u λ k u k. Let β ij = (u i, u j ) and let γ i = (v, u i ) for v V. Thus (v w, v w) = (v λ 1 u 1... λ k u k, v λ 1 w 1... λ k w k ) = (v, v) λ i λ j β ij 2 λ i γ i. This quadratic function in the λ s is nonnegative and so, by results from the calculus, has a minimum. The λ s for this minimum, λ () 1, λ () 2,..., λ () k give us the desired vector w = λ () 1 u λ () k u k in W. A second way of exhibiting such a minimizing w is as follows. In V define a metric ζ by ζ(x, y) = x y ; one shows that ζ is a proper metric on V, and V is now a metric space. Let S = {w W v w v }. In this metric S is a compact set (prove!) and so the continuous function f(w) = v w defined for w S takes on a minimum at some point w S. We leave it to the reader to verify that w is the desired vector satisfying v w v w for all w W. COROLLARY: If V is a finitedimensional inner product space and W is a subspace of V then (W ) = W. Proof: If w W then for any u W, (w, u) =, whence W (W ). Now V = W + W and V = W + (W ) ; from these we get, since the sums are direct, dim(w) = dim((w ) ). Since W (W ) and is of the same dimension as (W ), it follows that W = (W ). 8
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationSec 4.1 Vector Spaces and Subspaces
Sec 4. Vector Spaces and Subspaces Motivation Let S be the set of all solutions to the differential equation y + y =. Let T be the set of all 2 3 matrices with real entries. These two sets share many common
More informationMA106 Linear Algebra lecture notes
MA106 Linear Algebra lecture notes Lecturers: Martin Bright and Daan Krammer Warwick, January 2011 Contents 1 Number systems and fields 3 1.1 Axioms for number systems......................... 3 2 Vector
More information3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus ndimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More informationSection 6.1  Inner Products and Norms
Section 6.1  Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More informationRecall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 34 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
More informationChapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
More information1.3. DOT PRODUCT 19. 6. If θ is the angle (between 0 and π) between two nonzero vectors u and v,
1.3. DOT PRODUCT 19 1.3 Dot Product 1.3.1 Definitions and Properties The dot product is the first way to multiply two vectors. The definition we will give below may appear arbitrary. But it is not. It
More informationInner product. Definition of inner product
Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product
More informationNOTES on LINEAR ALGEBRA 1
School of Economics, Management and Statistics University of Bologna Academic Year 205/6 NOTES on LINEAR ALGEBRA for the students of Stats and Maths This is a modified version of the notes by Prof Laura
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More information17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function
17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function, : V V R, which is symmetric, that is u, v = v, u. bilinear, that is linear (in both factors):
More informationWe call this set an ndimensional parallelogram (with one vertex 0). We also refer to the vectors x 1,..., x n as the edges of P.
Volumes of parallelograms 1 Chapter 8 Volumes of parallelograms In the present short chapter we are going to discuss the elementary geometrical objects which we call parallelograms. These are going to
More informationDiagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions
Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationMATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
More information4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
More informationOrthogonal Projections and Orthonormal Bases
CS 3, HANDOUT A, 3 November 04 (adjusted on 7 November 04) Orthogonal Projections and Orthonormal Bases (continuation of Handout 07 of 6 September 04) Definition (Orthogonality, length, unit vectors).
More informationVectors, Gradient, Divergence and Curl.
Vectors, Gradient, Divergence and Curl. 1 Introduction A vector is determined by its length and direction. They are usually denoted with letters with arrows on the top a or in bold letter a. We will use
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 19967 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationSummary of week 8 (Lectures 22, 23 and 24)
WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry
More informationTHE DIMENSION OF A VECTOR SPACE
THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field
More informationChapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
More information5 =5. Since 5 > 0 Since 4 7 < 0 Since 0 0
a p p e n d i x e ABSOLUTE VALUE ABSOLUTE VALUE E.1 definition. The absolute value or magnitude of a real number a is denoted by a and is defined by { a if a 0 a = a if a
More informationAu = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
More informationWHEN DOES A CROSS PRODUCT ON R n EXIST?
WHEN DOES A CROSS PRODUCT ON R n EXIST? PETER F. MCLOUGHLIN It is probably safe to say that just about everyone reading this article is familiar with the cross product and the dot product. However, what
More information9 Multiplication of Vectors: The Scalar or Dot Product
Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation
More informationMath 4310 Handout  Quotient Vector Spaces
Math 4310 Handout  Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
More informationLinear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University
Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone
More informationLecture 14: Section 3.3
Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in
More informationSection 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
More information1 Inner Products and Norms on Real Vector Spaces
Math 373: Principles Techniques of Applied Mathematics Spring 29 The 2 Inner Product 1 Inner Products Norms on Real Vector Spaces Recall that an inner product on a real vector space V is a function from
More information28 CHAPTER 1. VECTORS AND THE GEOMETRY OF SPACE. v x. u y v z u z v y u y u z. v y v z
28 CHAPTER 1. VECTORS AND THE GEOMETRY OF SPACE 1.4 Cross Product 1.4.1 Definitions The cross product is the second multiplication operation between vectors we will study. The goal behind the definition
More informationT ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
More informationLEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. FiniteDimensional
More informationLEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 6 Woman teaching geometry, from a fourteenthcentury edition of Euclid s geometry book. Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition
More informationSection 1.2. Angles and the Dot Product. The Calculus of Functions of Several Variables
The Calculus of Functions of Several Variables Section 1.2 Angles and the Dot Product Suppose x = (x 1, x 2 ) and y = (y 1, y 2 ) are two vectors in R 2, neither of which is the zero vector 0. Let α and
More information1. Periodic Fourier series. The Fourier expansion of a 2πperiodic function f is:
CONVERGENCE OF FOURIER SERIES 1. Periodic Fourier series. The Fourier expansion of a 2πperiodic function f is: with coefficients given by: a n = 1 π f(x) a 0 2 + a n cos(nx) + b n sin(nx), n 1 f(x) cos(nx)dx
More informationBANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
More information1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
More informationMATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
More informationLecture L3  Vectors, Matrices and Coordinate Transformations
S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3  Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between
More informationChapter 5 Polar Coordinates; Vectors 5.1 Polar coordinates 1. Pole and polar axis
Chapter 5 Polar Coordinates; Vectors 5.1 Polar coordinates 1. Pole and polar axis 2. Polar coordinates A point P in a polar coordinate system is represented by an ordered pair of numbers (r, θ). If r >
More informationMAT 1341: REVIEW II SANGHOON BAEK
MAT 1341: REVIEW II SANGHOON BAEK 1. Projections and Cross Product 1.1. Projections. Definition 1.1. Given a vector u, the rectangular (or perpendicular or orthogonal) components are two vectors u 1 and
More informationLecture 1: Schur s Unitary Triangularization Theorem
Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections
More informationElementary Linear Algebra
Elementary Linear Algebra Kuttler January, Saylor URL: http://wwwsaylororg/courses/ma/ Saylor URL: http://wwwsaylororg/courses/ma/ Contents Some Prerequisite Topics Sets And Set Notation Functions Graphs
More informationLinear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
More information12. Inner Product Spaces
1. Inner roduct Spaces 1.1. Vector spaces A real vector space is a set of objects that you can do to things ith: you can add to of them together to get another such object, and you can multiply one of
More information1 Orthogonal projections and the approximation
Math 1512 Fall 2010 Notes on least squares approximation Given n data points (x 1, y 1 ),..., (x n, y n ), we would like to find the line L, with an equation of the form y = mx + b, which is the best fit
More informationRecall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.
ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the ndimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An ndimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0534405967. Systems of Linear Equations Definition. An ndimensional vector is a row or a column
More informationby the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the yaxis We observe that
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationLinear Algebra. A vector space (over R) is an ordered quadruple. such that V is a set; 0 V ; and the following eight axioms hold:
Linear Algebra A vector space (over R) is an ordered quadruple (V, 0, α, µ) such that V is a set; 0 V ; and the following eight axioms hold: α : V V V and µ : R V V ; (i) α(α(u, v), w) = α(u, α(v, w)),
More informationLectures notes on orthogonal matrices (with exercises) 92.222  Linear Algebra II  Spring 2004 by D. Klain
Lectures notes on orthogonal matrices (with exercises) 92.222  Linear Algebra II  Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n realvalued matrix A is said to be an orthogonal
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationInner Product Spaces. 7.1 Inner Products
7 Inner Product Spaces 71 Inner Products Recall that if z is a complex number, then z denotes the conjugate of z, Re(z) denotes the real part of z, and Im(z) denotes the imaginary part of z By definition,
More informationUnified Lecture # 4 Vectors
Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,
More informationv w is orthogonal to both v and w. the three vectors v, w and v w form a righthanded set of vectors.
3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with
More informationLinear Algebra: Vectors
A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector
More informationTangent and normal lines to conics
4.B. Tangent and normal lines to conics Apollonius work on conics includes a study of tangent and normal lines to these curves. The purpose of this document is to relate his approaches to the modern viewpoints
More informationInner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
More informationThe Notebook Series. The solution of cubic and quartic equations. R.S. Johnson. Professor of Applied Mathematics
The Notebook Series The solution of cubic and quartic equations by R.S. Johnson Professor of Applied Mathematics School of Mathematics & Statistics University of Newcastle upon Tyne R.S.Johnson 006 CONTENTS
More informationVector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a nonempty
More informationa 1 x + a 0 =0. (3) ax 2 + bx + c =0. (4)
ROOTS OF POLYNOMIAL EQUATIONS In this unit we discuss polynomial equations. A polynomial in x of degree n, where n 0 is an integer, is an expression of the form P n (x) =a n x n + a n 1 x n 1 + + a 1 x
More informationCross product and determinants (Sect. 12.4) Two main ways to introduce the cross product
Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product Geometrical definition Properties Expression in components. Definition in components Properties Geometrical expression.
More informationGeometry of Vectors. 1 Cartesian Coordinates. Carlo Tomasi
Geometry of Vectors Carlo Tomasi This note explores the geometric meaning of norm, inner product, orthogonality, and projection for vectors. For vectors in threedimensional space, we also examine the
More informationFUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied
More informationLinear Codes. In the V[n,q] setting, the terms word and vector are interchangeable.
Linear Codes Linear Codes In the V[n,q] setting, an important class of codes are the linear codes, these codes are the ones whose code words form a subvector space of V[n,q]. If the subspace of V[n,q]
More informationNotes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
More informationDot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product
Dot product and vector projections (Sect. 12.3) Two definitions for the dot product. Geometric definition of dot product. Orthogonal vectors. Dot product and orthogonal projections. Properties of the dot
More informationOrthogonal Projections
Orthogonal Projections and Reflections (with exercises) by D. Klain Version.. Corrections and comments are welcome! Orthogonal Projections Let X,..., X k be a family of linearly independent (column) vectors
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationTHREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
More informationIntroduction to finite fields
Introduction to finite fields Topics in Finite Fields (Fall 2013) Rutgers University Swastik Kopparty Last modified: Monday 16 th September, 2013 Welcome to the course on finite fields! This is aimed at
More informationLEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 6 Woman teaching geometry, from a fourteenthcentury edition of Euclid s geometry book. Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
More informationMath 215 HW #6 Solutions
Math 5 HW #6 Solutions Problem 34 Show that x y is orthogonal to x + y if and only if x = y Proof First, suppose x y is orthogonal to x + y Then since x, y = y, x In other words, = x y, x + y = (x y) T
More informationBindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
More informationWHICH LINEARFRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE?
WHICH LINEARFRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE? JOEL H. SHAPIRO Abstract. These notes supplement the discussion of linear fractional mappings presented in a beginning graduate course
More informationMath 333  Practice Exam 2 with Some Solutions
Math 333  Practice Exam 2 with Some Solutions (Note that the exam will NOT be this long) Definitions (0 points) Let T : V W be a transformation Let A be a square matrix (a) Define T is linear (b) Define
More informationLinear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
More informationLinear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)
MAT067 University of California, Davis Winter 2007 Linear Maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007) As we have discussed in the lecture on What is Linear Algebra? one of
More informationInteger roots of quadratic and cubic polynomials with integer coefficients
Integer roots of quadratic and cubic polynomials with integer coefficients Konstantine Zelator Mathematics, Computer Science and Statistics 212 Ben Franklin Hall Bloomsburg University 400 East Second Street
More information4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns
L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows
More informationMechanics 1: Vectors
Mechanics 1: Vectors roadly speaking, mechanical systems will be described by a combination of scalar and vector quantities. scalar is just a (real) number. For example, mass or weight is characterized
More informationTensors on a vector space
APPENDIX B Tensors on a vector space In this Appendix, we gather mathematical definitions and results pertaining to tensors. The purpose is mostly to introduce the modern, geometrical view on tensors,
More informationMethods for Finding Bases
Methods for Finding Bases Bases for the subspaces of a matrix Rowreduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,
More information1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
More informationNotes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 201112 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
More informationQuadratic Equations and Inequalities
MA 134 Lecture Notes August 20, 2012 Introduction The purpose of this lecture is to... Introduction The purpose of this lecture is to... Learn about different types of equations Introduction The purpose
More informationPYTHAGOREAN TRIPLES KEITH CONRAD
PYTHAGOREAN TRIPLES KEITH CONRAD 1. Introduction A Pythagorean triple is a triple of positive integers (a, b, c) where a + b = c. Examples include (3, 4, 5), (5, 1, 13), and (8, 15, 17). Below is an ancient
More informationLinear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices
MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two
More informationMATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all ndimensional column
More informationMATH 240 Fall, Chapter 1: Linear Equations and Matrices
MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS
More information