Chapter 5: Complex Inner Product Spaces
|
|
- Sophie Warren
- 7 years ago
- Views:
Transcription
1 Chapter 5: Complex Inner Product Spaces Let us recall the inner product for the real) n dimensional Euclidean space R n : for vectors x = x 1,x,...,x n ) and y = y 1,y,...,y n ) in R n, the inner product x,y is defined to be x,y = x 1 y 1 +x y + +x n y n, and the norm or the magnitude) x is given by x = x,x = x 1 +x + +x n. For complex vectors, we cannot copy this definition directly. We need to use complex conjugation to this definition in such a way that x,x so that the definition of magnitude x = x,x still makes sense. Denoteby C n thecomplexvectorspaceconsistingofalln tuples z = z 1,z,...,z n ). The addition and the scalar multiplication of vectors in C n are defined in the usual pointwise fashion: for vectors x = x 1,x,...,x n ) and y = y 1,y,...,y n ) in C n, and scalars a C, x+y = x 1 +y 1, x +y,..., x n +y n ) and ax = ax 1, ax,..., ax n ) The inner product or the scalar product) x,y of vectors x and y is defined via x,y = x 1 y 1 +x y + +x n y n 5.1) Notice that x,x = x 1 x 1 +x x + +x n x n = x 1 + x + + x n, which is what we ask for. The norm of x is given by x = x,x 1/ = x 1 + x + + x n. Remark: In 5.1), it is not clear why we prefer to take complex conjugates of components of y instead of components of x. Actually this is more or less due to the tradition of mathematics, rather than our preference. Physicists have a different tradition!) The space C n provides us with the typical example of complex inner product spaces, defined as follows: 1
2 Definition. By an inner product on a complex vector space we mean a device of assigning to each pair of vectors x and y a complex number denoted by x,y, such that the following conditions are satisfied: C1) x,y, and x,x = if and only if x =. C) y,x = x,y. C3) The inner product is a sesquelinear map, i.e. a 1 x 1 +a x, y = a 1 x 1,y +a x,y x, b 1 y 1 +b y = b 1 x,y 1 +b x,y. Actually the second identity of C3) above is the consequence of the first, together with C). Besides C n, another example of complex inner product space is given as follows. Consider a space F of well-behaved complex-valued functions over an interval, say [a, b]; here we do not specify the technical meaning of being well-behaved). The inner product f, g of f, g F is given by f,g = 1 b a b a ft)gt)dt, for f,g F. On the right hand side, 1/b a) is a normalization factor added for convenience in the future.) The norm induced by this inner product is f f,f 1/ 1 = b a b a ft) dt) 1/ for f F. In the future we will take F to be the space of trigonometric polynomials and [a,b] is any interval of length π, such as [, π] and [ π, π]. Now we return to the general discussion of a complex vector space V with an inner product,. We say that two vectors x and y in V are orthogonal or perpendicular if their inner product is zero and we write x y in this case. Thus, by our definition here, x y x,y =. From the definition of orthogonality you should recognize that, first, the zero vector is orthogonal to every vector; second, is the only vector orthogonal to itself and hence is the only vector orthogonal to every vector; third, x y implies y x.
3 A set of nonzero vectors S is called an orthogonal system if each vector in S is orthogonal to all other vectors in S. If, furthermore, each vector in S has length 1, then S is called an orthonormal system. Notice the difference of the endings of the words orthogonal and orthonormal.) We have the following generalized Pythagoras theorem: If v 1,v,,v n are an orthogonal system, then v 1 +v + +v n = v 1 + v + + v n. 5.) We prove this by induction on n. When n = 1, 5.) becomes v 1 = v 1 and there is nothing to prove. So let us assume that n and assume that the theorem is true for n 1 vectors. Let w = v +v 3 + +v n. Then, by our induction hypothesis, w = n k= v k. Thus 5.) becomes v 1 +w = v 1 + w which remains to be verified. Notice that v 1,w = w,v 1 = v,v 1 + v 3,v v n,v 1 =. Hence v 1 +w = v 1 +w,v 1 +w = v 1,v 1 + v 1,w + w,v 1 + w,w = v 1,v 1 + w,w = v 1 + w. Given an orthonormal system E = {e 1,e,...,e n } in V, and a vector v, which can be written as a linear combination of vectors in B, say v = v 1 e 1 +v e + +v n e n n v ke k. k= 1 we look for an explicit expression for the coefficients v k in this linear combination. By the linearity of the first slot of inner product, we have n v,e j = v ke k, e j = n v k e k,e j. k= 1 k= 1 Note that e k,e j are zeros except when k = j, which gives 1 in this case; in short, e k,e j = δ jk ). Hence we have v,e j = v j. Thus v = v,e 1 e 1 + v,e e + + v,e n e n, 5.3) Since v,e k e k = v,e k e k = v,e k, the generalized Pythagoras theorem gives v = v,e 1 + v,e + + v,e n, 5.4) if v is in the linear span of the orthonormal system E = {e 1,e,...,e n }. 3
4 Next weconsider aslightly moregeneral problem: given a vectorvinan innerproduct space V and a subspace W, spanned by an orthogonal system S = {w 1,w,...,w r } w k,w j = for k j and w k,w k, where k and j run between 1 and r) find the so-called orthogonal decomposition of v: v = w+h, 5.5) where w W and h W. The vector w here will be called the orthogonal) projection of v onto W. Since w is in W and since W is spanned by w 1,w,...,w r, we can write w = a 1 w 1 +a w + +a r w r. 5.6) We have to find a 1, a,..., a r. Identity 5.5) can be rewritten as v = w+h = r k= 1 a kw k +h. Take any vector from w 1,w,...,w r, say w j, and form the inner product of w j with each side of the above identity. By the linearity of the first slot of inner product, we have v,w j = r k= 1 a k w k,w j + h,w j. Note that w k,w j are zeros except when k = j. Hence r k= 1 a k w k,w j can be reduced to a j w j,w j. On the other hand, h,w j = because h is perpendicular to W and w j is in W. Thus we arrive at v,w j = a j w j,w j, or a j = v,w j / w j,w j. Substitute this expression of a j to D1), switching the index j to k, to obtain: w = r k= 1 v, w k w k,w k w k v, w 1 w 1,w 1 w v, w r w r,w r w r, 5.7) which is the required projection. Nowweconsidertwospecialcases: ThefirstcaseisthatS consistsofasinglenonzero) vector, say u. Write down the orthogonal decomposition v = v,u u+h where h u. u,u The generalized Pythagoras theorem gives v = v,u / u,u u + h. Using u,u = u, we rewrite the first term on the right-hand side as v,u / u. Then we show our generosity by dropping the second term h on the right to obtain the inequality v v,u / u. We can rearrange this into v,u v u, 4
5 which is the celebrated Cauchy Schwarz s inequality. The second special case is that S consists of an orthonormal system, say S = {e 1,e,...,e n }. In this case w = n k= 1 v,e k e k with w = n k= 1 v,e k. The orthogonal decomposition v = w +h tells us that v = w + h. Dropping h, we get v w, or w v. We have arrived at n This is called Bessel s inequality. k= 1 v,e k v. In the rest of this chapter, we give some important examples of orthonormal systems. Example 5.1. In C n, the vectors e 1 = 1,,,...,, ) e =, 1,,...,, ). e n =,,,...,, 1) clearly form an orthonormal basis, usually called the standard basis of C n. Example 5.. As usual, write ω = e πi/n. Consider the following vectors in C n : f k = 1 n 1, ω k 1, ω k 1),..., ω n 1)k 1)) ; 1 k n. We write down the first three of them to see the general pattern: f 1 = 1, 1, 1,..., 1)/ n, f = 1, ω, ω,..., ω n 1 )/ n, f 3 = 1, ω, ω 4,..., ω n 1) )/ n, We claim that f k 1 k n) form an orthonormal basis in C n. First we check that they are uni vectors: f k = 1 n 1 + ω k 1 + ω k 1) + + ω n 1)k 1) ) = 1. 5
6 Next we show that, for k l, f k,f l =. For definiteness, let us assume 1 l < k n. By using ω = ω 1, we get f k,f l = 1+ω k 1 ω l 1 +ω k 1) ω l 1) + +ω k 1)n 1) ω l 1)n 1)) /n = 1+ω k l +ω k l + +ω k l)n 1)) /n where η = ω k l. Now = 1+η +η + +η n 1 )/n, 1 η)1+η +η + +η n 1 ) = 1 η n = 1 ω k l)n = 1 ω n ) k l = 1 1 =. Since < k l < n, η ω k l 1, or 1 η. Hence 1+η+η + +η n 1 =. Now f k,f l = is clear. This example will be referred to in the next chapter when we discuss the finite Fourier transform in Example 6.3. Example 5.3. Consider the space of all periodic functions of period π. The inner product of two such functions f and g is defined to be f,g = 1 π π ft)gt)dt. We claim that the system e int < n < ), where n ranges over all integers, is orthonormal. First we check that each of them is of unit length: Next, for n m, we have e int == 1 π π e int dt = 1 π π 1dt = 1. e int,e imt = 1 π π e int e imt dt = 1 π π e in m)t dt =. The orthogonal decomposition of a function f in this space gives its Fourier series: where Bessel s inequality in this case says ft) = c n = f, e int = 1 π <n< <n< π c n 1 π π 6 c n e int, ft)e int dt. ft) dt,
7 showing that the infinite sum on the left hand side is a finite number. This prompts many mathematicians to say: the sequence of Fourier coefficients is square summable. We will present some basic theory of Fourier series in Chapter 8. Example 5.4. Consider the space of all even functions of period π. The inner product of two such functions f and g is defined to be f,g = 1 π π ft)gt)dt. Then the following functions 1, cost, cost, cos3t,... form an orthonormal system of this space. To show this, we need to check π π cosmt cosnt dt = δ mn = { 1 if m = n if m n 5.8) which is left to the reader as an exercise. Example 5.5. Recall that the Chebyshev s polynomials T n x) is defined via the identity cosnt = T n cost). The orthogonality relation 5.8) above gives π π T m cost) T n cost) dt = δ mn Now apply the change of variable x = cost. Notice that cos = 1, cosπ = 1 and dx = sintdt, which gives dt = dx/sint = dx/ 1 cos = dx/ 1 x ; notice that, for t π, we have sint ) Thus we have π 1 1 T m x)t n x) dx 1 x = δ mn. This shows that if we define the inner product of two polynomial functions f and g by f,g = π 1 1 fx)gx) dx 1 x then the Chebyshev s polynomials T n n = 1,,3,...) form an orthonormal system. 7
3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More informationInner product. Definition of inner product
Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationMath 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
More informationSection 4.4 Inner Product Spaces
Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationLecture 14: Section 3.3
Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in
More informationMATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
More informationRecall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationLEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional
More informationVector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
More informationi=(1,0), j=(0,1) in R 2 i=(1,0,0), j=(0,1,0), k=(0,0,1) in R 3 e 1 =(1,0,..,0), e 2 =(0,1,,0),,e n =(0,0,,1) in R n.
Length, norm, magnitude of a vector v=(v 1,,v n ) is v = (v 12 +v 22 + +v n2 ) 1/2. Examples v=(1,1,,1) v =n 1/2. Unit vectors u=v/ v corresponds to directions. Standard unit vectors i=(1,0), j=(0,1) in
More informationSection 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More information1 Inner Products and Norms on Real Vector Spaces
Math 373: Principles Techniques of Applied Mathematics Spring 29 The 2 Inner Product 1 Inner Products Norms on Real Vector Spaces Recall that an inner product on a real vector space V is a function from
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
More informationLinearly Independent Sets and Linearly Dependent Sets
These notes closely follow the presentation of the material given in David C. Lay s textbook Linear Algebra and its Applications (3rd edition). These notes are intended primarily for in-class presentation
More informationProblem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday.
Math 312, Fall 2012 Jerry L. Kazdan Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday. In addition to the problems below, you should also know how to solve
More informationBANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationInner Product Spaces and Normed Spaces
page 59 Chapter 4 Inner Product Spaces and Normed Spaces In the following all vector spaces are assumed to be over the real or complex field. 4.1 Inner Products and Inner Product Spaces Definition 4.1.
More information1. Let P be the space of all polynomials (of one real variable and with real coefficients) with the norm
Uppsala Universitet Matematiska Institutionen Andreas Strömbergsson Prov i matematik Funktionalanalys Kurs: F3B, F4Sy, NVP 005-06-15 Skrivtid: 9 14 Tillåtna hjälpmedel: Manuella skrivdon, Kreyszigs bok
More informationChapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationFactoring Quadratic Expressions
Factoring the trinomial ax 2 + bx + c when a = 1 A trinomial in the form x 2 + bx + c can be factored to equal (x + m)(x + n) when the product of m x n equals c and the sum of m + n equals b. (Note: the
More information4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More information1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
More informationr (t) = 2r(t) + sin t θ (t) = r(t) θ(t) + 1 = 1 1 θ(t) 1 9.4.4 Write the given system in matrix form x = Ax + f ( ) sin(t) x y 1 0 5 z = dy cos(t)
Solutions HW 9.4.2 Write the given system in matrix form x = Ax + f r (t) = 2r(t) + sin t θ (t) = r(t) θ(t) + We write this as ( ) r (t) θ (t) = ( ) ( ) 2 r(t) θ(t) + ( ) sin(t) 9.4.4 Write the given system
More information3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.
Exercise 1 1. Let A be an n n orthogonal matrix. Then prove that (a) the rows of A form an orthonormal basis of R n. (b) the columns of A form an orthonormal basis of R n. (c) for any two vectors x,y R
More informationHOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!
Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following
More informationLS.6 Solution Matrices
LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions
More informationFUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied
More informationCONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation
Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in
More informationSection 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationDIFFERENTIABILITY OF COMPLEX FUNCTIONS. Contents
DIFFERENTIABILITY OF COMPLEX FUNCTIONS Contents 1. Limit definition of a derivative 1 2. Holomorphic functions, the Cauchy-Riemann equations 3 3. Differentiability of real functions 5 4. A sufficient condition
More information1.5. Factorisation. Introduction. Prerequisites. Learning Outcomes. Learning Style
Factorisation 1.5 Introduction In Block 4 we showed the way in which brackets were removed from algebraic expressions. Factorisation, which can be considered as the reverse of this process, is dealt with
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationLEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 6 Woman teaching geometry, from a fourteenth-century edition of Euclid s geometry book. Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition
More informationTensor product of vector spaces
Tensor product of vector spaces Construction Let V,W be vector spaces over K = R or C. Let F denote the vector space freely generated by the set V W and let N F denote the subspace spanned by the elements
More informationLectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain
Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More informationChapter 9. Systems of Linear Equations
Chapter 9. Systems of Linear Equations 9.1. Solve Systems of Linear Equations by Graphing KYOTE Standards: CR 21; CA 13 In this section we discuss how to solve systems of two linear equations in two variables
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More information17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function
17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function, : V V R, which is symmetric, that is u, v = v, u. bilinear, that is linear (in both factors):
More informationMATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform
MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish
More informationAdding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors
1 Chapter 13. VECTORS IN THREE DIMENSIONAL SPACE Let s begin with some names and notation for things: R is the set (collection) of real numbers. We write x R to mean that x is a real number. A real number
More informationIntroduction. Appendix D Mathematical Induction D1
Appendix D Mathematical Induction D D Mathematical Induction Use mathematical induction to prove a formula. Find a sum of powers of integers. Find a formula for a finite sum. Use finite differences to
More informationMATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar
More informationSolutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationBindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
More informationLinear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
More informationLinear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
More informationNotes on metric spaces
Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.
More informationLinear Algebra: Vectors
A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1982, 2008. This chapter originates from material used by author at Imperial College, University of London, between 1981 and 1990. It is available free to all individuals,
More informationLecture 18 - Clifford Algebras and Spin groups
Lecture 18 - Clifford Algebras and Spin groups April 5, 2013 Reference: Lawson and Michelsohn, Spin Geometry. 1 Universal Property If V is a vector space over R or C, let q be any quadratic form, meaning
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More informationChapter 20. Vector Spaces and Bases
Chapter 20. Vector Spaces and Bases In this course, we have proceeded step-by-step through low-dimensional Linear Algebra. We have looked at lines, planes, hyperplanes, and have seen that there is no limit
More informationOrthogonal Projections
Orthogonal Projections and Reflections (with exercises) by D. Klain Version.. Corrections and comments are welcome! Orthogonal Projections Let X,..., X k be a family of linearly independent (column) vectors
More informationFOUNDATIONS OF ALGEBRAIC GEOMETRY CLASS 22
FOUNDATIONS OF ALGEBRAIC GEOMETRY CLASS 22 RAVI VAKIL CONTENTS 1. Discrete valuation rings: Dimension 1 Noetherian regular local rings 1 Last day, we discussed the Zariski tangent space, and saw that it
More informationFinite dimensional C -algebras
Finite dimensional C -algebras S. Sundar September 14, 2012 Throughout H, K stand for finite dimensional Hilbert spaces. 1 Spectral theorem for self-adjoint opertors Let A B(H) and let {ξ 1, ξ 2,, ξ n
More information9 Multiplication of Vectors: The Scalar or Dot Product
Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation
More informationExamination paper for TMA4115 Matematikk 3
Department of Mathematical Sciences Examination paper for TMA45 Matematikk 3 Academic contact during examination: Antoine Julien a, Alexander Schmeding b, Gereon Quick c Phone: a 73 59 77 82, b 40 53 99
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationLinear Algebra. A vector space (over R) is an ordered quadruple. such that V is a set; 0 V ; and the following eight axioms hold:
Linear Algebra A vector space (over R) is an ordered quadruple (V, 0, α, µ) such that V is a set; 0 V ; and the following eight axioms hold: α : V V V and µ : R V V ; (i) α(α(u, v), w) = α(u, α(v, w)),
More informationTHREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
More informationCITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION
No: CITY UNIVERSITY LONDON BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION ENGINEERING MATHEMATICS 2 (resit) EX2005 Date: August
More informationMathematics Review for MS Finance Students
Mathematics Review for MS Finance Students Anthony M. Marino Department of Finance and Business Economics Marshall School of Business Lecture 1: Introductory Material Sets The Real Number System Functions,
More informationDefinitions 1. A factor of integer is an integer that will divide the given integer evenly (with no remainder).
Math 50, Chapter 8 (Page 1 of 20) 8.1 Common Factors Definitions 1. A factor of integer is an integer that will divide the given integer evenly (with no remainder). Find all the factors of a. 44 b. 32
More informationPolynomial Invariants
Polynomial Invariants Dylan Wilson October 9, 2014 (1) Today we will be interested in the following Question 1.1. What are all the possible polynomials in two variables f(x, y) such that f(x, y) = f(y,
More informationOrthogonal Bases and the QR Algorithm
Orthogonal Bases and the QR Algorithm Orthogonal Bases by Peter J Olver University of Minnesota Throughout, we work in the Euclidean vector space V = R n, the space of column vectors with n real entries
More informationOrthogonal Projections and Orthonormal Bases
CS 3, HANDOUT -A, 3 November 04 (adjusted on 7 November 04) Orthogonal Projections and Orthonormal Bases (continuation of Handout 07 of 6 September 04) Definition (Orthogonality, length, unit vectors).
More informationReview Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
More informationT ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
More informationFactoring Trinomials: The ac Method
6.7 Factoring Trinomials: The ac Method 6.7 OBJECTIVES 1. Use the ac test to determine whether a trinomial is factorable over the integers 2. Use the results of the ac test to factor a trinomial 3. For
More informationMatrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
More informationEigenvalues, Eigenvectors, Matrix Factoring, and Principal Components
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they
More informationby the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that
More informationContinued Fractions. Darren C. Collins
Continued Fractions Darren C Collins Abstract In this paper, we discuss continued fractions First, we discuss the definition and notation Second, we discuss the development of the subject throughout history
More informationUnified Lecture # 4 Vectors
Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,
More informationQuotient Rings and Field Extensions
Chapter 5 Quotient Rings and Field Extensions In this chapter we describe a method for producing field extension of a given field. If F is a field, then a field extension is a field K that contains F.
More informationNegative Integer Exponents
7.7 Negative Integer Exponents 7.7 OBJECTIVES. Define the zero exponent 2. Use the definition of a negative exponent to simplify an expression 3. Use the properties of exponents to simplify expressions
More informationSection 1.4. Lines, Planes, and Hyperplanes. The Calculus of Functions of Several Variables
The Calculus of Functions of Several Variables Section 1.4 Lines, Planes, Hyperplanes In this section we will add to our basic geometric understing of R n by studying lines planes. If we do this carefully,
More informationThe Greatest Common Factor; Factoring by Grouping
296 CHAPTER 5 Factoring and Applications 5.1 The Greatest Common Factor; Factoring by Grouping OBJECTIVES 1 Find the greatest common factor of a list of terms. 2 Factor out the greatest common factor.
More informationDifferential Operators and their Adjoint Operators
Differential Operators and their Adjoint Operators Differential Operators inear functions from E n to E m may be described, once bases have been selected in both spaces ordinarily one uses the standard
More informationAu = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
More informationWHEN DOES A CROSS PRODUCT ON R n EXIST?
WHEN DOES A CROSS PRODUCT ON R n EXIST? PETER F. MCLOUGHLIN It is probably safe to say that just about everyone reading this article is familiar with the cross product and the dot product. However, what
More information5 Homogeneous systems
5 Homogeneous systems Definition: A homogeneous (ho-mo-jeen -i-us) system of linear algebraic equations is one in which all the numbers on the right hand side are equal to : a x +... + a n x n =.. a m
More informationChapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm.
Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm. We begin by defining the ring of polynomials with coefficients in a ring R. After some preliminary results, we specialize
More informationVector Spaces. Chapter 2. 2.1 R 2 through R n
Chapter 2 Vector Spaces One of my favorite dictionaries (the one from Oxford) defines a vector as A quantity having direction as well as magnitude, denoted by a line drawn from its original to its final
More informationGROUP ALGEBRAS. ANDREI YAFAEV
GROUP ALGEBRAS. ANDREI YAFAEV We will associate a certain algebra to a finite group and prove that it is semisimple. Then we will apply Wedderburn s theory to its study. Definition 0.1. Let G be a finite
More informationSECTION 10-2 Mathematical Induction
73 0 Sequences and Series 6. Approximate e 0. using the first five terms of the series. Compare this approximation with your calculator evaluation of e 0.. 6. Approximate e 0.5 using the first five terms
More information4.5 Chebyshev Polynomials
230 CHAP. 4 INTERPOLATION AND POLYNOMIAL APPROXIMATION 4.5 Chebyshev Polynomials We now turn our attention to polynomial interpolation for f (x) over [ 1, 1] based on the nodes 1 x 0 < x 1 < < x N 1. Both
More information12. Inner Product Spaces
1. Inner roduct Spaces 1.1. Vector spaces A real vector space is a set of objects that you can do to things ith: you can add to of them together to get another such object, and you can multiply one of
More informationCS3220 Lecture Notes: QR factorization and orthogonal transformations
CS3220 Lecture Notes: QR factorization and orthogonal transformations Steve Marschner Cornell University 11 March 2009 In this lecture I ll talk about orthogonal matrices and their properties, discuss
More information