Finite Dimensional Hilbert Spaces and Linear Inverse Problems
|
|
|
- Derick Potter
- 9 years ago
- Views:
Transcription
1 Finite Dimensional Hilbert Spaces and Linear Inverse Problems ECE 174 Lecture Supplement Spring 2009 Ken Kreutz-Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego Report Number ECE174LSHS-S2009V1.0 Copyright c , All Rights Reserved November 12, 2013 Linear Vector Space. A Vector Space, X, is a collection of vectors, x X, over a field, F, of scalars. Any two vectors x, y X can be added to form x+y X where the operation + of vector addition is associative and commutative. The vector space X must contain an additive identity (the zero vector 0) and, for every vector x, an additive inverse x. The required properties of vector addition are given on page 160 of the textbook by Meyer. 1 The properties of scalar multiplication, αx, of a vector x X by a scalar α F are also given on page 160 of the textbook. The scalars used in this course are either the field of real numbers, F = R, or the field of complex numbers, F = C. 2 Unless stated otherwise, in the remainder of this memo we assume that the scalars are from the field of complex numbers. In this course we essentially consider only finite dimensional vector spaces dim X = n < and give results appropriate for this restriction. 3 Any vector x in an n-dimensional vector 1 Matrix Analysis and Applied Linear Algebra, C.D. Meyer, SIAM, Other fields commonly encountered are the field of rational numbers F = Q, which is used in numerical analysis and approximation theory, and the Galois field F = {0, 1}, which is used in the theory of convolutional codes. 3 More general results applicable to the infinite dimensional case can be found in a variety of advanced textbooks, including Linear Operator Theory in Engineering and Science, A.W. Naylor and G.R. Sell, 1
2 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 2 space can be represented (with respect to an appropriate basis see below) as an n-tuple (n 1 column vector) over the field of scalars, ξ 1 x =. X = F n = C n or R n. ξ n We refer to this as a canonical representation of a finite-dimensional vector. We often assume that vectors in an n-dimensional vector space are canonically represented by n 1 column vectors. For example, the space, R m n of real m n matrices forms an mn-dimensional vector space. An element A R m n can be given a canonical representation as an mn 1 column vector by stacking its columns, stack(a) R mn. As this procedure can obviously be undone, we have R mn R m n. The complex conjugate of a scalar α is denoted by ᾱ. The complex conjugate of a matrix A (respectively a vector x) is denoted by Ā (respectively x) and is defined to be the matrix (vector) of the complex conjugates of the elements of the matrix A (of the vector x). The hermitian conjugate of a matrix A (vector x), denoted by A H (x H ), is defined to be the transpose of the complex conjugate, or equivalently the complex conjugate of the transpose, of the matrix (vector) A H = (Ā)T = (A) T (x H = ( x) T = (x) T ). A matrix is said to be hermitian if A H = A. 4 Linear Vector Subspace. A subset V X is a subspace of a vector space X if it is a vector space in its own right. It is understood that a subspace V inherits the vector addition and scalar multiplication operations from the ambient space X. Given this fact, to determine if a subset V is also a subspace one only needs to check that every linear combination of vectors in V yields a vector in V. 5 This latter property is called the property of closure of the subspace V under linear combinations of vectors in V. If V is a subspace of a vector space X, we call X the parent space or ambient space of V. Given two subsets V and W of vectors, we define their sum by 6 V + W = {v + w v V and w W}. If the sets V and W are in addition both subspaces of X, then V W and V + W are also subspaces of X, and V W V + W. 7 In general, {0} V W V + W X, Springer-Verlag, 1982, which is available in a relatively inexpensive paperback edition. Some of the necessary extensions are briefly mentioned as we proceed through our development. 4 Note that a real hermitian matrix is symmetric, A T = A. 5 Two important special cases of this test are that if the zero element does not belong to the subset or if multiplication of a vector by the scalar 1 violates closure then the subset is obviously not a subset. 6 Note the difference between sum and set union. In order for the sum to exist, the elements of the two sets must have a well-defined operation of addition, whereas the union exists whether or not the elements can be added. 7 In general V W is not a subspace.
3 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 3 where {0} is the trivial subspace consisting only of the zero vector (additive identity) of X. Linear Independence and Dimension. By definition r vectors x 1,, x r X are linearly independent when, α 1 x 1 + α r x r = 0 if and only if α 1 = α r = 0. Note that this definition can be written in matrix-vector form as, 8 α 1 Xα = ( ) x 1 x r. α r = 0 α = 0. Thus x 1,, x r are linearly independent iff the associated matrix X = (x 1 x r ) has full column rank (equivalently, iff the null space of X is trivial). If x 1,, x r are not linearly independent, then at least one of them can be written as a linear combination of the remaining vectors. In this case we say that the collection of vectors is linearly dependent. The span of the collection x 1,, x r X is the set of all linear combinations of the vectors, Span {x 1,, x r } = {y y = α 1 x α r x r = Xα, α F r } X. The subset V = Span {x 1,, x r } is a vector subspace of X. If, in addition, the spanning vectors x 1,, x r are linearly independent we say that the collection is a linearly independent spanning set or a basis for the subspace V. 9 We denote a basis for V by B V = {x 1,, x r }. Given a basis for a vector space or subspace, the number of basis vectors in the basis is unique. For a given space or subspace, there are many different bases, but they must all have the same number of vectors. This number, then, is an intrinsic property of the space itself and is called the dimension d = dim V of the space or subspace V. If the number of elements, d, in a basis is finite, we say that the space is finite dimensional, otherwise we say that the space is infinite dimensional. The dimension of the trivial subspace is zero, 0 = dim{0}. If V is a subspace of X, V X, we have dim V dim X. In general for two arbitrary subspaces V and W of X we have, dim (V + W) = dim V + dim W dim (V W), and 0 dim (V W) dim (V + W) dim X. 8 Assuming that x i F n, the resulting matrix X = (x 1 x r ) is n r. 9 Because of the assumption that we are working with finite dimensional spaces, we can ignore the distinction between a Hamel basis and other types of of bases. The interested reader can find an excellent discussion of this matter in Sections 4.6, 4.7, and 5.17 of the textbook by Naylor and Sell, op. cit.
4 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 4 Furthermore, if X = V + W then, with equality if and only if V W = {0}. dim X dim V + dim W, Linear algebra is the study of linear mappings between finite dimensional vector spaces. The study of linear mappings between infinite dimensional vector spaces is known as Linear Functional Analysis or Linear Operator Theory, and is the subject matter of courses which are usually taught in graduate school. 10 Independent Subspaces and Projection Operators. Two subspaces, V and W, of a vector space X are independent or disjoint when V W = {0}. In this case we have dim (V + W) = dim V + dim W. If X = V + W for two independent subspaces V and W we say that V and W are companion subspaces and we write, X = V W. In this case dim X = dim V + dim W. For two companion subspaces V and W any vector x X can be written uniquely as x = v + w, v V and w W. The unique component v is called the projection of x onto V along its companion space W. Similarly, the unique component w is called the projection of x onto W along its companion space V. Given the unique decomposition of a vector x along two companion subspaces V and W, x = v + w, we define the companion projection operators P V W and P W V by, Obviously, P V W x v and P W V x = w. P V W + P W V = I. Furthermore it is straightforward to show that P V W and P W V are idempotent, P 2 V W = P V W and P 2 W V = P W V. In addition, it can be shown that the projection operators P V W and P W V are linear operators. 10 E.g., see the textbook by Naylor and Sell, op. cit.
5 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 5 Linear Operators and Matrices. Consider a function A which maps between two vector spaces X and Y, A : X Y. The input space X is called the domain while the output space is called the codomain. The mapping or operator A is said to be linear when, A (α 1 x 1 + α 2 x 2 ) = α 1 Ax 1 + α 2 Ax 2 x 1, x 2 X, α 1, α 2 F. Note that in order for this definition to be well-posed the vector spaces X and Y both must have the same field of scalars F. 11 It is well-known that any linear operator between finite dimensional vectors spaces has a matrix representation. In particular if n = dim X < and m = dim Y < for two vector spaces over the field F, then a linear operator A which maps between these two spaces has an m n matrix representation over the field F. Note in particular that projection operators on finite-dimensional vector spaces must have matrix representations. Often, for convenience, we assume that any such linear mapping A is an m n matrix and we write A F m n. 12 Every linear operator has two natural vector subspaces associated with it. The range space, R(A) A(X ) {y y = Ax, x X } Y, and the nullspace, N (A) = {x Ax = 0} X. Given our assumption that all vector spaces are finite dimensional, both R(A) and N (A) are closed subspaces. 13 The dimension of the range space is called the rank of A, r(a) = rank(a) = dim R(A), while the dimension of the nullspace is called the nullity of A, ν(a) = nullity(a) = dim N (A), When solving a system of linear equations y = Ax a solution exists when y R(A), in which case we say that the problem is consistent. 11 For example, X and Y must be both real vectors spaces, or must be both complex vector spaces. 12 Thus if X and Y are both complex we write A C m n while if they are both real we write A R m n. 13 The definition of a closed set is given in Section of Moon and Stirling, op. cit. Conditions for closure of the range and null spaces of a linear operator can be found in Naylor and Sell, op. cit. If we take the operator A to be bounded when the vector spaces are infinite dimension, a property which is always satisfied in the finite dimensional situation, then N (A) is closed, while R(A) in general is not closed. However, every finite dimensional subspace is closed, so that if R(A) is finite dimensional, then it is closed. Sufficient conditions for R(A) = A(X ) Y to be finite dimensional are that either X or Y be finite dimensional. Thus the results in this note are also valid for the case when A is bounded and either X or Y are finite dimensional. A sufficient condition for this to be true is that X and Y are both finite dimensional; which is precisely the condition assumed at the outset of this note.
6 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 6 Linear Forward and Inverse Problem. Given a linear mapping between two vector spaces A : X Y the problem of computing an output y in the codomain given an input vector x in the domain, y = Ax, is called the forward problem. Note that the forward problem is always well-posed in that knowing A and given x one can construct y by straightforward matrix-vector multiplication. Given am m-dimensional vector y in the codomain, the problem of determining an n- dimensional vector x in the domain for which Ax = y is known as an inverse problem. The inverse problem is said to be well-posed if and only if the following three conditions are true for the linear mapping A: 1. y R(A) for all y Y so that a solution exists for all y. I.e., we demand that A be onto, R(A) = Y. 14 Equivalently, r(a) = m. If this condition is met, we say that the system y = Ax is robustly consistent, else the system is generically inconsistent. 2. If a solution exists, we demand that it be unique. I.e., that A is one-to-one, N (A) = {0}. Equivalently, ν(a) = The solution x does not depend sensitively on the value of y. 15 I.e., we demand that A be numerically well-conditioned. If any of these three conditions is violated we say that the inverse problem is ill-posed. Condition three is studied in great depth in courses on Numerical Linear Algebra. In this course, we ignore the numerical conditioning problem and focus on the first two conditions only. In particular, we will generalize the concept of solution by looking for a minimum-norm least-squares solution which will exist even when the first two conditions are violated. Normed Linear Vector Space, Banach Space. In a vector space it is useful to have a meaningful measure of size, distance, and neighborhood. The existence of a norm allows these concepts to be well-defined. A norm on a vector space X is a mapping from X to the the nonnegative real numbers which obeys the following three properties: 1. is homogeneous, αx = α x for all α F and x X, 2. is positive-definite, x 0 for all x X and x = 0 iff x = 0, and 3. satisfies the triangle-inequality, x + y x + y for all x, y X. The existence of a norm gives us a measure of size, size(x) = x, a measure of distance, d(x, y) = x y, and a well-defined ɛ-ball or neighborhood of a point x, N ɛ (x) = {y y x ɛ}. There are innumerable norms that one can define on a given vector space, 14 It it not enough to merely require consistency for a given y because even the tiniest error or misspecification in y can render the problem inconsistent. 15 So that a slight error or miss-specification in y does not cause a huge error in the solution x. If such a sensitive dependency exists, we say that A is numerically ill-conditioned
7 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 7 the most commonly used ones being the 1-norm, the 2-norm, and the -norm. In this course we focuss on the weighted 2-norm, x = x H Ωx, where the weighting matrix Ω is hermitian and positive-definite. A Banach Space is a complete normed linear vector space. Completeness is a technical condition which is the requirement that every so-called Cauchy convergent sequence is a convergent sequence. As this property is automatically guaranteed to be satisfied for every finite-dimensional normed linear vector space, it is not discussed in courses on Linear Algebra. Suffice it to say that the finite dimensional spaces normed-vector spaces considered in this course are perforce Banach Spaces. Inner Product Space, Hilbert Space. It is of great convenience, both conceptually and computationally, to work in a space which has well-defined concept of orthogonality, and of angle in general. For this reason, when possible, one attempts to define an inner product on the vector space of interest and then work with the associated norm induced by this inner product. Given a vector space X over the field of scalars F, an inner product is an F-valued binary operator on X X,, : X X F; {x, y} x, y F, x, y X. The inner product has the following three properties: 1. Linearity in the second argument Real positive-definiteness of x, x for all x X Conjugate-symmetry, x, y = y, x for all x, y X. 18 Given an inner product, one can construct the associated induced norm, x = x, x, as the right-hand side of the above can be shown to satisfy all the properties demanded of a norm. It is this norm that one uses in an inner product space. If the resulting normed 16 Although the developments in the Meyer textbook, this supplement, and the class lecture assume linearity in the second argument, this choice is arbitrary and many books take the inner product to alternatively be linear in the first argument. This distinction is only meaningful when F = C and is irrelevant when F = R, in which case the inner product is linear in both arguments (i.e., in the real case the inner product is bilinear). When the vector space is complex, most mathematicians (but not all, including Meyer) and many engineers (but not this class) tend to define linearity in the first argument, while most physicists and controls engineers tend to define linearity in the second argument. Serious confusion can occur if you do not take care to determine which definition is the case. 17 I.e., 0 x, x R for any vector x, and 0 = x, x iff x = which for a real vector space reduces to the property of symmetry, x, y = y, x for all x, y X.
8 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 8 vector space is a Banach space, one calls the inner product space a Hilbert space. All finitedimensional inner product spaces are Hilbert spaces, so the distinction between a general inner product space and a Hilbert space will not be an issue in this course. On a finite n-dimensional Hilbert space, a general inner product is given by the weighted inner product, x 1, x 2 = x H 1 Ωx 2, where the weighting matrix 19 Ω is hermitian and positive-definite. Note that the corresponding induced norm is the weighted 2-norm mentioned above, x = x H Ωx. When Ω = I we call the resulting inner product and induced norm the standard (or Cartesian) inner-product and the standard (or Cartesian) 2-norm respectively. Given an inner product, one then can define orthogonality of two vectors x y by the requirement that x, y = 0. Given two orthogonal vectors, it is straightforward to show that the induced norm satisfies the (generalized) Pythagorean Theorem x y x, y = 0 x + y 2 = x 2 + y 2. More generally, an inner-product induced norm satisfies the Parallelogram Law x + y 2 + x y 2 = 2 x y 2. If the vectors x and y are orthogonal, x, y = 0, then x + y 2 = x y 2 and the Parallelogram Law reduces to the Pythagorean Theorem. Note that the Parallelogram Law implies the useful inner-product induced norm inequality x + y 2 2 ( x 2 + y 2). This inequality and the Parallelogram Law do not hold for the norm on a general (non-inner product) normed vector space. 20 Another important result is that the inner-product induced norm satisfies the Cauchy- Schwarz (C-S) inequality, x, y x y for all x, y X, with equality holding if and only if y = αx for some scalar α. As a consequence of the C-S inequality, one can meaningfully define the angle θ between two vectors in a Hilbert space by x, y cos θ x y 19 Also known as the metric tensor in physics. 20 It is a remarkable fact that if a norm satisfies the Parallelogram Law, then the norm must be a norm induced by a unique inner product and that inner product can be inferred from the norm. See the Wikipedia article Parallelogram Law.
9 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V1.0 9 since as a consequence of the C-S inequality we must have 0 cos θ 1. Note that this definition is consistent with our earlier definition of orthogonality between two vectors. Two Hilbert subspaces are said to be orthogonal subspaces, V W if and only if every vector in V is orthogonal to every vector in W. If V W it must be the case that V are disjoint W, V W = {0}. Given a subspace V of X, one defines the orthogonal complement V to be all vectors in X which are perpendicular to V. The orthogonal complement (in the finite dimensional case assumed here) obeys the property V = V. The orthogonal complement V is unique and is a subspace in its own right for which 21 X = V V. Thus a subspace V and its orthogonal complement V are complementary subspaces. 22 Note that it must therefore be the case that dim X = dim V + dim V. In a Hilbert space the projection onto a subspace V along its (unique) orthogonal complement V is an orthogonal projection operator, denoted by P V = P V V. Note that for an orthogonal projection operator the complementary subspace does not have to be explicitly denoted. Furthermore if the subspace V is understood to be the case, one usually more simply denotes the orthogonal projection operator by P = P V. Of course, as is the case for all projection operators, an orthogonal projection operator is idempotent. Recall our discussion above on the range and null spaces of a linear mapping A : X Y. If X and Y are finite-dimensional Hilbert spaces, we must have that 23 and Y = R(A) R(A) X = N (A) N (A). The subspaces R(A), R(A), N (A), and N (A) are called the Four Fundamental Subspaces of the linear operator A. Projection Theorem, Orthogonality Principle. Suppose we are given a vector x in a Hilbert space X and are asked to find the best approximation, v, to x in a subspace V in the sense that the norm of the error e = x v, e = x v, is to be minimized over all possible vectors v V. We call the resulting optimal vector v the least-squares estimate of 21 Some modification of the subsequent discussion is required in the case of infinite-dimensional Hilbert spaces. See Naylor and Sell, op. cit. The modifications required here are V = V and X = V V where V denotes the closure of the subspace V. 22 Thus V is more than a complementary subspace to V; it is the orthogonally complementary subspace to V. 23 In this case, assuming the boundedness of A, the required infinite dimensional modification is Y = R(A) R(A). The second expression is unchanged.
10 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V x in V, because it is understood that in a Hilbert space minimizing the (induced norm) of the error is equivalent to minimizing the squared-error e 2 = e, e. Let v 0 = P V x be the orthogonal projection of x onto V. Note that P V x = (I P V )x = x P V x = x v 0 must be orthogonal to V. For any vector v V we have e 2 = x v 2 = (x v 0 ) + (v 0 v) 2 = x v v 0 v 2 x v 0 2, as a consequence of the Pythagorean theorem. 24 Thus the error is minimized when v = v 0. Because v 0 is the orthogonal projection of x onto V, the least-squares optimality of v 0 is known as the Projection Theorem, v 0 = P V x. Alternatively, recognizing that the optimal error must be orthogonal to V, (x v 0 ) V, this result is also known as the Orthogonality Principle, x v 0, v = 0 for all v V. We can obtain a generalized solution to an ill-posed inverse problem Ax = y by looking for the unique solution to the regularized problem, min x y Ax 2 + β x 2, β > 0. for appropriately chosen norms on the domain and codomain. The solution to this problem, ˆx β, is a function of the regularization parameter β. The choice of the precise value of the regularization parameter β is often a nontrivial problem. When the problem is posed as a mapping between Hilbert spaces (so that induced norms are used on the domain and codomain) the limiting solution, ˆx = lim β 0 ˆx β, is called the minimum norm least-squares solution. 25 It is also known as the pseudoinverse solution and the operator A + which maps y to this solution, ˆx = A + y is called the pseudoinverse of A. The pseudoinverse A + is itself a linear operator. Note that in the special case when A is square and full-rank, it must be the case that A + = A 1 showing that the pseudoinverse is a generalized inverse. The pseudoinverse solution, ˆx, is the unique least-squares solution to the linear inverse problem having minimum norm among all least-squares solutions to the least squares problem of minimizing e 2 = y Ax 2, { ˆx = arg min x x x arg min x y Ax 2 } 24 Note that the vector v v 0 must be in the subspace V. 25 Often this is called the weighted-minimum norm, weighted-least-squares solution as we generally assume weighted inner products (and hence weighted induced norms) on the domain and codomain spaces. This is because the standard terminology is to take minimum norm least-squares solution to mean that only standard (Cartesian or un-weighted) inner products hold on the domain and codomain. As we admit the richer class of weighted norms, our terminology has a richer meaning..
11 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V Because Ax R(A) we see that any particular least-squares solution, x, to the inverse problem y = Ax yields a value ŷ = Ax which is the unique least-squares approximation of y in R(A) Y, ŷ = P R(A) y. As discussed above, the Orthogonality Condition determines a least-squares solution ŷ = Ax from the geometric condition Geometric Condition for a Least-Squares Solution: e = y Ax R(A) (1) which can be equivalently written as e = y Ax R(A). One can write any particular least-squares solution, x arg min x y Ax 2, ŷ = Ax, as x = (P N (A) + P N (A) ) x = P N (A) x + P N (A) x = ˆx + x null, where ˆx = P N (A) x N (A) and the null space component x null homogeneous solution: Ax null = 0. Note that = P N (A) x N (A) is a ŷ = Ax = A (ˆx + x null) = Aˆx + Ax null = Aˆx. ˆx is unique, i.e., independent of the particular choice of least-squares solution x. leastsquares solution ˆx N (A) is the unique minimum norm least-squares solution. This is true because the Pythagorean theorem yields x 2 = ˆx 2 + x null 2 ˆx 2, showing that unique least squares solution ˆx is indeed the minimum norm least-squares solution. Thus the geometric condition that a least-squares solution x is also a minimum norm solution is that x N (A), or equivalently that Geometric Condition for a Minimum Norm LS Solution: x N (A) (2) The primary geometric condition (1) ensures that x is a least-squares solution. The additional requirement that x satisfy the secondary geometric condition (2) then ensures that x is the unique minimum norm least-squares solution. We now want to make the move from the insightful geometric conditions (1) and (2) to equivalent algebraic expressions which allow us to analytically solve for the minimum norm least-squares solution ˆx. To accomplish this we need to introduce the concept of the adjoint operator A. Adjoint Operator, Four Fundamental Subspaces of a Linear Operator. Given a linear operator A : X Y which maps between two finite dimensional Hilbert spaces, its adjoint operator A : Y X is defined by the condition, y, Ax = A y, x for all x X, y Y.
12 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V If X has a weighted inner-product with weighting matrix Ω and Y has a weighted innerproduct with weighting matrix W, it is shown in class lecture that the adjoint operator can be determined to be unique and is given by the linear operator A = Ω 1 A H W. The adjoint A is a companion operator associated with the operator A. 26 Note that if the standard inner product is used on both the domain and codomain we have A = A H and if furthermore the Hilbert spaces are real we have A = A T. Thus the deeper meaning of the transpose of a real matrix A is that it forms the adjoint ( companion ) operator to A when A is viewed as a linear mapping between two real Hilbert spaces each having the standard inner product. It is discussed in class lecture that the adjoint operator gives a description of R(A) and N (A) which is symmetrical to that of N (A) and R(A), 27 R(A) = N (A ) and N (A) = R(A ). It can be shown that A = A showing that the four fundamental subspaces of A are (with an appropriate renaming) identical to those of A, and thus the geometric relationships between A and A are entirely symmetrical. We can view the operator A and its adjoint A (or, equivalently, the operator A and its adjoint A = A) as being companion matrices which have associated with them the four fundamental subspaces R(A), R(A ), N (A), and N (A ) which are related as A : X Y, A : Y X y, Ax = A y, x, x, A y = Ax, y Y = R(A) N (A ), R(A) = N (A ), X = R(A ) N (A) R(A ) = N (A) Again, note that the relationships between A and A, and their associated four fundamental subspaces, are entirely symmetrical. 28 In class we discuss the fact that if P is an orthogonal projection operator then, in addition to being idempotent, it must also be self-adjoint, P = P. An operator is an orthogonal projection operator if and only if it is both idempotent and self-adjoint. 26 In the infinite dimensional case, the adjoint of a bounded linear operator is itself a bounded linear operator. 27 In the infinite dimensional case, taking A to be bounded, the second expression is modified to N (A) = R(A ). 28 When generalized to the general infinite-dimensional case, but still assuming that A is a bounded linear operator, these relations must be modified by replacing the range spaces by their closures. However, as discussed in an earlier footnote, as long as A is bounded and either X or Y is finite dimensional then the shown relationships remain true. Every linear operator A encountered in the homework and exam problems given in this course will satisfy this latter condition.
13 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V The Minimum Norm Least-Squares Solution and the Pseudoinverse. Combining the geometric conditions (1) and (2) for optimality of a minimum least-squares solution with the characterization given above of the four fundamental subspaces provided by an operator A and its adjoint A we determine the conditions for optimality to be e = y Ax N (A ) and x R(A ). The first optimality condition is equivalent to 0 = A e = A (y Ax) yielding the so-called normal equations, A Ax = A y. (3) The normal equations are a consistent set of equations, every solution of which constitutes a least-squares solution to the inverse problem y = Ax. The second optimality condition is equivalent to, x = A λ, λ Y. (4) The vector λ is a nuisance parameter (which can be interpreted as a Lagrange multiplier) which is usually solved for as an intermediate solution on the way to determining the optimal minimum norm least-square solution ˆx. The two sets of equations (3) and (4) can be solved simultaneously to determine both λ and the optimal solution ˆx. In principle, solving for the pseudoinverse solution ˆx corresponds to determining the pseudoinverse operator A + which yields the solution as ˆx = A + y. The pseudoinverse A + is unique and linear. Although it is not always numerically advisable to determine the pseudoinverse A + itself, special-case closed form expressions for A + and general numerical procedures for constructing A + do exist. In Matlab the command pinv(a) numerically constructs the pseudoinverse matrix for A assuming that the standard inner product holds on the domain and codomain of A. As shown in class, closed form expressions exist for A + for the following two special cases: rank(a) = n : A + = (A A) 1 A, rank(a) = m : A + = A (AA ) 1. In both these cases if A is square one obtains A + = A 1. Generally it is not numerically sound to construct A + using these closed form expressions, although they are quite useful for analysis purposes. From knowledge of the pseudoinverse A +, one obtains the orthogonal projection operators onto the subspaces R(A) and R(A ) respectively as P R(A) = AA + and P R(A ) = A + A. As a consequence, we also have that the orthogonal projection operators onto the subspaces N (A ) and N (A) are respectively given by P N (A ) = I AA + and P N (A) = I A + A.
14 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V One way to identify the pseudoinverse A + is that a candidate matrix M is the unique pseudoinverse for A if and only if it satisfies the four Moore-Penrose (M-P) pseudoinverse conditions: (MA) = MA, (AM) = AM, AMA = A, MAM = M. For instance, using these conditions one can ascertain that the pseudoinverse of the 1 1 (scalar) matrix A = 0 has pseudoinverse A + = 0 + = 0. As discussed below, an alternative way to obtain the pseudoinverse when standard (unweighted) inner products are assumed on the domain and codomain is to obtain the Singular Value Decomposition (SVD) of A, which in Matlab is obtained by the simple function call svd(a). Completing the Square. An alternative, non-geometric way to determine a weighted least-squares solution to a full column-rank linear inverse problem is by completing the square. Consider the quadratic form l(x) = x H Πx 2Re x H By + y H W y (5) where Π is an n n hermitian, positive-definite matrix, Π = Π H > 0, W is an m m hermitian, positive-definite matrix and B is n m. With the positive-definite assumptions on Π and W, the quadratic form l(x) is a real-valued function of the complex vector x C n. With some minor algebra, one can rewrite the quadratic form as Thus for all x we have that l(x) = ( x Π 1 By ) H Π ( x Π 1 By ) + y H ( W B H Π 1 B ) y. (6) l(x) y H ( W B H Π 1 B ) y with equality if and only if x = Π 1 By. Thus we have proved that ˆx = Π 1 By = arg min x l(x). (7) It is straightforward to apply this result to the full column-rank, weighted least-squares problem. Noting that l(x) = y Ax 2 W = (y Ax) H W (y Ax) = x H A H W Ax x H A H W y y H W Ax + y H W y = x H Πx x H By y H B H x + y H W y (Setting Π A H W A and B A H W ) = x H Πx 2Re x H By + y H W y = ( x Π 1 By ) H Π ( x Π 1 By ) + y H ( W B H Π 1 B ) y it is evident that the weighted least-squares estimate of x is given by ˆx = Π 1 By = ( A H W A ) 1 A H W y.
15 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V Use of Derivatives. A third method of finding the least-squares solution to a full columnrank linear inverse problem is provided by taking the derivative of the the weighted leastsquares loss function and setting the resulting expression to zero. Solving the resulting equation then yields the weighted least-squares solution, which is unique under the full column-rank assumption. This method is discussed in the course lectures and in a lecture Supplement. The Singular Value Decomposition (SVD). Given a complex m n matrix A, we can view it as a linear mapping between the unweighted finite dimensional Hilbert spaces X = C n and Y = C m, A : X Y. Note that with the use of the unweighted (standard) inner products the adjoint of A is given by its hermitian transpose, A = A H. The rank of the matrix A is denoted by r = r(a), which is the dimension of the range of A, dim (R(A)) = r and the dimension of the range of the adjoint of A, dim ( R(A H ) ) = r. The nullity ν = n r of the matrix A denotes the dimension of the nullspace of A, dim (N (A)) = ν, and µ = m r denotes the dimension of the null space of the adjoint of A, dim ( N (A H ) ) = µ (this latter nullspace is also called the left-nullspace of A.) The Singular Value Decomposition (SVD) of A is given by the factorization 29 A = UΣV H = ( ) ( ) ( ) S 0 V H r U 1 U 2 = U SV1 H = σ i u i vi H (8) where U is m m, V is n n, Σ is m n, S is r r, U 1 is m r, U 2 is m µ, V 1 is n r, V 2 is n ν, u i denotes a column of U, and v i denotes a column of V. There are precisely r (nonzero) real singular values of A rank ordered as and 1 V H 2 σ 1 σ r > 0, S = diag (σ 1 σ r ). The columns of U form an orthonormal basis for Y = C m, u i, u j = u H i u j = δ i,j, i, j = 1, m. Equivalently, U is a unitary matrix, U H = U 1. The r columns of U 1 form an orthonormal basis for R(A) and the µ columns of U 2 form an orthonormal basis for N (A H ). The columns of V form an orthonormal basis for X = C n, v i, v j = v H i v j = δ i,j, i, j = 1, n, or, equivalently, V H = V 1. The r columns of V 1 form an orthonormal basis for R(A H ) and the ν columns of V 2 form an orthonormal basis for N (A). With the SVD at hand the Moore-Penrose pseudoinverse can be determined as r A + = V 1 S 1 U1 H 1 = v i u H i. (9) σ i 29 See page 412 of the textbook by Meyer or the useful survey paper The Singular Value Decomposition: Its Computation and Some Applications, V.C. Klema & A.J. Laub, IEEE Transactions on Automatic Control, Vol. AC-25, No. 2, pp , April i=1 i=1
16 K. Kreutz-Delgado Copyright c , All Rights Reserved Ver. ECE174LSHS-S2009V Note that (9) does indeed obeys the four Moore-Penrose necessary and sufficient conditions for A + to be the pseudoinverse of A, (AA + ) H = AA +, (A + A) H = A + A, AA + A = A, A + AA + = A + which can be used as a check that one has correctly computed the pseudoinverse. The various projection operators can be determined as P R(A) = U 1 U H 1 = AA + = I P N (A H ), P N (A H ) = U 2 U H 2 = I P R(A), P R(A H ) = V 1 V1 H = A + A = I P N (A), P N (A) = V 2 V2 H = I P R(A H ). Because it is possible to construct the projection operators in more than one way, one can test for correctness by doing so and checking for consistency of the answers. This is also another way to determine that the pseudoinverse has been correctly determined. The columns of U are the m eigenvectors of (AA H ) and are known as the left singular vectors of A. Specifically, the columns of U 1 are the r eigenvectors of (AA H ) having associated nonzero eigenvalues σ 2 i > 0, i = 1, r, while the columns of U 2 are the remaining µ eigenvectors of (AA H ), which have zero eigenvalues. The n columns of V are the eigenvectors of (A H A) and are known as the right singular vectors of A. Specifically, the columns of V 1 are the r eigenvectors of (A H A) having associated nonzero eigenvalues σi 2 > 0, i = 1, r, while the ν columns of V 2 are the eigenvectors of (A H A), which have zero eigenvalues.
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
Linear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka [email protected] http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
BANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
NOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
Inner product. Definition of inner product
Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
Section 4.4 Inner Product Spaces
Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied
Orthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
Numerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 [email protected] 1 Course G63.2010.001 / G22.2420-001, Fall 2010 September 30th, 2010 A. Donev (Courant Institute)
Inner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
Methods for Finding Bases
Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
Linear Algebraic Equations, SVD, and the Pseudo-Inverse
Linear Algebraic Equations, SVD, and the Pseudo-Inverse Philip N. Sabes October, 21 1 A Little Background 1.1 Singular values and matrix inversion For non-smmetric matrices, the eigenvalues and singular
Lecture 5: Singular Value Decomposition SVD (1)
EEM3L1: Numerical and Analytical Techniques Lecture 5: Singular Value Decomposition SVD (1) EE3L1, slide 1, Version 4: 25-Sep-02 Motivation for SVD (1) SVD = Singular Value Decomposition Consider the system
5. Orthogonal matrices
L Vandenberghe EE133A (Spring 2016) 5 Orthogonal matrices matrices with orthonormal columns orthogonal matrices tall matrices with orthonormal columns complex matrices with orthonormal columns 5-1 Orthonormal
1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column
Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms
1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
LINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
Linear Algebra: Vectors
A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector
Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University
Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone
T ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain
Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal
Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
CS3220 Lecture Notes: QR factorization and orthogonal transformations
CS3220 Lecture Notes: QR factorization and orthogonal transformations Steve Marschner Cornell University 11 March 2009 In this lecture I ll talk about orthogonal matrices and their properties, discuss
3 Orthogonal Vectors and Matrices
3 Orthogonal Vectors and Matrices The linear algebra portion of this course focuses on three matrix factorizations: QR factorization, singular valued decomposition (SVD), and LU factorization The first
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
Chapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
Continued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
Matrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
Chapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
Applied Linear Algebra I Review page 1
Applied Linear Algebra Review 1 I. Determinants A. Definition of a determinant 1. Using sum a. Permutations i. Sign of a permutation ii. Cycle 2. Uniqueness of the determinant function in terms of properties
October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix
Linear Algebra & Properties of the Covariance Matrix October 3rd, 2012 Estimation of r and C Let rn 1, rn, t..., rn T be the historical return rates on the n th asset. rn 1 rṇ 2 r n =. r T n n = 1, 2,...,
α = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
Dot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product
Dot product and vector projections (Sect. 12.3) Two definitions for the dot product. Geometric definition of dot product. Orthogonal vectors. Dot product and orthogonal projections. Properties of the dot
Inner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
by the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that
Linear Algebra: Determinants, Inverses, Rank
D Linear Algebra: Determinants, Inverses, Rank D 1 Appendix D: LINEAR ALGEBRA: DETERMINANTS, INVERSES, RANK TABLE OF CONTENTS Page D.1. Introduction D 3 D.2. Determinants D 3 D.2.1. Some Properties of
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University
Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
1. Let P be the space of all polynomials (of one real variable and with real coefficients) with the norm
Uppsala Universitet Matematiska Institutionen Andreas Strömbergsson Prov i matematik Funktionalanalys Kurs: F3B, F4Sy, NVP 005-06-15 Skrivtid: 9 14 Tillåtna hjälpmedel: Manuella skrivdon, Kreyszigs bok
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
Solutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
Linear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
Orthogonal Projections and Orthonormal Bases
CS 3, HANDOUT -A, 3 November 04 (adjusted on 7 November 04) Orthogonal Projections and Orthonormal Bases (continuation of Handout 07 of 6 September 04) Definition (Orthogonality, length, unit vectors).
Finite dimensional C -algebras
Finite dimensional C -algebras S. Sundar September 14, 2012 Throughout H, K stand for finite dimensional Hilbert spaces. 1 Spectral theorem for self-adjoint opertors Let A B(H) and let {ξ 1, ξ 2,, ξ n
Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)
MAT067 University of California, Davis Winter 2007 Linear Maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007) As we have discussed in the lecture on What is Linear Algebra? one of
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
Recall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
Lecture L3 - Vectors, Matrices and Coordinate Transformations
S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between
Systems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices
MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two
MATH 551 - APPLIED MATRIX THEORY
MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
Unified Lecture # 4 Vectors
Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,
Lecture 1: Schur s Unitary Triangularization Theorem
Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections
MA106 Linear Algebra lecture notes
MA106 Linear Algebra lecture notes Lecturers: Martin Bright and Daan Krammer Warwick, January 2011 Contents 1 Number systems and fields 3 1.1 Axioms for number systems......................... 3 2 Vector
A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form
Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...
IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction
IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL R. DRNOVŠEK, T. KOŠIR Dedicated to Prof. Heydar Radjavi on the occasion of his seventieth birthday. Abstract. Let S be an irreducible
1 2 3 1 1 2 x = + x 2 + x 4 1 0 1
(d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which
THREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
Math 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function
17. Inner product spaces Definition 17.1. Let V be a real vector space. An inner product on V is a function, : V V R, which is symmetric, that is u, v = v, u. bilinear, that is linear (in both factors):
MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =
MAT 200, Midterm Exam Solution. (0 points total) a. (5 points) Compute the determinant of the matrix 2 2 0 A = 0 3 0 3 0 Answer: det A = 3. The most efficient way is to develop the determinant along the
LINEAR ALGEBRA. September 23, 2010
LINEAR ALGEBRA September 3, 00 Contents 0. LU-decomposition.................................... 0. Inverses and Transposes................................. 0.3 Column Spaces and NullSpaces.............................
Notes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
LS.6 Solution Matrices
LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions
Math 550 Notes. Chapter 7. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010
Math 550 Notes Chapter 7 Jesse Crawford Department of Mathematics Tarleton State University Fall 2010 (Tarleton State University) Math 550 Chapter 7 Fall 2010 1 / 34 Outline 1 Self-Adjoint and Normal Operators
MATH2210 Notebook 1 Fall Semester 2016/2017. 1 MATH2210 Notebook 1 3. 1.1 Solving Systems of Linear Equations... 3
MATH0 Notebook Fall Semester 06/07 prepared by Professor Jenny Baglivo c Copyright 009 07 by Jenny A. Baglivo. All Rights Reserved. Contents MATH0 Notebook 3. Solving Systems of Linear Equations........................
( ) which must be a vector
MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are
Name: Section Registered In:
Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
FUNCTIONAL ANALYSIS LECTURE NOTES CHAPTER 2. OPERATORS ON HILBERT SPACES
FUNCTIONAL ANALYSIS LECTURE NOTES CHAPTER 2. OPERATORS ON HILBERT SPACES CHRISTOPHER HEIL 1. Elementary Properties and Examples First recall the basic definitions regarding operators. Definition 1.1 (Continuous
Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.
ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?
Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product
Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product Geometrical definition Properties Expression in components. Definition in components Properties Geometrical expression.
Math 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
Moving Least Squares Approximation
Chapter 7 Moving Least Squares Approimation An alternative to radial basis function interpolation and approimation is the so-called moving least squares method. As we will see below, in this method the
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Examination paper for TMA4205 Numerical Linear Algebra
Department of Mathematical Sciences Examination paper for TMA4205 Numerical Linear Algebra Academic contact during examination: Markus Grasmair Phone: 97580435 Examination date: December 16, 2015 Examination
THE DIMENSION OF A VECTOR SPACE
THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field
