Outline. Scientific Computing: An Introductory Survey. Eigenvalue Problems. Eigenvalues and Eigenvectors. Geometric Interpretation

Size: px
Start display at page:

Download "Outline. Scientific Computing: An Introductory Survey. Eigenvalue Problems. Eigenvalues and Eigenvectors. Geometric Interpretation"

Transcription

1 Outline Scientific Computing: An Introductory Survey Chapter 4 Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c Reproduction permitted for noncommercial, educational use only. 2 3 Michael T. Heath Scientific Computing / 87 Michael T. Heath Scientific Computing 2 / 87 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues are also important in analyzing numerical methods Theory and algorithms apply to complex matrices as well as real matrices With complex matrices, we use conjugate transpose, A H, instead of usual transpose, A T Standard eigenvalue problem : Given n n matrix A, find scalar λ and nonzero vector x such that A x λ x λ is eigenvalue, and x is corresponding eigenvector λ may be complex even if A is real Spectrum λ(a) set of eigenvalues of A Spectral radius ρ(a) max{ λ : λ λ(a)} Michael T. Heath Scientific Computing 3 / 87 Michael T. Heath Scientific Computing 4 / 87 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors Examples: Eigenvalues and Eigenvectors Matrix expands or shrinks any vector lying in direction of eigenvector by scalar factor Expansion or contraction factor is given by corresponding eigenvalue λ Eigenvalues and eigenvectors decompose complicated behavior of general linear transformation into simpler actions 0 0 A : λ 0 2, x, λ 0 2 2, x 2 A : λ 0 2, x, λ 0 2 2, x 2 3 A : λ 3 2, x, λ 2 4, x A : λ , x, λ 2, x 2 0 i A : λ 0 i, x, λ i 2 i, x 2 where i Michael T. Heath Scientific Computing 5 / 87 Michael T. Heath Scientific Computing 6 / 87 Example: Equation Ax λx is equivalent to (A λi)x 0 which has nonzero solution x if, and only if, its matrix is singular Eigenvalues of A are roots λ i of characteristic polynomial det(a λi) 0 in λ of degree n Fundamental Theorem of Algebra implies that n n matrix A always has n eigenvalues, but they may not be real nor distinct Complex eigenvalues of real matrix occur in complex conjugate pairs: if α + iβ is eigenvalue of real matrix, then so is α iβ, where i Michael T. Heath Scientific Computing 7 / 87 Characteristic polynomial of previous example matrix is ( ) 3 0 det λ 3 0 () 3 λ det 3 λ (3 λ)(3 λ) ( )( ) λ 2 6λ so eigenvalues are given by λ 6 ± 36 32, or λ 2, λ Michael T. Heath Scientific Computing 8 / 87

2 Companion Matrix, continued Monic polynomial p(λ) c 0 + c λ + + c n λ n + λ n is characteristic polynomial of companion matrix c c C n 0 0 c c n Roots of polynomial of degree > 4 cannot always computed in finite number of steps So in general, computation of eigenvalues of matrices of order > 4 requires (theoretically infinite) iterative process Michael T. Heath Scientific Computing 9 / 87 Example: Computing eigenvalues using characteristic polynomial is not recommended because of work in computing coefficients of characteristic polynomial sensitivity of coefficients of characteristic polynomial work in solving for roots of characteristic polynomial Characteristic polynomial is powerful theoretical tool but usually not useful computationally Multiplicity and Diagonalizability Michael T. Heath Scientific Computing 0 / 87 Consider A ɛ ɛ where ɛ is positive number slightly smaller than ɛ mach Exact eigenvalues of A are + ɛ and ɛ Computing characteristic polynomial in floating-point arithmetic, we obtain det(a λi) λ 2 2λ + ( ɛ 2 ) λ 2 2λ + which has as double root Thus, eigenvalues cannot be resolved by this method even though they are distinct in working precision Michael T. Heath Scientific Computing / 87 Eigenspaces and Invariant Subspaces Eigenvectors can be scaled arbitrarily: if Ax λx, then A(γx) λ(γx) for any scalar γ, so γx is also eigenvector corresponding to λ Eigenvectors are usually normalized by requiring some norm of eigenvector to be Eigenspace S λ {x : Ax λx} Multiplicity is number of times root appears when polynomial is written as product of linear factors Eigenvalue of multiplicity is simple Defective matrix has eigenvalue of multiplicity k > with fewer than k linearly independent corresponding eigenvectors Nondefective matrix A has n linearly independent eigenvectors, so it is diagonalizable X AX D where X is nonsingular matrix of eigenvectors Michael T. Heath Scientific Computing 2 / 87 Properties of matrix A relevant to eigenvalue problems Property Definition diagonal a ij 0 for i j tridiagonal a ij 0 for i j > triangular a ij 0 for i > j (upper) a ij 0 for i < j (lower) Hessenberg a ij 0 for i > j + (upper) a ij 0 for i < j (lower) Subspace S of R n (or C n ) is invariant if AS S For eigenvectors x x p, span([x x p ]) is invariant subspace orthogonal unitary symmetric Hermitian normal A T A AA T I A H A AA H I A A T A A H A H A AA H Examples: Matrix Properties Michael T. Heath Scientific Computing 3 / 87 T 2 3 Transpose: [ + i + 2i Conjugate transpose: 2 i 2 2i 2 Symmetric: Nonsymmetric: i Hermitian: NonHermitian: i 2 + i + i 2 ] H [ i ] 2 + i 2i 2 + 2i Examples, continued Michael T. Heath Scientific Computing 4 / Orthogonal:,, 0 0 [ ] i 2/2 2/2 Unitary: 2/2 i 2/2 Nonorthogonal: Normal: Nonnormal: 0 [ ] 2/2 2/2 2/2 2/2 Michael T. Heath Scientific Computing 5 / 87 Michael T. Heath Scientific Computing 6 / 87

3 Properties of of Properties of eigenvalue problem affecting choice of algorithm and software Are all eigenvalues needed, or only a few? Are only eigenvalues needed, or are corresponding eigenvectors also needed? Is matrix real or complex? Is matrix relatively small and dense, or large and sparse? Condition of eigenvalue problem is sensitivity of eigenvalues and eigenvectors to changes in matrix of eigenvalue problem is not same as conditioning of solution to linear system for same matrix Different eigenvalues and eigenvectors are not necessarily equally sensitive to perturbations in matrix Does matrix have any special properties, such as symmetry, or is it general matrix? of Eigenvalues Michael T. Heath Scientific Computing 7 / 87 of Eigenvalues Michael T. Heath Scientific Computing 8 / 87 If µ is eigenvalue of perturbation A + E of nondefective matrix A, then µ λ k cond 2 (X) E 2 where λ k is closest eigenvalue of A to µ and X is nonsingular matrix of eigenvectors of A Absolute condition number of eigenvalues is condition number of matrix of eigenvectors with respect to solving linear equations Eigenvalues may be sensitive if eigenvectors are nearly linearly dependent (i.e., matrix is nearly defective) For normal matrix (A H A AA H ), eigenvectors are orthogonal, so eigenvalues are well-conditioned Michael T. Heath Scientific Computing 9 / 87 If (A + E)(x + x) (λ + λ)(x + x), where λ is simple eigenvalue of A, then λ y 2 x 2 y H E 2 x cos(θ) E 2 where x and y are corresponding right and left eigenvectors and θ is angle between them For symmetric or Hermitian matrix, right and left eigenvectors are same, so cos(θ) and eigenvalues are inherently well-conditioned Eigenvalues of nonnormal matrices may be sensitive For multiple or closely clustered eigenvalues, corresponding eigenvectors may be sensitive Michael T. Heath Scientific Computing 20 / 87 Similarity Transformation Shift : If Ax λx and σ is any scalar, then (A σi)x (λ σ)x, so eigenvalues of shifted matrix are shifted eigenvalues of original matrix Inversion : If A is nonsingular and Ax λx with x 0, then λ 0 and A x (/λ)x, so eigenvalues of inverse are reciprocals of eigenvalues of original matrix Powers : If Ax λx, then A k x λ k x, so eigenvalues of power of matrix are same power of eigenvalues of original matrix Polynomial : If Ax λx and p(t) is polynomial, then p(a)x p(λ)x, so eigenvalues of polynomial in matrix are values of polynomial evaluated at eigenvalues of original matrix Michael T. Heath Scientific Computing 2 / 87 B is similar to A if there is nonsingular matrix T such that Then B T A T By λy T AT y λy A(T y) λ(t y) so A and B have same eigenvalues, and if y is eigenvector of B, then x T y is eigenvector of A Similarity transformations preserve eigenvalues and eigenvectors are easily recovered Michael T. Heath Scientific Computing 22 / 87 Example: Similarity Transformation Diagonal Form From eigenvalues and eigenvectors for previous example, and hence So original matrix is similar to diagonal matrix, and eigenvectors form columns of similarity transformation matrix Eigenvalues of diagonal matrix are diagonal entries, and eigenvectors are columns of identity matrix Diagonal form is desirable in simplifying eigenvalue problems for general matrices by similarity transformations But not all matrices are diagonalizable by similarity transformation Closest one can get, in general, is Jordan form, which is nearly diagonal but may have some nonzero entries on first superdiagonal, corresponding to one or more multiple eigenvalues Michael T. Heath Scientific Computing 23 / 87 Michael T. Heath Scientific Computing 24 / 87

4 Triangular Form Block Triangular Form Any matrix can be transformed into triangular (Schur ) form by similarity, and eigenvalues of triangular matrix are diagonal entries Eigenvectors of triangular matrix less obvious, but still straightforward to compute If U u U 3 A λi 0 0 v T O 0 U 33 is triangular, then U y u can be solved for y, so that y x 0 is corresponding eigenvector Forms Attainable by Similarity Michael T. Heath Scientific Computing 25 / 87 A T B distinct eigenvalues nonsingular diagonal real symmetric orthogonal real diagonal complex Hermitian unitary real diagonal normal unitary diagonal arbitrary real orthogonal real block triangular (real Schur) arbitrary unitary upper triangular (Schur) arbitrary nonsingular almost diagonal (Jordan) Given matrix A with indicated property, matrices B and T exist with indicated properties such that B T AT If B is diagonal or triangular, eigenvalues are its diagonal entries If B is diagonal, eigenvectors are columns of T Convergence of Power Iteration Michael T. Heath Scientific Computing 27 / 87 To see why power iteration converges to dominant eigenvector, express starting vector x 0 as linear combination n x 0 α i v i i where v i are eigenvectors of A Then x k Ax k A 2 x k 2 A k x 0 ( ) n n λ k i α i v i λ k α v + (λ i /λ ) k α i v i i Since λ i /λ < for i >, successively higher powers go to zero, leaving only component corresponding to v Limitations of Power Iteration i2 Michael T. Heath Scientific Computing 29 / 87 If A A 2 A p A 22 A 2p A.... with square diagonal blocks, then p λ(a) λ(a jj ) j A pp so eigenvalue problem breaks into p smaller eigenvalue problems Real Schur form has diagonal blocks corresponding to real eigenvalues and 2 2 diagonal blocks corresponding to pairs of complex conjugate eigenvalues Power Iteration Michael T. Heath Scientific Computing 26 / 87 Simplest method for computing one eigenvalueeigenvector pair is power iteration, which repeatedly multiplies matrix times initial starting vector Assume A has unique eigenvalue of maximum modulus, say λ, with corresponding eigenvector v Then, starting from nonzero vector x 0, iteration scheme x k Ax k converges to multiple of eigenvector v corresponding to dominant eigenvalue λ Example: Power Iteration Michael T. Heath Scientific Computing 28 / 87 Ratio of values of given component of x k from one iteration to next converges to [ dominant ] eigenvalue [ λ ] For example, if A and x , we obtain k x T k ratio Ratio is converging to dominant eigenvalue, which is 2 Normalized Power Iteration Michael T. Heath Scientific Computing 30 / 87 Power iteration can fail for various reasons Starting vector may have no component in dominant eigenvector v (i.e., α 0) not problem in practice because rounding error usually introduces such component in any case There may be more than one eigenvalue having same (maximum) modulus, in which case iteration may converge to linear combination of corresponding eigenvectors For real matrix and starting vector, iteration can never converge to complex vector Geometric growth of components at each iteration risks eventual overflow (or underflow if λ < ) Approximate eigenvector should be normalized at each iteration, say, by requiring its largest component to be in modulus, giving iteration scheme y k Ax k x k y k / y k With normalization, y k λ, and x k v / v Michael T. Heath Scientific Computing 3 / 87 Michael T. Heath Scientific Computing 32 / 87

5 Example: Normalized Power Iteration Repeating previous example with normalized scheme, k x T k y k < interactive example > Michael T. Heath Scientific Computing 33 / 87 Behavior of power iteration depicted geometrically Initial vector x 0 v + v 2 contains equal components in eigenvectors v and v 2 (dashed arrows) Repeated multiplication by A causes component in v (corresponding to larger eigenvalue, 2) to dominate, so sequence of vectors x k converges to v Michael T. Heath Scientific Computing 34 / 87 Power Iteration with Shift Example: Power Iteration with Shift Convergence rate of power iteration depends on ratio λ 2 /λ, where λ 2 is eigenvalue having second largest modulus May be possible to choose shift, A σi, such that λ 2 σ λ σ < λ 2 so convergence is accelerated λ In earlier example, for instance, if we pick shift of σ, (which is equal to other eigenvalue) then ratio becomes zero and method converges in one iteration In general, we would not be able to make such fortuitous choice, but shifts can still be extremely useful in some contexts, as we will see later Shift must then be added to result to obtain eigenvalue of original matrix Michael T. Heath Scientific Computing 35 / 87 Michael T. Heath Scientific Computing 36 / 87 Inverse Iteration Inverse Iteration, continued If smallest eigenvalue of matrix required rather than largest, can make use of fact that eigenvalues of A are reciprocals of those of A, so smallest eigenvalue of A is reciprocal of largest eigenvalue of A This leads to inverse iteration scheme Ay k x k x k y k / y k Inverse iteration converges to eigenvector corresponding to smallest eigenvalue of A Eigenvalue obtained is dominant eigenvalue of A, and hence its reciprocal is smallest eigenvalue of A in modulus which is equivalent to power iteration applied to A Inverse of A not computed explicitly, but factorization of A used to solve system of linear equations at each iteration Michael T. Heath Scientific Computing 37 / 87 Michael T. Heath Scientific Computing 38 / 87 Example: Inverse Iteration Inverse Iteration with Shift Applying inverse iteration to previous example to compute smallest eigenvalue yields sequence k x T k y k which is indeed converging to (which is its own reciprocal in this case) < interactive example > As before, shifting strategy, working with A σi for some scalar σ, can greatly improve convergence Inverse iteration is particularly useful for computing eigenvector corresponding to approximate eigenvalue, since it converges rapidly when applied to shifted matrix A λi, where λ is approximate eigenvalue Inverse iteration is also useful for computing eigenvalue closest to given value β, since if β is used as shift, then desired eigenvalue corresponds to smallest eigenvalue of shifted matrix Michael T. Heath Scientific Computing 39 / 87 Michael T. Heath Scientific Computing 40 / 87

6 Rayleigh Quotient Example: Rayleigh Quotient Given approximate eigenvector x for real matrix A, determining best estimate for corresponding eigenvalue λ can be considered as n linear least squares approximation problem xλ Ax From normal equation x T xλ x T Ax, least squares solution is given by λ xt Ax x T x This quantity, known as Rayleigh quotient, has many useful properties Rayleigh Quotient Iteration Michael T. Heath Scientific Computing 4 / 87 Rayleigh quotient can accelerate convergence of iterative methods such as power iteration, since Rayleigh quotient x T k Ax k/x T k x k gives better approximation to eigenvalue at iteration k than does basic method alone For previous example using power iteration, value of Rayleigh quotient at each iteration is shown below k x T k y k x T k Ax k/x T k x k Michael T. Heath Scientific Computing 42 / 87 Rayleigh Quotient Iteration, continued Given approximate eigenvector, Rayleigh quotient yields good estimate for corresponding eigenvalue Conversely, inverse iteration converges rapidly to eigenvector if approximate eigenvalue is used as shift, with one iteration often sufficing These two ideas combined in Rayleigh quotient iteration σ k x T k Ax k/x T k x k (A σ k I)y k+ x k x k+ y k+ / y k+ Rayleigh quotient iteration is especially effective for symmetric matrices and usually converges very rapidly Using different shift at each iteration means matrix must be refactored each time to solve linear system, so cost per iteration is high unless matrix has special form that makes factorization easy Same idea also works for complex matrices, for which transpose is replaced by conjugate transpose, so Rayleigh quotient becomes x H Ax/x H x starting from given nonzero vector x 0 Michael T. Heath Scientific Computing 43 / 87 Michael T. Heath Scientific Computing 44 / 87 Example: Rayleigh Quotient Iteration Deflation Using same matrix as previous examples and randomly chosen starting vector x 0, Rayleigh quotient iteration converges in two iterations k x T k σ k Michael T. Heath Scientific Computing 45 / 87 After eigenvalue λ and corresponding eigenvector x have been computed, then additional eigenvalues λ 2,..., λ n of A can be computed by deflation, which effectively removes known eigenvalue Let H be any nonsingular matrix such that Hx αe, scalar multiple of first column of identity matrix (Householder transformation is good choice for H) Then similarity transformation determined by H transforms A into form [ HAH λ b T ] 0 B where B is matrix of order n having eigenvalues λ 2,..., λ n Michael T. Heath Scientific Computing 46 / 87 Deflation, continued Deflation, continued Thus, we can work with B to compute next eigenvalue λ 2 Moreover, if y 2 is eigenvector of B corresponding to λ 2, then x 2 H [ α y 2 ], where α bt y 2 λ 2 λ is eigenvector corresponding to λ 2 for original matrix A, provided λ λ 2 Process can be repeated to find additional eigenvalues and eigenvectors Alternative approach lets u be any vector such that u T x λ Then A x u T has eigenvalues 0, λ 2,..., λ n Possible choices for u include u λ x, if A is symmetric and x is normalized so that x 2 u λ y, where y is corresponding left eigenvector (i.e., A T y λ y ) normalized so that y T x u A T e k, if x is normalized so that x and kth component of x is Michael T. Heath Scientific Computing 47 / 87 Michael T. Heath Scientific Computing 48 / 87

7 Simultaneous Iteration Orthogonal Iteration Simplest method for computing many eigenvalueeigenvector pairs is simultaneous iteration, which repeatedly multiplies matrix times matrix of initial starting vectors Starting from n p matrix X 0 of rank p, iteration scheme is X k AX k span(x k ) converges to invariant subspace determined by p largest eigenvalues of A, provided λ p > λ p+ Also called subspace iteration Michael T. Heath Scientific Computing 49 / 87 As with power iteration, normalization is needed with simultaneous iteration Each column of X k converges to dominant eigenvector, so columns of X k become increasingly ill-conditioned basis for span(x k ) Both issues can be addressed by computing QR factorization at each iteration ˆQ k R k X k X k A ˆQ k where ˆQ k R k is reduced QR factorization of X k This orthogonal iteration converges to block triangular form, and leading block is triangular if moduli of consecutive eigenvalues are distinct Michael T. Heath Scientific Computing 50 / 87 QR Iteration QR Iteration, continued For p n and X 0 I, matrices A k ˆQ H k A ˆQ k generated by orthogonal iteration converge to triangular or block triangular form, yielding all eigenvalues of A QR iteration computes successive matrices A k without forming above product explicitly Starting with A 0 A, at iteration k compute QR factorization Q k R k A k and form reverse product A k R k Q k Successive matrices A k are unitarily similar to each other A k R k Q k Q H k A k Q k Diagonal entries (or eigenvalues of diagonal blocks) of A k converge to eigenvalues of A Product of orthogonal matrices Q k converges to matrix of corresponding eigenvectors If A is symmetric, then symmetry is preserved by QR iteration, so A k converge to matrix that is both triangular and symmetric, hence diagonal Example: QR Iteration Michael T. Heath Scientific Computing 5 / Let A Compute QR factorization A 0 Q R and form reverse product A R Q [ 7.83 ] Off-diagonal entries are now smaller, and diagonal entries closer to eigenvalues, 8 and 3 Process continues until matrix is within tolerance of being diagonal, and diagonal entries then closely approximate eigenvalues Michael T. Heath Scientific Computing 53 / 87 QR Iteration with Shifts Michael T. Heath Scientific Computing 52 / 87 Convergence rate of QR iteration can be accelerated by incorporating shifts Q k R k A k σ k I A k R k Q k + σ k I where σ k is rough approximation to eigenvalue Good shift can be determined by computing eigenvalues of 2 2 submatrix in lower right corner of matrix Michael T. Heath Scientific Computing 54 / 87 Example: QR Iteration with Shifts Preliminary Reduction Repeat previous example, but with shift of σ 4, which is lower right corner entry of matrix We compute QR factorization A 0 σ I Q R and form reverse product, adding back shift to obtain A R Q + σ I After one iteration, off-diagonal entries smaller compared with unshifted algorithm, and eigenvalues closer approximations to eigenvalues < interactive example > Efficiency of QR iteration can be enhanced by first transforming matrix as close to triangular form as possible before beginning iterations Hessenberg matrix is triangular except for one additional nonzero diagonal immediately adjacent to main diagonal Any matrix can be reduced to Hessenberg form in finite number of steps by orthogonal similarity transformation, for example using Householder transformations Symmetric Hessenberg matrix is tridiagonal Hessenberg or tridiagonal form is preserved during successive QR iterations Michael T. Heath Scientific Computing 55 / 87 Michael T. Heath Scientific Computing 56 / 87

8 Preliminary Reduction, continued Advantages of initial reduction to upper Hessenberg or tridiagonal form Work per QR iteration is reduced from O(n 3 ) to O(n 2 ) for general matrix or O(n) for symmetric matrix Fewer QR iterations are required because matrix nearly triangular (or diagonal) already If any zero entries on first subdiagonal, then matrix is block triangular and problem can be broken into two or more smaller subproblems Michael T. Heath Scientific Computing 57 / 87 Preliminary Reduction, continued QR iteration is implemented in two-stages symmetric tridiagonal diagonal or general Hessenberg triangular Preliminary reduction requires definite number of steps, whereas subsequent iterative stage continues until convergence In practice only modest number of iterations usually required, so much of work is in preliminary reduction Cost of accumulating eigenvectors, if needed, dominates total cost Michael T. Heath Scientific Computing 58 / 87 Cost of QR Iteration Krylov Subspace Methods Approximate overall cost of preliminary reduction and QR iteration, counting both additions and multiplications Symmetric matrices 4 3 n3 for eigenvalues only 9n 3 for eigenvalues and eigenvectors General matrices 0n 3 for eigenvalues only 25n 3 for eigenvalues and eigenvectors Michael T. Heath Scientific Computing 59 / 87 Krylov subspace methods reduce matrix to Hessenberg (or tridiagonal) form using only matrix-vector multiplication For arbitrary starting vector x 0, if K k [ x 0 Ax 0 A k x 0 ] then Kn AK n C n where C n is upper Hessenberg (in fact, companion matrix) To obtain better conditioned basis for span(k n ), compute QR factorization Q n R n K n so that Q H n AQ n R n C n Rn H with H upper Hessenberg Michael T. Heath Scientific Computing 60 / 87 Krylov Subspace Methods Arnoldi Iteration Equating kth columns on each side of equation AQ n Q n H yields recurrence Aq k h k q + + h kk q k + h k+,k q k+ relating q k+ to preceding vectors q,..., q k Premultiplying by q H j and using orthonormality, h jk q H j Aq k, j,..., k These relationships yield Arnoldi iteration, which produces unitary matrix Q n and upper Hessenberg matrix H n column by column using only matrix-vector multiplication by A and inner products of vectors x 0 arbitrary nonzero starting vector q x 0 / x 0 2 for k, 2,... u k Aq k for j to k h jk q H j u k u k u k h jk q j end h k+,k u k 2 if h k+,k 0 then stop q k+ u k /h k+,k end Arnoldi Iteration, continued Michael T. Heath Scientific Computing 6 / 87 Arnoldi Iteration, continued Michael T. Heath Scientific Computing 62 / 87 If then is upper Hessenberg matrix Q k [ q q k ] H k Q H k AQ k Eigenvalues of H k, called Ritz values, are approximate eigenvalues of A, and Ritz vectors given by Q k y, where y is eigenvector of H k, are corresponding approximate eigenvectors of A Eigenvalues of H k must be computed by another method, such as QR iteration, but this is easier problem if k n Arnoldi iteration fairly expensive in work and storage because each new vector q k must be orthogonalized against all previous columns of Q k, and all must be stored for that purpose. So Arnoldi process usually restarted periodically with carefully chosen starting vector Ritz values and vectors produced are often good approximations to eigenvalues and eigenvectors of A after relatively few iterations Michael T. Heath Scientific Computing 63 / 87 Michael T. Heath Scientific Computing 64 / 87

9 Lanczos Iteration Work and storage costs drop dramatically if matrix is symmetric or Hermitian, since recurrence then has only three terms and H k is tridiagonal (so usually denoted T k ) q 0 0 β 0 0 x 0 arbitrary nonzero starting vector q x 0 / x 0 2 for k, 2,... u k Aq k α k q H k u k u k u k β k q k α k q k β k u k 2 if β k 0 then stop q k+ u k /β k end Lanczos Iteration, continued Michael T. Heath Scientific Computing 65 / 87 Lanczos Iteration, continued α k and β k are diagonal and subdiagonal entries of symmetric tridiagonal matrix T k As with Arnoldi, Lanczos iteration does not produce eigenvalues and eigenvectors directly, but only tridiagonal matrix T k, whose eigenvalues and eigenvectors must be computed by another method to obtain Ritz values and vectors If β k 0, then algorithm appears to break down, but in that case invariant subspace has already been identified (i.e., Ritz values and vectors are already exact at that point) Michael T. Heath Scientific Computing 66 / 87 Krylov Subspace Methods, continued In principle, if Lanczos algorithm were run until k n, resulting tridiagonal matrix would be orthogonally similar to A In practice, rounding error causes loss of orthogonality, invalidating this expectation Problem can be overcome by reorthogonalizing vectors as needed, but expense can be substantial Alternatively, can ignore problem, in which case algorithm still produces good eigenvalue approximations, but multiple copies of some eigenvalues may be generated Great virtue of Arnoldi and Lanczos iterations is their ability to produce good approximations to extreme eigenvalues for k n Moreover, they require only one matrix-vector multiplication by A per step and little auxiliary storage, so are ideally suited to large sparse matrices If eigenvalues are needed in middle of spectrum, say near σ, then algorithm can be applied to matrix (A σi), assuming it is practical to solve systems of form (A σi)x y Example: Lanczos Iteration Michael T. Heath Scientific Computing 67 / 87 For symmetric matrix with eigenvalues,..., 29, behavior of Lanczos iteration is shown below Michael T. Heath Scientific Computing 69 / 87 Jacobi Method Michael T. Heath Scientific Computing 68 / 87 One of oldest methods for computing eigenvalues is Jacobi method, which uses similarity transformation based on plane rotations Sequence of plane rotations chosen to annihilate symmetric pairs of matrix entries, eventually converging to diagonal form Choice of plane rotation slightly more complicated than in Givens method for QR factorization To annihilate given off-diagonal pair, choose c and s so that J T c s a b c s AJ s c b d s c [ c 2 a 2csb + s 2 d c 2 b + cs(a d) s 2 ] b c 2 b + cs(a d) s 2 b c 2 d + 2csb + s 2 a is diagonal Michael T. Heath Scientific Computing 70 / 87 Jacobi Method, continued Example: Plane Rotation Transformed matrix diagonal if c 2 b + cs(a d) s 2 b 0 Dividing both sides by c 2 b, we obtain + s (a d) s2 c b c 2 0 Making substitution t s/c, we get quadratic equation (a d) + t t 2 0 b for tangent t of angle of rotation, from which we can recover c /( + t 2 ) and s c t Advantageous numerically to use root of smaller magnitude Michael T. Heath Scientific Computing 7 / 87 Consider 2 2 matrix A 2 2 Quadratic equation for tangent reduces to t 2, so t ± Two roots of same magnitude, so we arbitrarily choose t, which yields c / 2 and s / 2 Resulting plane rotation J gives [ ] [ ] J T / 2 / 2 AJ / 2 / 2 / 2 / / 2 / Michael T. Heath Scientific Computing 72 / 87

10 Jacobi Method, continued Jacobi Method, continued Starting with symmetric matrix A 0 A, each iteration has form A k+ J T k A kj k where J k is plane rotation chosen to annihilate a symmetric pair of entries in A k Plane rotations repeatedly applied from both sides in systematic sweeps through matrix until off-diagonal mass of matrix is reduced to within some tolerance of zero Resulting diagonal matrix orthogonally similar to original matrix, so diagonal entries are eigenvalues, and eigenvectors are given by product of plane rotations Jacobi method is reliable, simple to program, and capable of high accuracy, but converges rather slowly and difficult to generalize beyond symmetric matrices Except for small problems, more modern methods usually require 5 to 0 times less work than Jacobi One source of inefficiency is that previously annihilated entries can subsequently become nonzero again, thereby requiring repeated annihilation Newer methods such as QR iteration preserve zero entries introduced into matrix Example: Jacobi Method 0 2 Let A Michael T. Heath Scientific Computing 73 / 87 First annihilate (,3) and (3,) entries using rotation J to obtain A J0 T A 0 J Example, continued Michael T. Heath Scientific Computing 74 / 87 Next annihilate (,2) and (2,) entries using rotation J to obtain A 2 J T A J Michael T. Heath Scientific Computing 75 / 87 Michael T. Heath Scientific Computing 76 / 87 Example, continued Example, continued Next annihilate (2,3) and (3,2) entries using rotation 0 0 J to obtain A 3 J2 T A 2 J Beginning new sweep, again annihilate (,3) and (3,) entries using rotation J to obtain A 4 J3 T A 3 J Michael T. Heath Scientific Computing 77 / 87 Michael T. Heath Scientific Computing 78 / 87 Example, continued Bisection or Spectrum-Slicing Process continues until off-diagonal entries reduced to as small as desired Result is diagonal matrix orthogonally similar to original matrix, with the orthogonal similarity transformation given by product of plane rotations < interactive example > For real symmetric matrix, can determine how many eigenvalues are less than given real number σ By systematically choosing various values for σ (slicing spectrum at σ) and monitoring resulting count, any eigenvalue can be isolated as accurately as desired For example, symmetric indefinite factorization A LDL T makes inertia (numbers of positive, negative, and zero eigenvalues) of symmetric matrix A easy to determine By applying factorization to matrix A σi for various values of σ, individual eigenvalues can be isolated as accurately as desired using interval bisection technique Michael T. Heath Scientific Computing 79 / 87 Michael T. Heath Scientific Computing 80 / 87

11 Sturm Sequence Divide-and-Conquer Method Another spectrum-slicing method for computing individual eigenvalues is based on Sturm sequence property of symmetric matrices Let A be symmetric matrix and let p r (σ) denote determinant of leading principal minor of order r of A σi Then zeros of p r (σ) strictly separate those of p r (σ), and number of agreements in sign of successive members of sequence p r (σ), for r,..., n, equals number of eigenvalues of A strictly greater than σ Determinants p r (σ) are easy to compute if A is transformed to tridiagonal form before applying Sturm sequence technique Express symmetric tridiagonal matrix T as T O T + β uu T O T 2 Can now compute eigenvalues and eigenvectors of smaller matrices T and T 2 To relate these back to eigenvalues and eigenvectors of original matrix requires solution of secular equation, which can be done reliably and efficiently Applying this approach recursively yields divide-and-conquer algorithm for symmetric tridiagonal eigenproblems Relatively Robust Representation Michael T. Heath Scientific Computing 8 / 87 With conventional methods, cost of computing eigenvalues of symmetric tridiagonal matrix is O(n 2 ), but if orthogonal eigenvectors are also computed, then cost rises to O(n 3 ) Another possibility is to compute eigenvalues first at O(n 2 ) cost, and then compute corresponding eigenvectors separately using inverse iteration with computed eigenvalues as shifts Key to making this idea work is computing eigenvalues and corresponding eigenvectors to very high relative accuracy so that expensive explicit orthogonalization of eigenvectors is not needed RRR algorithm exploits this approach to produce eigenvalues and orthogonal eigenvectors at O(n 2 ) cost Michael T. Heath Scientific Computing 83 / 87 Michael T. Heath Scientific Computing 82 / 87 Generalized Generalized eigenvalue problem has form Ax λbx where A and B are given n n matrices If either A or B is nonsingular, then generalized eigenvalue problem can be converted to standard eigenvalue problem, either (B A)x λx or (A B)x (/λ)x This is not recommended, since it may cause loss of accuracy due to rounding error loss of symmetry if A and B are symmetric Better alternative for generalized eigenvalue problems is QZ algorithm Michael T. Heath Scientific Computing 84 / 87 QZ Algorithm QZ Algorithm, continued If A and B are triangular, then eigenvalues are given by λ i a ii /b ii, for b ii 0 QZ algorithm reduces A and B simultaneously to upper triangular form by orthogonal transformations First, B is reduced to upper triangular form by orthogonal transformation from left, which is also applied to A Next, transformed A is reduced to upper Hessenberg form by orthogonal transformation from left, while maintaining triangular form of B, which requires additional transformations from right Finally, analogous to QR iteration, A is reduced to triangular form while still maintaining triangular form of B, which again requires transformations from both sides Eigenvalues can now be determined from mutually triangular form, and eigenvectors can be recovered from products of left and right transformations, denoted by Q and Z Michael T. Heath Scientific Computing 85 / 87 Michael T. Heath Scientific Computing 86 / 87 Computing SVD Singular values of A are nonnegative square roots of eigenvalues of A T A, and columns of U and V are orthonormal eigenvectors of AA T and A T A, respectively Algorithms for computing SVD work directly with A, without forming AA T or A T A, to avoid loss of information associated with forming these matrix products explicitly SVD is usually computed by variant of QR iteration, with A first reduced to bidiagonal form by orthogonal transformations, then remaining off-diagonal entries are annihilated iteratively SVD can also be computed by variant of Jacobi method, which can be useful on parallel computers or if matrix has special structure Michael T. Heath Scientific Computing 87 / 87

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course G63.2010.001 / G22.2420-001, Fall 2010 September 30th, 2010 A. Donev (Courant Institute)

More information

ALGEBRAIC EIGENVALUE PROBLEM

ALGEBRAIC EIGENVALUE PROBLEM ALGEBRAIC EIGENVALUE PROBLEM BY J. H. WILKINSON, M.A. (Cantab.), Sc.D. Technische Universes! Dsrmstedt FACHBEREICH (NFORMATiK BIBL1OTHEK Sachgebieto:. Standort: CLARENDON PRESS OXFORD 1965 Contents 1.

More information

Similarity and Diagonalization. Similar Matrices

Similarity and Diagonalization. Similar Matrices MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

More information

Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

More information

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013 Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

More information

Factorization Theorems

Factorization Theorems Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

3 Orthogonal Vectors and Matrices

3 Orthogonal Vectors and Matrices 3 Orthogonal Vectors and Matrices The linear algebra portion of this course focuses on three matrix factorizations: QR factorization, singular valued decomposition (SVD), and LU factorization The first

More information

7 Gaussian Elimination and LU Factorization

7 Gaussian Elimination and LU Factorization 7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method

More information

LINEAR ALGEBRA. September 23, 2010

LINEAR ALGEBRA. September 23, 2010 LINEAR ALGEBRA September 3, 00 Contents 0. LU-decomposition.................................... 0. Inverses and Transposes................................. 0.3 Column Spaces and NullSpaces.............................

More information

Applied Linear Algebra I Review page 1

Applied Linear Algebra I Review page 1 Applied Linear Algebra Review 1 I. Determinants A. Definition of a determinant 1. Using sum a. Permutations i. Sign of a permutation ii. Cycle 2. Uniqueness of the determinant function in terms of properties

More information

by the matrix A results in a vector which is a reflection of the given

by the matrix A results in a vector which is a reflection of the given Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

More information

Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

More information

Nonlinear Iterative Partial Least Squares Method

Nonlinear Iterative Partial Least Squares Method Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for

More information

Lecture 1: Schur s Unitary Triangularization Theorem

Lecture 1: Schur s Unitary Triangularization Theorem Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections

More information

Vector and Matrix Norms

Vector and Matrix Norms Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

More information

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression The SVD is the most generally applicable of the orthogonal-diagonal-orthogonal type matrix decompositions Every

More information

13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions. 3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,

More information

CS3220 Lecture Notes: QR factorization and orthogonal transformations

CS3220 Lecture Notes: QR factorization and orthogonal transformations CS3220 Lecture Notes: QR factorization and orthogonal transformations Steve Marschner Cornell University 11 March 2009 In this lecture I ll talk about orthogonal matrices and their properties, discuss

More information

Orthogonal Bases and the QR Algorithm

Orthogonal Bases and the QR Algorithm Orthogonal Bases and the QR Algorithm Orthogonal Bases by Peter J Olver University of Minnesota Throughout, we work in the Euclidean vector space V = R n, the space of column vectors with n real entries

More information

The Characteristic Polynomial

The Characteristic Polynomial Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem

More information

Chapter 6. Orthogonality

Chapter 6. Orthogonality 6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively. Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry

More information

A note on companion matrices

A note on companion matrices Linear Algebra and its Applications 372 (2003) 325 33 www.elsevier.com/locate/laa A note on companion matrices Miroslav Fiedler Academy of Sciences of the Czech Republic Institute of Computer Science Pod

More information

Notes on Determinant

Notes on Determinant ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

More information

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices. Exercise 1 1. Let A be an n n orthogonal matrix. Then prove that (a) the rows of A form an orthonormal basis of R n. (b) the columns of A form an orthonormal basis of R n. (c) for any two vectors x,y R

More information

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in

More information

Examination paper for TMA4205 Numerical Linear Algebra

Examination paper for TMA4205 Numerical Linear Algebra Department of Mathematical Sciences Examination paper for TMA4205 Numerical Linear Algebra Academic contact during examination: Markus Grasmair Phone: 97580435 Examination date: December 16, 2015 Examination

More information

Orthogonal Diagonalization of Symmetric Matrices

Orthogonal Diagonalization of Symmetric Matrices MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

More information

Linear Algebra: Determinants, Inverses, Rank

Linear Algebra: Determinants, Inverses, Rank D Linear Algebra: Determinants, Inverses, Rank D 1 Appendix D: LINEAR ALGEBRA: DETERMINANTS, INVERSES, RANK TABLE OF CONTENTS Page D.1. Introduction D 3 D.2. Determinants D 3 D.2.1. Some Properties of

More information

Solution of Linear Systems

Solution of Linear Systems Chapter 3 Solution of Linear Systems In this chapter we study algorithms for possibly the most commonly occurring problem in scientific computing, the solution of linear systems of equations. We start

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length

More information

Similar matrices and Jordan form

Similar matrices and Jordan form Similar matrices and Jordan form We ve nearly covered the entire heart of linear algebra once we ve finished singular value decompositions we ll have seen all the most central topics. A T A is positive

More information

Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems

Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course G63.2010.001 / G22.2420-001,

More information

Operation Count; Numerical Linear Algebra

Operation Count; Numerical Linear Algebra 10 Operation Count; Numerical Linear Algebra 10.1 Introduction Many computations are limited simply by the sheer number of required additions, multiplications, or function evaluations. If floating-point

More information

Direct Methods for Solving Linear Systems. Matrix Factorization

Direct Methods for Solving Linear Systems. Matrix Factorization Direct Methods for Solving Linear Systems Matrix Factorization Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011

More information

Lecture 5: Singular Value Decomposition SVD (1)

Lecture 5: Singular Value Decomposition SVD (1) EEM3L1: Numerical and Analytical Techniques Lecture 5: Singular Value Decomposition SVD (1) EE3L1, slide 1, Version 4: 25-Sep-02 Motivation for SVD (1) SVD = Singular Value Decomposition Consider the system

More information

ASEN 3112 - Structures. MDOF Dynamic Systems. ASEN 3112 Lecture 1 Slide 1

ASEN 3112 - Structures. MDOF Dynamic Systems. ASEN 3112 Lecture 1 Slide 1 19 MDOF Dynamic Systems ASEN 3112 Lecture 1 Slide 1 A Two-DOF Mass-Spring-Dashpot Dynamic System Consider the lumped-parameter, mass-spring-dashpot dynamic system shown in the Figure. It has two point

More information

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 10

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 10 Lecture Notes to Accompany Scientific Computing An Introductory Survey Second Edition by Michael T. Heath Chapter 10 Boundary Value Problems for Ordinary Differential Equations Copyright c 2001. Reproduction

More information

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n. ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?

More information

NUMERICAL METHODS FOR LARGE EIGENVALUE PROBLEMS

NUMERICAL METHODS FOR LARGE EIGENVALUE PROBLEMS NUMERICAL METHODS FOR LARGE EIGENVALUE PROBLEMS Second edition Yousef Saad Copyright c 2011 by the Society for Industrial and Applied Mathematics Contents Preface to Classics Edition Preface xiii xv 1

More information

SOLVING LINEAR SYSTEMS

SOLVING LINEAR SYSTEMS SOLVING LINEAR SYSTEMS Linear systems Ax = b occur widely in applied mathematics They occur as direct formulations of real world problems; but more often, they occur as a part of the numerical analysis

More information

Section 1.1. Introduction to R n

Section 1.1. Introduction to R n The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to

More information

Inner Product Spaces

Inner Product Spaces Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number

More information

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two

More information

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued). MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0

More information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1. MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

More information

Systems of Linear Equations

Systems of Linear Equations Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

More information

DATA ANALYSIS II. Matrix Algorithms

DATA ANALYSIS II. Matrix Algorithms DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where

More information

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8 Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e

More information

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

Chapter 12 Modal Decomposition of State-Space Models 12.1 Introduction The solutions obtained in previous chapters, whether in time domain or transfor

Chapter 12 Modal Decomposition of State-Space Models 12.1 Introduction The solutions obtained in previous chapters, whether in time domain or transfor Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology 1 1 c Chapter

More information

Derivative Free Optimization

Derivative Free Optimization Department of Mathematics Derivative Free Optimization M.J.D. Powell LiTH-MAT-R--2014/02--SE Department of Mathematics Linköping University S-581 83 Linköping, Sweden. Three lectures 1 on Derivative Free

More information

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Review Jeopardy. Blue vs. Orange. Review Jeopardy Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?

More information

Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

More information

1 Sets and Set Notation.

1 Sets and Set Notation. LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most

More information

[1] Diagonal factorization

[1] Diagonal factorization 8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:

More information

1 Solving LPs: The Simplex Algorithm of George Dantzig

1 Solving LPs: The Simplex Algorithm of George Dantzig Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

More information

5. Orthogonal matrices

5. Orthogonal matrices L Vandenberghe EE133A (Spring 2016) 5 Orthogonal matrices matrices with orthonormal columns orthogonal matrices tall matrices with orthonormal columns complex matrices with orthonormal columns 5-1 Orthonormal

More information

LS.6 Solution Matrices

LS.6 Solution Matrices LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions

More information

1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES 1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

More information

Lecture Topic: Low-Rank Approximations

Lecture Topic: Low-Rank Approximations Lecture Topic: Low-Rank Approximations Low-Rank Approximations We have seen principal component analysis. The extraction of the first principle eigenvalue could be seen as an approximation of the original

More information

Metric Spaces. Chapter 7. 7.1. Metrics

Metric Spaces. Chapter 7. 7.1. Metrics Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some

More information

ISOMETRIES OF R n KEITH CONRAD

ISOMETRIES OF R n KEITH CONRAD ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x

More information

Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

More information

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

More information

UNCOUPLING THE PERRON EIGENVECTOR PROBLEM

UNCOUPLING THE PERRON EIGENVECTOR PROBLEM UNCOUPLING THE PERRON EIGENVECTOR PROBLEM Carl D Meyer INTRODUCTION Foranonnegative irreducible matrix m m with spectral radius ρ,afundamental problem concerns the determination of the unique normalized

More information

Solving Systems of Linear Equations

Solving Systems of Linear Equations LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Chapter 6 Eigenvalues and Eigenvectors 6. Introduction to Eigenvalues Linear equations Ax D b come from steady state problems. Eigenvalues have their greatest importance in dynamic problems. The solution

More information

Towards a deeper understanding of Chris Paige s error analysis of the finite precision Lanczos process

Towards a deeper understanding of Chris Paige s error analysis of the finite precision Lanczos process Towards a deeper understanding of Chris Paige s error analysis of the finite precision Lanczos process Jens-Peter M. Zemke zemke@tu-harburg.de Institut für Numerische Simulation Technische Universität

More information

Matrix Differentiation

Matrix Differentiation 1 Introduction Matrix Differentiation ( and some other stuff ) Randal J. Barnes Department of Civil Engineering, University of Minnesota Minneapolis, Minnesota, USA Throughout this presentation I have

More information

9 MATRICES AND TRANSFORMATIONS

9 MATRICES AND TRANSFORMATIONS 9 MATRICES AND TRANSFORMATIONS Chapter 9 Matrices and Transformations Objectives After studying this chapter you should be able to handle matrix (and vector) algebra with confidence, and understand the

More information

Chapter 7. Lyapunov Exponents. 7.1 Maps

Chapter 7. Lyapunov Exponents. 7.1 Maps Chapter 7 Lyapunov Exponents Lyapunov exponents tell us the rate of divergence of nearby trajectories a key component of chaotic dynamics. For one dimensional maps the exponent is simply the average

More information

MATH 551 - APPLIED MATRIX THEORY

MATH 551 - APPLIED MATRIX THEORY MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: SVD revisited; Software for Linear Algebra Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 9 Outline 1 Computing

More information

The Matrix Elements of a 3 3 Orthogonal Matrix Revisited

The Matrix Elements of a 3 3 Orthogonal Matrix Revisited Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation

More information

1 Determinants and the Solvability of Linear Systems

1 Determinants and the Solvability of Linear Systems 1 Determinants and the Solvability of Linear Systems In the last section we learned how to use Gaussian elimination to solve linear systems of n equations in n unknowns The section completely side-stepped

More information

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In

More information

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone

More information

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

More information

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

More information

October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix

October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix Linear Algebra & Properties of the Covariance Matrix October 3rd, 2012 Estimation of r and C Let rn 1, rn, t..., rn T be the historical return rates on the n th asset. rn 1 rṇ 2 r n =. r T n n = 1, 2,...,

More information

Elementary Linear Algebra

Elementary Linear Algebra Elementary Linear Algebra Kuttler January, Saylor URL: http://wwwsaylororg/courses/ma/ Saylor URL: http://wwwsaylororg/courses/ma/ Contents Some Prerequisite Topics Sets And Set Notation Functions Graphs

More information

LINEAR ALGEBRA W W L CHEN

LINEAR ALGEBRA W W L CHEN LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,

More information

6. Cholesky factorization

6. Cholesky factorization 6. Cholesky factorization EE103 (Fall 2011-12) triangular matrices forward and backward substitution the Cholesky factorization solving Ax = b with A positive definite inverse of a positive definite matrix

More information

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0. Cross product 1 Chapter 7 Cross product We are getting ready to study integration in several variables. Until now we have been doing only differential calculus. One outcome of this study will be our ability

More information

Lecture L3 - Vectors, Matrices and Coordinate Transformations

Lecture L3 - Vectors, Matrices and Coordinate Transformations S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between

More information

4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

More information

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they

More information

Solving Linear Systems, Continued and The Inverse of a Matrix

Solving Linear Systems, Continued and The Inverse of a Matrix , Continued and The of a Matrix Calculus III Summer 2013, Session II Monday, July 15, 2013 Agenda 1. The rank of a matrix 2. The inverse of a square matrix Gaussian Gaussian solves a linear system by reducing

More information

Using row reduction to calculate the inverse and the determinant of a square matrix

Using row reduction to calculate the inverse and the determinant of a square matrix Using row reduction to calculate the inverse and the determinant of a square matrix Notes for MATH 0290 Honors by Prof. Anna Vainchtein 1 Inverse of a square matrix An n n square matrix A is called invertible

More information

Notes on Symmetric Matrices

Notes on Symmetric Matrices CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.

More information

Linear Algebra: Vectors

Linear Algebra: Vectors A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector

More information

Math Review. for the Quantitative Reasoning Measure of the GRE revised General Test

Math Review. for the Quantitative Reasoning Measure of the GRE revised General Test Math Review for the Quantitative Reasoning Measure of the GRE revised General Test www.ets.org Overview This Math Review will familiarize you with the mathematical skills and concepts that are important

More information

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem

More information