We shall turn our attention to solving linear systems of equations. Ax = b



Similar documents
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

Linear Algebra Notes for Marsden and Tromba Vector Calculus

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

CS3220 Lecture Notes: QR factorization and orthogonal transformations

(67902) Topics in Theory and Complexity Nov 2, Lecture 7

General Framework for an Iterative Solution of Ax b. Jacobi s Method

Orthogonal Projections

Inner product. Definition of inner product

Numerical Analysis Lecture Notes

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

Data Mining: Algorithms and Applications Matrix Math Review

8. Linear least-squares

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Vector and Matrix Norms

Similarity and Diagonalization. Similar Matrices

Lecture 1: Systems of Linear Equations

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

1 Introduction to Matrices

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices

Section 1.1. Introduction to R n

3. INNER PRODUCT SPACES

Lecture 3: Finding integer solutions to systems of linear equations

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8

Applied Algorithm Design Lecture 5

LINEAR ALGEBRA W W L CHEN

Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday.

Linear Algebra: Vectors

SOLVING LINEAR SYSTEMS

1 Norms and Vector Spaces

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

Math 312 Homework 1 Solutions

How To Prove The Dirichlet Unit Theorem

Numerical Analysis Lecture Notes

Factorization Theorems

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Metric Spaces. Chapter Metrics

Lecture 2 Matrix Operations

1 Review of Least Squares Solutions to Overdetermined Systems

Recall that two vectors in are perpendicular or orthogonal provided that their dot

Math 4310 Handout - Quotient Vector Spaces

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Class Meeting # 1: Introduction to PDEs

1 VECTOR SPACES AND SUBSPACES

NOTES ON LINEAR TRANSFORMATIONS

Section 6.1 Factoring Expressions

Linear Codes. Chapter Basics

Solving Systems of Linear Equations

Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product

What is Linear Programming?

Notes on Factoring. MA 206 Kurt Bryan

Inner Product Spaces and Orthogonality

DERIVATIVES AS MATRICES; CHAIN RULE

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

LECTURE: INTRO TO LINEAR PROGRAMMING AND THE SIMPLEX METHOD, KEVIN ROSS MARCH 31, 2005

Matrix Calculations: Applications of Eigenvalues and Eigenvectors; Inner Products

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

α = u v. In other words, Orthogonal Projection

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Metrics on SO(3) and Inverse Kinematics

1 Lecture: Integration of rational functions by decomposition

Manifold Learning Examples PCA, LLE and ISOMAP

1 Determinants and the Solvability of Linear Systems

Lecture 1: Schur s Unitary Triangularization Theorem

Methods for Finding Bases

Linear Programming I

Systems of Linear Equations

8 Square matrices continued: Determinants

160 CHAPTER 4. VECTOR SPACES

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

A linear algebraic method for pricing temporary life annuities

Chapter 7. Matrices. Definition. An m n matrix is an array of numbers set out in m rows and n columns. Examples. (

South Carolina College- and Career-Ready (SCCCR) Pre-Calculus

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

2.3 Convex Constrained Optimization Problems

Continued Fractions and the Euclidean Algorithm

5. Orthogonal matrices

Solution of Linear Systems

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = i.

On the representability of the bi-uniform matroid

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

Ideal Class Group and Units

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

Chapter 4, Arithmetic in F [x] Polynomial arithmetic and the division algorithm.

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

MATH1231 Algebra, 2015 Chapter 7: Linear maps

Brief Introduction to Vectors and Matrices

Arithmetic and Algebra of Matrices

Orthogonal Diagonalization of Symmetric Matrices

3 Orthogonal Vectors and Matrices

The Characteristic Polynomial

The van Hoeij Algorithm for Factoring Polynomials

MATH APPLIED MATRIX THEORY

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances

1 Sets and Set Notation.

Transcription:

59 Linear Algebra We shall turn our attention to solving linear systems of equations Ax = b where A R m n, x R n, and b R m. We already saw examples of methods that required the solution of a linear system as part of the overall algorithm, e.g. the Vandermonde system for interpolation, which was a square system (m = n). Another category of methods that leads to rectangular systems with m > n is least square methods. They answer questions of the form: What is the best n-order polynomial we can use to approximate (not interpolate) m + 1 data points (where m > n) More generally, find the solution that most closely satisfies m equations, in the presence of n (n < m) unknowns. All these algorithms need to be conscious about error and there are at least 3 sources where error stems from: Lecture of: 19 Mar 213 Some algorithms are imperfect in the sense that they require several iterations to generate a good quality approximation. Thus, intermediate results are subject to error. Sometimes, it is not possible to fine an ideal solution, e.g. because we have more equations than unknowns. In this case, not all equations will be satisfied exactly, and we need a notion of the error incurred in not satisfying certain equations fully. Inputs to an algorithm are often corrupted by noise, roundoff error, etc. For example, instead of solving an intended system Ax = b we may be solving A x = b where the entries A and b have been subject to noise and inaccuracy. It is important to know how these translate to errors in determining x. Vector Norms Norms are valuable tools in arguing about the extent and magnitude of error. We will introduce some concepts that we will use broadly later on.

6 Definition: A vector norm is a function from R n to R, with a certain number of properties. If x R n, we symbolize its norm by x. The defining properties of a norm are : (i) x for all x R n and also x = iff x = (ii) αx = α x for all α R, x R n (iii) x + y x + y for all x, y R n (triangle inequality) Note that the properties above do not determine a unique form of a norm function, in fact many different valid norms exist. Typically we will use subscripts ( a, b ) to denote different types of norms. So, why do we need to introduce the concept of a vector norm? The reason has to do with the following fact: When dealing with the solution of a nonlinear equation f(x) =, the error e = xapprox x exact is a single number, thus the absolute value e gives us a good idea of the extent of the error. When solving a system of linear equations Ax = b, the exact solution x exact as well as any approximation xapprox are vectors, and the error: e = xapprox x exact is a vector, too. It is not straightforward to assess the magnitude of such a vector-valued error. e.g. Consider e 1, e 2 R 1, and e 1 =. e 2 = Which one is worse? e 1 has a modest amount of error, distributed over all components. In e 2, all but one component are exact, but one of them has a huge discrepancy. Exactly how we quantify and assess the extent of error is application-dependent. Vector norms are alternative ways to measure this magnitude and different norms would be appropriate for different tasks. There are, however, some specific definitions of vector norms, which are commonly used. All of them can be proven to satisfy the norm properties, and their definitions are: 1.

61 (In all definitions below, x = (x 1, x 2,, x n )) 1. The L 1 -norm (or 1-norm) x 1 = x i 2. The L 2 -norm (or 2-norm, or Euclidean norm) x 2 = n 3. The infinity norm (or max-norm) 4. (Less common) L p norm x 2 i x = max 1 i n x i ( n ) 1/p x p = x i p It is relatively easy to show that these satisfy the defining properties of a norm e.g. for 1 : x = n x i if x =, then x = if x = n x i = x i = i x = αx = n αx i = α x i = α x 1 Similar proofs can be given for (just as easy), 2 (a bit more difficult) and p (rather complicated). Matrix Norms We can actually define norms for (square) matrices, as well. Definition: A matrix norm is a function from R n n to R which satisfies a certain number of properties. If A R n n, we symbolize its norm by A. The defining properties of a matrix norm are :

62 (i) M for all M R n n, and also M = iff M = (ii) αm = α M for all α R (iii) M + N M + N (iv) M N M N Property (iv) is the one that has slightly different flavor than the properties of vector norms. Although more types of matrix norms do exist, one common category is that of matrix norms induced by vector norms. Definition: defined as or equivalently: If is a valid vector norm, its induced matrix norm is M = max x R n x { } Mx x M = max { Mx } x R n x =1 Note, again, that not all valid matrix norms are induced by vector norms. One notable example is the very commonly used Frobenius norm: M F = Mij 2 i,j=1 We can easily show though that induced norms satisfy properties (i) through (iv). (i)-(iii) are rather trivial, e.g,: (M + N)x M + N = max x x = max x Mx x + max x Mx + Nx max x x Nx x Property (iv) is slightly trickier to show. First a lemma: = M + N Lemma: If is a matrix norm induced by a vector norm, then: Ax A x

63 Proof: Since A = maxx ( Ax / x ), we have that for an arbitrary y R n (with y ) : Ax A = max Ay x x y Ay A y This holds for y, but we can see it is also true for y =. We can now show property (iv): MN = max MNx max M Nx x =1 x =1 = M max Nx = M N x =1 MN M N Although the definition of an induced norm allowed us to prove certain properties it does not necessarily provide a convenient formula for evaluating the matrix norm. Fortunately, such formulas do exist for the L 1 and L induced matrix norms. Given here without proof: A 1 = max j A = max i A ij max. absolute column sum A ij max. absolute row sum The formula for 2 is much more involved; we will settle with the general definition instead of a closed-form formula in this case. Useful Properties of Matrix and Vector Norms We previously saw that Lecture of: 21 Mar 213 Ax A x (11) for any matrix A and any vector x of dimensions m m and m 1, respectively). Note that, when writing an expression such as 11, the matrix norm A is understood to be the inferred norm from the vector norm used in Ax and x. Thus, Ax 1 A 1 x 1 and Ax A x are both valid but we can not mix and match, e.g.: Ax A 2 x 1 NOT CORRECT