SECTION 8.3: THE INVERSE OF A SQUARE MATRIX



Similar documents
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Using row reduction to calculate the inverse and the determinant of a square matrix

Section 8.2 Solving a System of Equations Using Matrices (Guassian Elimination)

Lecture 2 Matrix Operations

Solving Linear Systems, Continued and The Inverse of a Matrix

Solving Systems of Linear Equations

1.2 Solving a System of Linear Equations

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

DETERMINANTS TERRY A. LORING

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison

Linear Equations ! $ & " % & " 11,750 12,750 13,750% MATHEMATICS LEARNING SERVICE Centre for Learning and Professional Development

5.5. Solving linear systems by the elimination method

7.4. The Inverse of a Matrix. Introduction. Prerequisites. Learning Style. Learning Outcomes

8 Square matrices continued: Determinants

Solving Systems of Linear Equations Using Matrices

Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication).

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Typical Linear Equation Set and Corresponding Matrices

Orthogonal Projections

1 Introduction to Matrices

Linear Algebra Review. Vectors

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

Similarity and Diagonalization. Similar Matrices

6. Cholesky factorization

Direct Methods for Solving Linear Systems. Matrix Factorization

Elementary Matrices and The LU Factorization

MULTIPLICATION AND DIVISION OF REAL NUMBERS In this section we will complete the study of the four basic operations with real numbers.

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Similar matrices and Jordan form

MATH APPLIED MATRIX THEORY

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

CS3220 Lecture Notes: QR factorization and orthogonal transformations

Lecture 5: Singular Value Decomposition SVD (1)

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

Notes on Determinant

Solving Systems of Linear Equations

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices

Question 2: How do you solve a matrix equation using the matrix inverse?

Chapter 6. Orthogonality

LS.6 Solution Matrices

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = i.

University of Lille I PC first year list of exercises n 7. Review

5.3 The Cross Product in R 3

Solution to Homework 2

Lecture 4: Partitioned Matrices and Determinants

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

9 MATRICES AND TRANSFORMATIONS

Data Mining: Algorithms and Applications Matrix Math Review

Introduction to Matrix Algebra

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Linear Algebra: Determinants, Inverses, Rank

LINEAR ALGEBRA. September 23, 2010

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Introduction to Matrices for Engineers

MAT188H1S Lec0101 Burbulla

x = + x 2 + x

Numerical Analysis Lecture Notes

Linear Algebraic Equations, SVD, and the Pseudo-Inverse

How To Understand And Solve A Linear Programming Problem

MATHEMATICS FOR ENGINEERS BASIC MATRIX THEORY TUTORIAL 2

Brief Introduction to Vectors and Matrices

7 Gaussian Elimination and LU Factorization

Algebraic expressions are a combination of numbers and variables. Here are examples of some basic algebraic expressions.

Factorization Theorems

Chapter 7. Matrices. Definition. An m n matrix is an array of numbers set out in m rows and n columns. Examples. (

8.2. Solution by Inverse Matrix Method. Introduction. Prerequisites. Learning Outcomes

Eigenvalues and Eigenvectors

Section 4.1 Rules of Exponents

Rotation Matrices and Homogeneous Transformations

1 Sets and Set Notation.

α = u v. In other words, Orthogonal Projection

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

[1] Diagonal factorization

MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =

1 Solving LPs: The Simplex Algorithm of George Dantzig

Solution of Linear Systems

is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

Solving simultaneous equations using the inverse matrix

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Name: Section Registered In:

Math 312 Homework 1 Solutions

Geometric description of the cross product of the vectors u and v. The cross product of two vectors is a vector! u x v is perpendicular to u and v

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

2.3. Finding polynomial functions. An Introduction:

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

DATA ANALYSIS II. Matrix Algorithms

Continued Fractions and the Euclidean Algorithm

General Framework for an Iterative Solution of Ax b. Jacobi s Method

Lecture notes on linear algebra

Applied Linear Algebra I Review page 1

Operation Count; Numerical Linear Algebra

Vector and Matrix Norms

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Some Lecture Notes and In-Class Examples for Pre-Calculus:

1 Determinants and the Solvability of Linear Systems

A note on companion matrices

Transcription:

(Section 8.3: The Inverse of a Square Matrix) 8.47 SECTION 8.3: THE INVERSE OF A SQUARE MATRIX PART A: (REVIEW) THE INVERSE OF A REAL NUMBER If a is a nonzero real number, then aa 1 = a 1 a = 1. a 1, or 1, is the multiplicative inverse of a, because its product with a is 1, a the multiplicative identity. Example 3 1 3 = 1, so 3 and 1 3 are multiplicative inverses of each other. PART B: THE INVERSE OF A SQUARE MATRIX If A is a square n n matrix, sometimes there exists a matrix A 1 ( A inverse ) such that AA 1 = I n and A 1 A = I n. An invertible matrix and its inverse commute with respect to matrix multiplication. Then, A is invertible (or nonsingular), and A 1 is unique. In this course, an invertible matrix is assumed to be square. Technical Note: A nonsquare matrix may have a left inverse matrix or a right inverse matrix that works on one side of the product and produces an identity matrix. They cannot be the same matrix, however.

(Section 8.3: The Inverse of a Square Matrix) 8.48 PART C : FINDING A -1 We will discuss a shortcut for 2 2 matrices in Part F. Assume that A is a given n n (square) matrix. Example A is invertible Its RRE Form is the identity matrix I n (or simply I). It turns out that a sequence of EROs that takes you from an invertible matrix A down to I will also take you from I down to A 1. (A good Linear Algebra book will have a proof for this.) We can use this fact to efficiently find A 1. We construct A I. We say that A is in the left square of this matrix, and I is in the right square. We apply EROs to A I until we obtain the RRE Form I A 1. That is, as soon as you obtain I in the left square, you grab the matrix in the right square as your A 1. If you ever get a row of 0 s in the left square, then it will be impossible to obtain I A 1, and A is noninvertible (or singular). Let s go back to our A matrix from Section 8.2: Notes 8.44. Find A 1. 1 1 3 A = 1 2 0 1 1 1

(Section 8.3: The Inverse of a Square Matrix) 8.49 Solution We construct A I : 1 1 3 1 2 0 1 1 1 1 0 0 0 1 0 0 0 1 We perform Gauss-Jordan Elimination to take the left square down to I. The right square will be affected in the process, because we perform EROs on entire rows all the way across. We will show a couple of row replacement EROs, and then we will leave the remaining steps to you. We will kill off the purple entries and put 0 s in their places. old R 2 1 2 0 0 1 0 +R 1 1 1 3 1 0 0 new R 2 0 3-3 1 1 0 old R 3 1 1 1 0 0 1 + ( 1) R 1 1 1 3 1 0 0 new R 3 0-2 4-1 0 1 New matrix: 1 1 3 0 3 3 0 2 4 1 0 0 1 1 0 1 0 1 (Your turn! Keep going.)

(Section 8.3: The Inverse of a Square Matrix) 8.50 RRE Form: (Remember, this form is unique.) 1 0 0 0 1 0 0 0 1 I 1/ 3 1/ 3 1 1/ 6 2 / 3 1/ 2 1/ 6 1/ 3 1/ 2 A 1 Check. (Optional) You can check that AA 1 = I. If that holds, then it is automatically true that A 1 A = I. (The right inverse and the left inverse of an invertible matrix must be the same. An invertible matrix must commute with its inverse.)

PART D: USING INVERSES TO SOLVE SYSTEMS (Section 8.3: The Inverse of a Square Matrix) 8.51 In Section 8.2: Notes 8.44, we expressed a system in the matrix form AX = B : 1 1 3 x 1 9 1 2 0 x 2 = 6 1 1 1 x 3 5 A X B A should look familiar. In Part C, we found its inverse, A 1. How can we express X directly in terms of A and B? Review from Algebra I (Optional) Let s say we want to solve ax = b, where a 0 and a and b are real constants. ax = b 1 a a x = 1 a b =1 x = b a Because a 1 represents the multiplicative inverse of a, we can say that a 1 = 1, and the steps can be rewritten as follows: a ax = b a 1 a x = a 1 b =1 x = a 1 b

Solving the Matrix Equation AX = B Assume that A is invertible. (Section 8.3: The Inverse of a Square Matrix) 8.52 Note: Even though 0 is the only real number that is noninvertible (in a multiplicative sense), there are many matrices other than zero matrices that are noninvertible. It is assumed that A, X, and B have compatible sizes. That is, AX is defined, and AX and B have the same size. The steps should look familiar: AX = B A 1 A X = A 1 B = I X = A 1 B The Inverse Matrix Method for Solving a System of Linear Equations If A is invertible, then the system AX = B has a unique solution given by X = A 1 B.

(Section 8.3: The Inverse of a Square Matrix) 8.53 Comments We must left multiply both sides of AX = B by A 1. If we were to right multiply, then we would obtain AXA 1 = BA 1 ; both sides of that equation are undefined, unless A is 1 1. Remember that matrix multiplication is not commutative. Although x = ba 1 would have been acceptable in our Algebra I discussion (because multiplication of real numbers is commutative), X = BA 1 would be inappropriate here. I is the identity matrix that is the same size as A. It plays the role that 1 did in our Algebra I discussion, because 1 was the multiplicative identity for the set of real numbers. This result is of more theoretical significance than practical significance. The Gaussian Elimination (with Back-Substitution) method we discussed earlier is often more efficient than this inversebased process. However, X = A 1 B is good to know if you re using software you re not familiar with. Also, there s an important category of matrices called orthogonal matrices, for which A 1 = A T ; this makes matters a whole lot easier, since A T is trivial to find.

(Section 8.3: The Inverse of a Square Matrix) 8.54 Back to Our Example We will solve the system from Section 8.2: Notes 8.44 using this Inverse Matrix Method. Solution 1 1 3 x 1 9 1 2 0 x 2 = 6 1 1 1 x 3 5 A X B It helps a lot that we ve already found A 1 in Part C; that s the bulk of the work. X = A 1 B 1/ 3 1/ 3 1 9 = 1/ 6 2 / 3 1/ 2 6 1/ 6 1/ 3 1/ 2 5 0 = 3 2 This agrees with our result from the Gauss-Jordan Elimination method we used in Section 8.2: Notes 8.44-8.46.

(Section 8.3: The Inverse of a Square Matrix) 8.55 PART E : THE DETERMINANT OF A 2 2 MATRIX ("BUTTERFLY RULE" ) If A = a c b d, then the determinant of A, denoted by det( A) or A, is given by: det( A) = ad bc i.e., det( A) = ( product along main diagonal) ( product along skew diagonal) The following butterfly image may help you recall this formula. Warning: A should not be confused with absolute value notation. See Section 8.4. We will further discuss determinants in Section 8.4.

(Section 8.3: The Inverse of a Square Matrix) 8.56 PART F : SHORTCUT FORMULA FOR THE INVERSE OF A 2 2 MATRIX If A = a c b d, then: 1 A 1 = det A ( ) d c b a If det( A) = 0, then A 1 does not exist. Remember that we: Switch the entries along the main diagonal. Flip the signs on (i.e., take the opposite of) the entries along the skew diagonal. This formula is consistent with the method from Part C. Example If A = 1 2 3 4, find A 1. Solution First off: Now: det( A) = ( 1) ( 4) ( 2) ( 3) = 2 A 1 = 1 4 2 2 3 1 = 2 1 3 / 2 1/ 2