# LS.6 Solution Matrices

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 LS.6 Solution Matrices In the literature, solutions to linear systems often are expressed using square matrices rather than vectors. You need to get used to the terminology. As before, we state the definitions and results for a 2 2 system, but they generalize immediately to n n systems.. Fundamental matrices. We return to the system () x = A(t) x, with the general solution (2) x = c x (t) + c 2 x 2 (t), where x and x 2 are two independent solutions to (), and c and c 2 are arbitrary constants. We form the matrix whose columns are the solutions x and x 2 : x x (3) X(t) = ( x 2 x 2 ) =. y y 2 Since the solutions are linearly independent, we called them in LS.5 a fundamental set of solutions, and therefore we call the matrix in (3) a fundamental matrix for the system (). Writing the general solution using X(t). As a first application of X(t), we can use it to write the general solution (2) efficiently. For according to (2), it is x x2 x x x = c + c 2 = 2 c, y y 2 y y 2 c 2 which becomes using the fundamental matrix c (4) x = X(t) c where c =, (general solution to ()). c 2 Note that the vector c must be written on the right, even though the c s are usually written on the left when they are the coefficients of the solutions x i. Solving the IVP using X(t). We can now write down the solution to the IVP (5) x = A(t) x, x(t 0 ) = x 0. Starting from the general solution (4), we have to choose the c so that the initial condition in (6) is satisfied. Substituting t 0 into (5) gives us the matrix equation for c : X(t 0 ) c = x 0. Since the determinant X(t 0 ) is the value at t 0 of the Wronskian of x amd x 2, it is non-zero since the two solutions are linearly independent (Theorem 5.2C). Therefore the inverse matrix exists (by LS.), and the matrix equation above can be solved for c: c = X(t 0 ) x 0 ; using the above value of c in (4), the solution to the IVP () can now be written (6) x = X(t)X(t 0 ) x 0. 25

2 NOTES: LS. LINEAR SYSTEMS Note that when the solution is written in this form, it s obvious that x(t 0 ) = x 0, i.e., that the initial condition in (5) is satisfied. An equation for fundamental matrices We have been saying a rather than the fundamental matrix since the system () doesn t have a unique fundamental matrix: thare are many different ways to pick two independent solutions of x = A x to form the columns of X. It is therefore useful to have a way of recognizing a fundamental matrix when you see one. The following theorem is good for this; we ll need it shortly. Theorem 6. X(t) is a fundamental matrix for the system () if its determinant X(t) is non-zero and it satisfies the matrix equation (7) X = A X, where X means that each entry of X has been differentiated. Proof. Since X 0, its columns x and x 2 are linearly independent, by section LS.5. And writing X = ( x x 2 ), (7) becomes, according to the rules for matrix multiplication, ( x x 2 ) = A ( x x 2 ) = ( Ax Ax 2 ), which shows that x = A x and x 2 = A x 2 ; this last line says that x and x 2 are solutions to the system (). 2. The normalized fundamental matrix. Is there a best choice for fundamental matrix? There are two common choices, each with its advantages. If the ODE system has constant coefficients, and its eigenvalues are real and distinct, then a natural choice for the fundamental matrix would be the one whose columns are the normal modes the solutions of the form x i = i e i t, i =, 2. There is another choice however which is suggested by (6) and which is particularly useful in showing how the solution depends on the initial conditions. Suppose we pick X(t) so that 0 (8) X(t 0 ) = I =. Referring to the definition (3), this means the solutions x and x 2 are picked so (8 ) x (t 0 ) = 0 0, x 2 (t 0 ) =. Since the x i (t) are uniquely determined by these initial conditions, the fundamental matrix X(t) satisfying (8) is also unique; we give it a name. Definition 6.2 The unique matrix X t0 (t) satisfying (9) X = A X t0, X t0 (t 0 ) = I t 0

3 LS.6 SOLUTION MATRICES 27 is called the normalized fundamental matrix at t 0 for A. For convenience in use, the definition uses Theorem 6. to guarantee X t0 will actually be a fundamental matrix; the condition X t0 (t) = 0 in Theorem 6. is satisfied, since the definition implies X t0 (t 0 ) =. To keep the notation simple, we will assume in the rest of this section that t 0 = 0, as it almost always is; then X 0 is the normalized fundamental matrix. Since X 0 (0) = I, we get from (6) the matrix form for the solution to an IVP: (0) The solution to the IVP x = A(t) x, x(0) = x 0 is x(t) = X 0 (t)x 0. Calculating X 0. One way is to find the two solutions in (8 ), and use them as the columns of X 0. This is fine if the two solutions can be determined by inspection. If not, a simpler method is this: find any fundamental matrix X(t); then () X 0 (t) = X(t) X(0). To verify this, we have to see that the matrix on the right of () satisfies the two conditions in Definition 6.2. The second is trivial; the first is easy using the rule for matrix differentiation: If M = M(t) and B, C are constant matrices, then (BM) = BM, (MC) = M C, from which we see that since X is a fundamental matrix, (X(t)X(0) ) = X(t) X(0) = AX(t)X(0) = A(X(t)X(0) ), showing that X(t)X(0) also satisfies the first condition in Definition 6.2. Example 6.2A Find the solution to the IVP: x = x, x(0) = x 0. 0 Solution Since the system is x = y, y = x, we can find by inspection the fundamental set of solutions satisfying (8 ) : x = cos t x = sin t and. y = sin t y = cos t Thus by (0) the normalized fundamental matrix at 0 and solution to the IVP is cos t sin t x 0 cos t sin t x = X x 0 = = x 0 + y 0. sin t cos t y 0 sin t cos t 3 Example 6.2B Give the normalized fundamental matrix at 0 for x = x. Solution. This time the solutions (8 ) cannot be obtained by inspection, so we use the second method. We calculated the normal modes for this sytem at the beginning of LS.2; using them as the columns of a fundamental matrix gives us 3e 2t e X(t) = 2t 2t 2t. e e

4 NOTES: LS. LINEAR SYSTEMS Using () and the formula for calculating the inverse matrix given in LS., we get 3 X(0) =, X(0) = 4, 3 so that 2t e 2t 2t 2t 3e + e 2t 3e 2t 3e X(t) 3e = 4 2t 2t = 2t e 2t + 3e 2t. e e 3 4 e 2t e 6.3 The Exponential matrix. The work in the preceding section with fundamental matrices was valid for any linear homogeneous square system of ODE s, x = A(t) x. However, if the system has constant coefficients, i.e., the matrix A is a constant matrix, the results are usually expressed by using the exponential matrix, which we now define. Recall that if x is any real number, then 2 n x x (2) e x = + x ! n! Definition 6.3 Given an n n constant matrix A, the exponential matrix e A is the n n matrix defined by (3) e A = I + A ! n! A 2 Each term on the right side of (3) is an n n matrix; adding up the ij-th entry of each of these matrices gives you an infinite series whose sum is the ij-th entry of e A. (The series always converges.) In the applications, an independent variable t is usually included: A n (4) e = I + A t + A 2 t A n tn ! n! This is not a new definition, it s just (3) above applied to the matrix A t in which every element of A has been multiplied by t, since for example 2 () 2 = = A A t 2 = A 2 t. Try out (3) and (4) on these two examples; the first is worked out in your book (Example 2, p. 47); the second is easy, since it is not an infinite series. at a 0 e 0 a 0 e Example 6.3A Let A =, show: e A = ; e = 0 b 0 e b 0 e bt t Example 6.3B Let A =, show: e 0 0 A = ; e = What s the point of the exponential matrix? The answer is given by the theorem below, which says that the exponential matrix provides a royal road to the solution of a square

5 LS.6 SOLUTION MATRICES 29 system with constant coefficients: no eigenvectors, no eigenvalues, you just write down the answer! Theorem 6.3 Let A be a square constant matrix. Then (5) (a) e = X 0 (t), the normalized fundamental matrix at 0; (6) (b) the unique solution to the IVP x = Ax, x(0) = x 0 is x = e x 0. Proof. Statement (6) follows immediately from (5), in view of (0). We prove (5) is true, by using the description of a normalized fundamental matrix given in Definition 6.2: letting X = e, we must show X = AX and X(0) = I. The second of these follows from substituting t = 0 into the infinite series definition (4) for e. To show X = AX, we assume that we can differentiate the series (4) term-by-term; then we have for the individual terms d t n t n A n = A n, dt n! (n )! since A n is a constant matrix. Differentiating (4) term-by-term then gives dx d = e = A + A 2 t A n t n +... (8) dt dt (n )! = A e = A X. Calculation of e. The main use of the exponential matrix is in (6) writing down explicitly the solution to an IVP. If e has to be actually calculated for a specific system, several techniques are available. a) In simple cases, it can be calculated directly as an infinite series of matrices. b) It can always be calculated, according to Theorem 6.3, as the normalized fundamental matrix X 0 (t), using (): X 0 (t) = X(t)X(0). c) A third technique uses the exponential law Bt Ct (9) e (B+C)t = e e, valid if BC = CB. To use it, one looks for constant matrices B and C such that Bt (20) A = B + C, BC = CB, e and e Ct are computable; then B t C t (2) e = e e. 2 Example 6.3C Let A =. Solve x = A x, x(0) =, using e Solution. We set B = and C = ; then (20) is satisfied, and

6 NOTES: LS. LINEAR SYSTEMS e = e 2t 0 t 2t t 0 e 2t = e, by (2) and Examples 6.3A and 6.3B. Therefore, by (6), we get 2t t 2t + 2t x = e x 0 = e = e. 2 2 Exercises: Sections 4G,H

### Solutions to Linear First Order ODE s

First Order Linear Equations In the previous session we learned that a first order linear inhomogeneous ODE for the unknown function x = x(t), has the standard form x + p(t)x = q(t) () (To be precise we

### 1 Gaussian Elimination

Contents 1 Gaussian Elimination 1.1 Elementary Row Operations 1.2 Some matrices whose associated system of equations are easy to solve 1.3 Gaussian Elimination 1.4 Gauss-Jordan reduction and the Reduced

### Linear Dependence Tests

Linear Dependence Tests The book omits a few key tests for checking the linear dependence of vectors. These short notes discuss these tests, as well as the reasoning behind them. Our first test checks

### Basic Terminology for Systems of Equations in a Nutshell. E. L. Lady. 3x 1 7x 2 +4x 3 =0 5x 1 +8x 2 12x 3 =0.

Basic Terminology for Systems of Equations in a Nutshell E L Lady A system of linear equations is something like the following: x 7x +4x =0 5x +8x x = Note that the number of equations is not required

### 1 Eigenvalues and Eigenvectors

Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x

### C 1 x(t) = e ta C = e C n. 2! A2 + t3

Matrix Exponential Fundamental Matrix Solution Objective: Solve dt A x with an n n constant coefficient matrix A x (t) Here the unknown is the vector function x(t) x n (t) General Solution Formula in Matrix

### r (t) = 2r(t) + sin t θ (t) = r(t) θ(t) + 1 = 1 1 θ(t) 1 9.4.4 Write the given system in matrix form x = Ax + f ( ) sin(t) x y 1 0 5 z = dy cos(t)

Solutions HW 9.4.2 Write the given system in matrix form x = Ax + f r (t) = 2r(t) + sin t θ (t) = r(t) θ(t) + We write this as ( ) r (t) θ (t) = ( ) ( ) 2 r(t) θ(t) + ( ) sin(t) 9.4.4 Write the given system

### 22 Matrix exponent. Equal eigenvalues

22 Matrix exponent. Equal eigenvalues 22. Matrix exponent Consider a first order differential equation of the form y = ay, a R, with the initial condition y) = y. Of course, we know that the solution to

### Chapter 4 - Systems of Equations and Inequalities

Math 233 - Spring 2009 Chapter 4 - Systems of Equations and Inequalities 4.1 Solving Systems of equations in Two Variables Definition 1. A system of linear equations is two or more linear equations to

### 3.7 Non-autonomous linear systems of ODE. General theory

3.7 Non-autonomous linear systems of ODE. General theory Now I will study the ODE in the form ẋ = A(t)x + g(t), x(t) R k, A, g C(I), (3.1) where now the matrix A is time dependent and continuous on some

### Matrices: 2.3 The Inverse of Matrices

September 4 Goals Define inverse of a matrix. Point out that not every matrix A has an inverse. Discuss uniqueness of inverse of a matrix A. Discuss methods of computing inverses, particularly by row operations.

### We know a formula for and some properties of the determinant. Now we see how the determinant can be used.

Cramer s rule, inverse matrix, and volume We know a formula for and some properties of the determinant. Now we see how the determinant can be used. Formula for A We know: a b d b =. c d ad bc c a Can we

### 1 2 3 1 1 2 x = + x 2 + x 4 1 0 1

(d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which

### by the matrix A results in a vector which is a reflection of the given

Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

### UNIT 2 MATRICES - I 2.0 INTRODUCTION. Structure

UNIT 2 MATRICES - I Matrices - I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress

### SYSTEMS OF EQUATIONS

SYSTEMS OF EQUATIONS 1. Examples of systems of equations Here are some examples of systems of equations. Each system has a number of equations and a number (not necessarily the same) of variables for which

### Similar matrices and Jordan form

Similar matrices and Jordan form We ve nearly covered the entire heart of linear algebra once we ve finished singular value decompositions we ll have seen all the most central topics. A T A is positive

### Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Pearson Education, Inc.

2 Matrix Algebra 2.3 CHARACTERIZATIONS OF INVERTIBLE MATRICES Theorem 8: Let A be a square matrix. Then the following statements are equivalent. That is, for a given A, the statements are either all true

### Matrix Methods for Linear Systems of Differential Equations

Matrix Methods for Linear Systems of Differential Equations We now present an application of matrix methods to linear systems of differential equations. We shall follow the development given in Chapter

### MAT Solving Linear Systems Using Matrices and Row Operations

MAT 171 8.5 Solving Linear Systems Using Matrices and Row Operations A. Introduction to Matrices Identifying the Size and Entries of a Matrix B. The Augmented Matrix of a System of Equations Forming Augmented

### 4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows

### 4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

### Systems of Linear Equations

Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

### Solutions to Linear Algebra Practice Problems

Solutions to Linear Algebra Practice Problems. Find all solutions to the following systems of linear equations. (a) x x + x 5 x x x + x + x 5 (b) x + x + x x + x + x x + x + 8x Answer: (a) We create the

### Notes on Determinant

ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

### CHAPTER 2. Eigenvalue Problems (EVP s) for ODE s

A SERIES OF CLASS NOTES FOR 005-006 TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS DE CLASS NOTES 4 A COLLECTION OF HANDOUTS ON PARTIAL DIFFERENTIAL EQUATIONS

### B such that AB = I and BA = I. (We say B is an inverse of A.) Definition A square matrix A is invertible (or nonsingular) if matrix

Matrix inverses Recall... Definition A square matrix A is invertible (or nonsingular) if matrix B such that AB = and BA =. (We say B is an inverse of A.) Remark Not all square matrices are invertible.

### DETERMINANTS. b 2. x 2

DETERMINANTS 1 Systems of two equations in two unknowns A system of two equations in two unknowns has the form a 11 x 1 + a 12 x 2 = b 1 a 21 x 1 + a 22 x 2 = b 2 This can be written more concisely in

### Definition A square matrix M is invertible (or nonsingular) if there exists a matrix M 1 such that

0. Inverse Matrix Definition A square matrix M is invertible (or nonsingular) if there exists a matrix M such that M M = I = M M. Inverse of a 2 2 Matrix Let M and N be the matrices: a b d b M =, N = c

### 4.1 VECTOR SPACES AND SUBSPACES

4.1 VECTOR SPACES AND SUBSPACES What is a vector space? (pg 229) A vector space is a nonempty set, V, of vectors together with two operations; addition and scalar multiplication which satisfies the following

### Math 2280 Section 002 [SPRING 2013] 1

Math 2280 Section 002 [SPRING 2013] 1 Today well learn about a method for solving systems of differential equations, the method of elimination, that is very similar to the elimination methods we learned

### Presentation 3: Eigenvalues and Eigenvectors of a Matrix

Colleen Kirksey, Beth Van Schoyck, Dennis Bowers MATH 280: Problem Solving November 18, 2011 Presentation 3: Eigenvalues and Eigenvectors of a Matrix Order of Presentation: 1. Definitions of Eigenvalues

### Section 2.1. Section 2.2. Exercise 6: We have to compute the product AB in two ways, where , B =. 2 1 3 5 A =

Section 2.1 Exercise 6: We have to compute the product AB in two ways, where 4 2 A = 3 0 1 3, B =. 2 1 3 5 Solution 1. Let b 1 = (1, 2) and b 2 = (3, 1) be the columns of B. Then Ab 1 = (0, 3, 13) and

### MATH 2030: SYSTEMS OF LINEAR EQUATIONS. ax + by + cz = d. )z = e. while these equations are not linear: xy z = 2, x x = 0,

MATH 23: SYSTEMS OF LINEAR EQUATIONS Systems of Linear Equations In the plane R 2 the general form of the equation of a line is ax + by = c and that the general equation of a plane in R 3 will be we call

### Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

### Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices Solving square systems of linear equations; inverse matrices. Linear algebra is essentially about solving systems of linear equations,

### [1] Diagonal factorization

8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:

### A Second Course in Elementary Differential Equations: Problems and Solutions. Marcel B. Finan Arkansas Tech University c All Rights Reserved

A Second Course in Elementary Differential Equations: Problems and Solutions Marcel B. Finan Arkansas Tech University c All Rights Reserved Contents 8 Calculus of Matrix-Valued Functions of a Real Variable

### 1.5 Elementary Matrices and a Method for Finding the Inverse

.5 Elementary Matrices and a Method for Finding the Inverse Definition A n n matrix is called an elementary matrix if it can be obtained from I n by performing a single elementary row operation Reminder:

### ( ) which must be a vector

MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are

### MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

### Lecture L19 - Vibration, Normal Modes, Natural Frequencies, Instability

S. Widnall 16.07 Dynamics Fall 2009 Version 1.0 Lecture L19 - Vibration, Normal Modes, Natural Frequencies, Instability Vibration, Instability An important class of problems in dynamics concerns the free

### Matrix Calculations: Inverse and Basis Transformation

Matrix Calculations: Inverse and asis Transformation A. Kissinger (and H. Geuvers) Institute for Computing and Information ciences Intelligent ystems Version: spring 25 A. Kissinger (and H. Geuvers) Version:

### CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in

### Linearly Independent Sets and Linearly Dependent Sets

These notes closely follow the presentation of the material given in David C. Lay s textbook Linear Algebra and its Applications (3rd edition). These notes are intended primarily for in-class presentation

### Methods for Finding Bases

Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,

### (4.8) Solving Systems of Linear DEs by Elimination

INTRODUCTION: (4.8) Solving Systems of Linear DEs by Elimination Simultaneous dinary differential equations involve two me equations that contain derivatives of two me dependent variables the unknown functions

### 2.1: Determinants by Cofactor Expansion. Math 214 Chapter 2 Notes and Homework. Evaluate a Determinant by Expanding by Cofactors

2.1: Determinants by Cofactor Expansion Math 214 Chapter 2 Notes and Homework Determinants The minor M ij of the entry a ij is the determinant of the submatrix obtained from deleting the i th row and the

### Inverses and powers: Rules of Matrix Arithmetic

Contents 1 Inverses and powers: Rules of Matrix Arithmetic 1.1 What about division of matrices? 1.2 Properties of the Inverse of a Matrix 1.2.1 Theorem (Uniqueness of Inverse) 1.2.2 Inverse Test 1.2.3

### Chapter 20. Vector Spaces and Bases

Chapter 20. Vector Spaces and Bases In this course, we have proceeded step-by-step through low-dimensional Linear Algebra. We have looked at lines, planes, hyperplanes, and have seen that there is no limit

### Chapter 8. Matrices II: inverses. 8.1 What is an inverse?

Chapter 8 Matrices II: inverses We have learnt how to add subtract and multiply matrices but we have not defined division. The reason is that in general it cannot always be defined. In this chapter, we

### 1. LINEAR EQUATIONS. A linear equation in n unknowns x 1, x 2,, x n is an equation of the form

1. LINEAR EQUATIONS A linear equation in n unknowns x 1, x 2,, x n is an equation of the form a 1 x 1 + a 2 x 2 + + a n x n = b, where a 1, a 2,..., a n, b are given real numbers. For example, with x and

### SECOND-ORDER LINEAR HOMOGENEOUS DIFFERENTIAL EQUATIONS

L SECOND-ORDER LINEAR HOOGENEOUS DIFFERENTIAL EQUATIONS SECOND-ORDER LINEAR HOOGENEOUS DIFFERENTIAL EQUATIONS WITH CONSTANT COEFFICIENTS A second-order linear differential equation is one of the form d

### 4 Solving Systems of Equations by Reducing Matrices

Math 15 Sec S0601/S060 4 Solving Systems of Equations by Reducing Matrices 4.1 Introduction One of the main applications of matrix methods is the solution of systems of linear equations. Consider for example

### MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

### MATH 551 - APPLIED MATRIX THEORY

MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

### Chapter 6. Orthogonality

6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

### Question 2: How do you solve a matrix equation using the matrix inverse?

Question : How do you solve a matrix equation using the matrix inverse? In the previous question, we wrote systems of equations as a matrix equation AX B. In this format, the matrix A contains the coefficients

### Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two

### Determinants. Dr. Doreen De Leon Math 152, Fall 2015

Determinants Dr. Doreen De Leon Math 52, Fall 205 Determinant of a Matrix Elementary Matrices We will first discuss matrices that can be used to produce an elementary row operation on a given matrix A.

### University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients

### Sec 4.1 Vector Spaces and Subspaces

Sec 4. Vector Spaces and Subspaces Motivation Let S be the set of all solutions to the differential equation y + y =. Let T be the set of all 2 3 matrices with real entries. These two sets share many common

### Chapter 12 Modal Decomposition of State-Space Models 12.1 Introduction The solutions obtained in previous chapters, whether in time domain or transfor

Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology 1 1 c Chapter

### Lecture 14: Section 3.3

Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in

### Inner Product Spaces

Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

### LINEAR ALGEBRA W W L CHEN

LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,

### Similarity and Diagonalization. Similar Matrices

MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

### ANALYTICAL MATHEMATICS FOR APPLICATIONS 2016 LECTURE NOTES Series

ANALYTICAL MATHEMATICS FOR APPLICATIONS 206 LECTURE NOTES 8 ISSUED 24 APRIL 206 A series is a formal sum. Series a + a 2 + a 3 + + + where { } is a sequence of real numbers. Here formal means that we don

### ( % . This matrix consists of \$ 4 5 " 5' the coefficients of the variables as they appear in the original system. The augmented 3 " 2 2 # 2 " 3 4&

Matrices define matrix We will use matrices to help us solve systems of equations. A matrix is a rectangular array of numbers enclosed in parentheses or brackets. In linear algebra, matrices are important

### NON SINGULAR MATRICES. DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that

NON SINGULAR MATRICES DEFINITION. (Non singular matrix) An n n A is called non singular or invertible if there exists an n n matrix B such that AB = I n = BA. Any matrix B with the above property is called

### Some Notes on Taylor Polynomials and Taylor Series

Some Notes on Taylor Polynomials and Taylor Series Mark MacLean October 3, 27 UBC s courses MATH /8 and MATH introduce students to the ideas of Taylor polynomials and Taylor series in a fairly limited

### Lecture 10: Invertible matrices. Finding the inverse of a matrix

Lecture 10: Invertible matrices. Finding the inverse of a matrix Danny W. Crytser April 11, 2014 Today s lecture Today we will Today s lecture Today we will 1 Single out a class of especially nice matrices

### We seek a factorization of a square matrix A into the product of two matrices which yields an

LU Decompositions We seek a factorization of a square matrix A into the product of two matrices which yields an efficient method for solving the system where A is the coefficient matrix, x is our variable

### MATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix.

MATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix. Inverse matrix Definition. Let A be an n n matrix. The inverse of A is an n n matrix, denoted

### The Method of Partial Fractions Math 121 Calculus II Spring 2015

Rational functions. as The Method of Partial Fractions Math 11 Calculus II Spring 015 Recall that a rational function is a quotient of two polynomials such f(x) g(x) = 3x5 + x 3 + 16x x 60. The method

### Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

### Taylor and Maclaurin Series

Taylor and Maclaurin Series In the preceding section we were able to find power series representations for a certain restricted class of functions. Here we investigate more general problems: Which functions

### 10.3 POWER METHOD FOR APPROXIMATING EIGENVALUES

55 CHAPTER NUMERICAL METHODS. POWER METHOD FOR APPROXIMATING EIGENVALUES In Chapter 7 we saw that the eigenvalues of an n n matrix A are obtained by solving its characteristic equation n c n n c n n...

### Chapter 1 - Matrices & Determinants

Chapter 1 - Matrices & Determinants Arthur Cayley (August 16, 1821 - January 26, 1895) was a British Mathematician and Founder of the Modern British School of Pure Mathematics. As a child, Cayley enjoyed

### Solving Systems of Linear Equations

LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

### Dynamics. Figure 1: Dynamics used to generate an exemplar of the letter A. To generate

Dynamics Any physical system, such as neurons or muscles, will not respond instantaneously in time but will have a time-varying response termed the dynamics. The dynamics of neurons are an inevitable constraint

### Math 54. Selected Solutions for Week Is u in the plane in R 3 spanned by the columns

Math 5. Selected Solutions for Week 2 Section. (Page 2). Let u = and A = 5 2 6. Is u in the plane in R spanned by the columns of A? (See the figure omitted].) Why or why not? First of all, the plane in

### HW6 Solutions. MATH 20D Fall 2013 Prof: Sun Hui TA: Zezhou Zhang (David) November 14, 2013. Checklist: Section 7.8: 1c, 2, 7, 10, [16]

HW6 Solutions MATH D Fall 3 Prof: Sun Hui TA: Zezhou Zhang David November 4, 3 Checklist: Section 7.8: c,, 7,, [6] Section 7.9:, 3, 7, 9 Section 7.8 In Problems 7.8. thru 4: a Draw a direction field and

### Solving Systems of Linear Equations. Substitution

Solving Systems of Linear Equations There are two basic methods we will use to solve systems of linear equations: Substitution Elimination We will describe each for a system of two equations in two unknowns,

### Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions

Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential

### Inverses. Stephen Boyd. EE103 Stanford University. October 27, 2015

Inverses Stephen Boyd EE103 Stanford University October 27, 2015 Outline Left and right inverses Inverse Solving linear equations Examples Pseudo-inverse Left and right inverses 2 Left inverses a number

### 7 - Linear Transformations

7 - Linear Transformations Mathematics has as its objects of study sets with various structures. These sets include sets of numbers (such as the integers, rationals, reals, and complexes) whose structure

### Solving Linear Systems, Continued and The Inverse of a Matrix

, Continued and The of a Matrix Calculus III Summer 2013, Session II Monday, July 15, 2013 Agenda 1. The rank of a matrix 2. The inverse of a square matrix Gaussian Gaussian solves a linear system by reducing

### MATH 304 Linear Algebra Lecture 4: Matrix multiplication. Diagonal matrices. Inverse matrix.

MATH 304 Linear Algebra Lecture 4: Matrix multiplication. Diagonal matrices. Inverse matrix. Matrices Definition. An m-by-n matrix is a rectangular array of numbers that has m rows and n columns: a 11

### Additional Topics in Linear Algebra Supplementary Material for Math 540. Joseph H. Silverman

Additional Topics in Linear Algebra Supplementary Material for Math 540 Joseph H Silverman E-mail address: jhs@mathbrownedu Mathematics Department, Box 1917 Brown University, Providence, RI 02912 USA Contents

### 13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms

### 8 Square matrices continued: Determinants

8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You

### Lecture 6. Inverse of Matrix

Lecture 6 Inverse of Matrix Recall that any linear system can be written as a matrix equation In one dimension case, ie, A is 1 1, then can be easily solved as A x b Ax b x b A 1 A b A 1 b provided that

### 1 Introduction to Matrices

1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

### Lecture L3 - Vectors, Matrices and Coordinate Transformations

S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between

### 4.6 Null Space, Column Space, Row Space

NULL SPACE, COLUMN SPACE, ROW SPACE Null Space, Column Space, Row Space In applications of linear algebra, subspaces of R n typically arise in one of two situations: ) as the set of solutions of a linear

### 10.3 POWER METHOD FOR APPROXIMATING EIGENVALUES

58 CHAPTER NUMERICAL METHODS. POWER METHOD FOR APPROXIMATING EIGENVALUES In Chapter 7 you saw that the eigenvalues of an n n matrix A are obtained by solving its characteristic equation n c nn c nn...

### MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +