1 0 5 3 3 A = 0 0 0 1 3 0 0 0 0 0 0 0 0 0 0



Similar documents
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

NOTES ON LINEAR TRANSFORMATIONS

Name: Section Registered In:

Solutions to Math 51 First Exam January 29, 2015

Linear Algebra Notes

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =

Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj

Matrix Representations of Linear Transformations and Changes of Coordinates

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MAT 242 Test 2 SOLUTIONS, FORM T

Similarity and Diagonalization. Similar Matrices

Systems of Linear Equations

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

by the matrix A results in a vector which is a reflection of the given

Methods for Finding Bases

Lecture Notes 2: Matrices as Systems of Linear Equations

T ( a i x i ) = a i T (x i ).

Solving Systems of Linear Equations

18.06 Problem Set 4 Solution Due Wednesday, 11 March 2009 at 4 pm in Total: 175 points.

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

MATH1231 Algebra, 2015 Chapter 7: Linear maps

( ) which must be a vector

160 CHAPTER 4. VECTOR SPACES

These axioms must hold for all vectors ū, v, and w in V and all scalars c and d.

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

THE DIMENSION OF A VECTOR SPACE

Linear Algebra Review. Vectors

4.5 Linear Dependence and Linear Independence

Solving Linear Systems, Continued and The Inverse of a Matrix

Linear Algebra Notes for Marsden and Tromba Vector Calculus

8 Square matrices continued: Determinants

1 Introduction to Matrices

Mathematics Course 111: Algebra I Part IV: Vector Spaces

LINEAR ALGEBRA. September 23, 2010

Section Continued

Orthogonal Projections

Orthogonal Diagonalization of Symmetric Matrices

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

LS.6 Solution Matrices

IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction

Solving Systems of Linear Equations

Math 312 Homework 1 Solutions

MATH2210 Notebook 1 Fall Semester 2016/ MATH2210 Notebook Solving Systems of Linear Equations... 3

α = u v. In other words, Orthogonal Projection

1 VECTOR SPACES AND SUBSPACES

1 Determinants and the Solvability of Linear Systems

Section 8.2 Solving a System of Equations Using Matrices (Guassian Elimination)

1 Sets and Set Notation.

Section Inner Products and Norms

MATH APPLIED MATRIX THEORY

University of Lille I PC first year list of exercises n 7. Review

Recall that two vectors in are perpendicular or orthogonal provided that their dot

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Linearly Independent Sets and Linearly Dependent Sets

LINEAR ALGEBRA W W L CHEN

ISOMETRIES OF R n KEITH CONRAD

Lecture 14: Section 3.3

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

Math 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = i.

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Examination paper for TMA4115 Matematikk 3

3. INNER PRODUCT SPACES

Math 215 HW #6 Solutions

Solution to Homework 2

Solving Systems of Linear Equations Using Matrices

Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication).

[1] Diagonal factorization

Inner Product Spaces and Orthogonality

Subspaces of R n LECTURE Subspaces

x y The matrix form, the vector form, and the augmented matrix form, respectively, for the system of equations are

x = + x 2 + x

Notes on Determinant

Solutions to Homework Section 3.7 February 18th, 2005

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Using row reduction to calculate the inverse and the determinant of a square matrix

Vector Spaces 4.4 Spanning and Independence

Question 2: How do you solve a matrix equation using the matrix inverse?

Chapter 20. Vector Spaces and Bases

Vector and Matrix Norms

Chapter 17. Orthogonal Matrices and Symmetries of Space

Inner Product Spaces

5.3 The Cross Product in R 3

Chapter 6. Orthogonality

Data Mining: Algorithms and Applications Matrix Math Review

Lecture 5: Singular Value Decomposition SVD (1)

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

MA106 Linear Algebra lecture notes

is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

Similar matrices and Jordan form

Introduction to Matrix Algebra

Applied Linear Algebra I Review page 1

Lecture notes on linear algebra

MAT188H1S Lec0101 Burbulla

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Brief Introduction to Vectors and Matrices

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

Direct Methods for Solving Linear Systems. Matrix Factorization

6. Cholesky factorization

Lecture 2 Matrix Operations

Transcription:

Solutions: Assignment 4.. Find the redundant column vectors of the given matrix A by inspection. Then find a basis of the image of A and a basis of the kernel of A. 5 A The second and third columns are mutliples of the first. And the fifth column is times the third column minus times the first. So the second, third, and fifth columns are redundant. And a basis for the image is just and. To get a basis for the kernel we look at Ax. This tells us that relationship between entries of x are just x 5x x 4 +x 5 and x 4 x 5. So this gives us a basis of elements: x x x 5 x x x 5 5 x x x 5.. Find the reduced row-echelon form of the given matrix A. Then find a basis of the image of A and a basis for the kernel of A. 4 8 A 4 5 7 9 4 8 4 4 4 4 5 4 5 5 5 7 9 7 9 5 5 5 5 6 5 There are leading ones in thefirst two columns of rref(a), So a basis for 4 the image is 4 and 5. We know that x is in the kernel is equivalent 7 9 to Ax. That s equivalent to rref(a)x. Which is equivalent

to x 6x and x 5x. Which is equivalent to x x 6 5 is a basis for the kernel. 6 5. So.. A subspace V of R n is called a hyperplane if V is defined by the homogeneous equation c x + c x +... + c n x n where at least one of the coefficients c i is nonzero. What is the dimension of a hyperplane in R n? Justify your answer carefully. What is a hyperplane in R? What is it in R. Define C c c... c n. Notice that v is in V exactly when Cv. So ker(c) V. Call the first nonzero entry of C by k. Then rref(c) k C. So rref(c) has one leading nonzero, i.e. rank(c). By the Rank-Nullity Theorem, dim(ker(c)) + rk(c) n. So dim(v ) dim(ker(c) n rk(c) n. So in R, a hyperplane is dimensional, or a plane. In R, a hyperplane is dimensional it s a line...9 We are told that a certain 5 5 matrix A can be written as A BC where B is 5 4 and C is 4 5. Explain how you know that A is not invertible. Since C is 4 5, we know that im(c) is a subspace of R 4. So dim(im(c)) 4. By the Rank-Nullity theorem, we know dim(ker(c)) + dim(im(c)) 5. This means dim(ker(c)). So there is some vector v in the kernel of C. Now for any 5 5 matrix E at all, EAv EBCv EB v. So no E can make it true that EA I 5. In other words, A is not invertible...5 Find a basis for the row space of the matrix A 4 5 7 Row reducing does not change the row space. So we want to row reduce A: 4 4 6 5 7 4 6

The two remaining nonzero rows arenot multiples of each other. So a basis for the row space is and...56 An n n matrix A is called nilpotent if A m for some positive integer m. Examples are triangular matrices whose entries on the diagonal are all. Consider a nilpotent n n matrix A and choose the small number m such that A m. Pick a vector v in R n such that A m v. Show that the vectors v, Av, A v,..., A m v are linearly independent. Suppose that c v + c Av + c A v +... + c m A m v If all the c s before c i were, we will show that c i also. This means that all the c s must be. Consider the situation where all the c s before c i are. Then we get: c i A i v + c i+ A i+ v +... + c m A m v If we multiply this equation by A m i we get: c i A m v + c i+ A m v +... + c m A m i v c i A m v + A m ( c i+ v + c i+ Av +... + c m A m i v ) c i A m v + ( c i+ v + c i+ Av +... + c m A m i v ) c i A m v But we picked v so that A m v. Thus c i must also be, which results in all the c s being. So the vectors v, Av, A v,..., A m v are linearly independent..4. Find the matrix of the linear transformation T (x) Ax with respect to the basis B (v, v ). A ; v, v B S AS.4.6 Find the matrix of the linear transformation T (x) Ax with respect to the basis B (v,..., v m ). A ; v 4, v

B S AS 4 4 4 6.4.8 Find a basis B of R n such that the B-matrix B of the given linear transformation T is diagonal. T is the reflection about the line in R spanned by. We want to pick vectors v so that T (v) c v v for some c v. that way, the off-diagonal ( ) entries of B will be zero. Vectors on the line don t move, so T. Also, if we pick something perpendicular to the ( ) line, its reflection is just its negation. So T ( ). So according to the basis we will get B which is of the form we want. Ch. Ex. True or False? B, The span of vectors v, v,..., v n consists of all linear combinations of vectors v, v,.... True. This is the definition of span. 7 If u + v + 4w 5u + 6v + 7w, then the vectors u, v, w must be linearly dependent. True. w u v, so w is redundant. If A is a 5 6 matrix of rank 4, then the nullity of A is. False. We should get rank plus nullity equalling 6, not 5. a 8 The vectors of the form b (where a and b are arbitrary real a numbers) form a subspace of R 4. True. This set includes and is closed under both addition and multiplication. 4

4.. Is the following subset of P a subsapce? Find a basis for it if it is. {p(t) p() } Yes. () so is included. If p () and p () then (p + p )() p () + p () + so it is closed under addition. And if p() then (rp)() rp() r, so it is closed under scalar multiplication. Since dim (P ) and is not in the subspace, we know the dimension of the subspace must be at most. Since x and x 4 are in the subspace and linearly independent, they must be a basis. 4..6 Is this subset of R a subspace? The invertible matrices No. This set actually fails all three requirements of a subspace. It does not have. It has both I and I, but it doesn t I + ( I ), so it isn t closed under addition. It has I but it does not have I, so it is not closed under multiplication either. a b 4.. Find a basis for the space of all matrices A in R c d such that a d. If A is as written above, then A a + b + c As those three matrices are linearly independent, we get a basis of,, So the dimension is. 4..4 If B is a diagonal 4 4 matrix, what are the possible dimensions of the space of all 4 4 matrices A that commute with B. b B b b b 4 We need to have that (AB) ij (BA) ij. This tells us that A ij b j b i A ij. This tells us (b j b i )A ij. I.e. b i b j or A ij. So we can make a basis out of matrices E ij which have a in the ijth position and elsewhere, only including the E ij s for which b i b j. If all the b s are the same, we include all 6 E s. If of the b s are the same, we include the 9 E s for those along with the E for the remaining B with itself. This totals to. If there are two pairs of b s, then inside for each pair we have 4 E s. If there is one pair, and two other values, we get 4 E s for the pair plus one for each of the others, totalling 6. And finally, if all for b s are different, we get one E value for each, for 4 total. So the possible dimensions of this space are 4, 6, or 6. 5