On the general equation of the second degree

Save this PDF as:

Size: px
Start display at page:

Download "On the general equation of the second degree"

Transcription

1 On the general equation of the second degree S Kesavan The Institute of Mathematical Sciences, CIT Campus, Taramani, Chennai Abstract We give a unified treatment of the general equation of the second degree in two real variables in terms of the eigenvalues of the matrix associated to the quadratic terms and describe the solution sets in all cases 1

2 2

3 1 Introduction The study of the general equation of the second degree in two variables used to be a major chapter in a course on analytic geometry in the undergraduate mathematics curriculum for a long time The equation usually represents a pair of straight lines or a conic In the latter case the method of tracing a conic was to compute the trigonometric ratios of the angle that the axes of the conic make with the coordinate axes and then rotate the coordinate axes to reduce the equation to the normal form These computations could be tedious Further in most classical text books the treatment is rather incomplete and the cases when the solution set is degenerate (especially when it contains a single point or is empty) are not carefully explained The aim of this note is to study all cases of the equation in a unified manner By just computing the eigenvalues and eigenvectors of the 2 2 real symmetric matrix associated to the quadratic terms, we can just read off the properties of the solution set and also write down the equations of various features of the set very easily This approach neatly brings out some of the connections between linear algebra and geometry 2 Some linear algebra Consider the following 2 2 real symmetric matrix: [ ] a h A = h b Its characteristic equation is The discriminant is λ 2 (a + b)λ + (ab h 2 ) = 0 (a + b) 2 4ab + 4h 2 = (a b) 2 + 4h 2 0 and so both its eigenvalues are real and are given by (a + b) ± (a b) 2 + 4h 2 2 3

4 These eigenvalues are coincident if, and only if, a = b and h = 0 Let us denote these eigenvalues by λ 1 and λ 2 Then we can find an eigenvector u = (u 1, u 2 ) associated to λ 1 and an eigenvector v = (v 1, v 2 ) associated to λ 2 If (, ) denotes the usual euclidean scalar product in R 2, we have λ 1 (u, v) = (λ 1 u, v) = (Au, v) = au 1 v 1 + h(u 2 v 1 + u 1 v 2 ) + bu 2 v 2 = (u, Av) = λ 2 (u, v) If λ 1 λ 2, it then follows that (u, v) = 0 (21) If λ 1 = λ 2, then, as we already observed, A = ai, where I is the 2 2 identity matrix, and so every vector in R 2 is an eigenvector and we can always choose the two eigenvectors u = (1, 0) and v = (0, 1) so that (21) is still valid Thus, we always have a pair of orthogonal eigenvectors Let us normalize them so that their euclidean norms are unity Thus let u and v satisfy Now define u u 2 2 = 1 = v v 2 2 P = [ u1 v 1 u 2 v 2 Then we have P P T = P T P = I (where by B T we denote the transpose of a given matrix B) Thus P is an orthogonal matrix Now [ ] [ ] [ ] λ1 u AP = 1 λ 2 v 1 u1 v = 1 λ1 0 λ 1 u 2 λ 2 v 2 u 2 v 2 0 λ 2 Thus, if we have D = [ λ1 0 0 λ 2 ] ], AP = P D, or equivalently A = P DP T Remark This is a particular case (when n = 2) of the following general result: If A is an n n real symmetric matrix, then there exists an orthonormal basis of eigenvectors; if P is the orthogonal matrix whose columns are these eigenvectors and if D is the diagonal matrix whose diagonal entries are the 4

5 eigenvalues of A (in the same order corresponding to the column vectors of P ), then A = P DP T The same result is true if A is a complex hermitian (ie self-adjoint) matrix, in which case we replace P T in the preceding relation by P, the conjugate transpose of P (and P will be a unitary matrix) Example 21 Let A = [ Then, its characteristic equation is λ 2 10λ + 16 = 0 and its eigenvalues are, therefore, λ 1 = 2 and λ 2 = 8 Corresponding to λ 1 = 2, we get the equation 5x 3y = 2x or, equivalently, x = y Thus the eigenvectors corresponding to λ 1 are scalar multiples of (1,1) Normalizing, we get (u 1, u 2 ) = ( 1 2, ] ) 1 2 Corresponding to the eigenvalue λ 2 = 8, we get the equation 5x 3y = 8x, or, equivalently, x = y Thus all eigenvectors corresponding to this eigenvalue are scalar multiples of (1, 1) and, normalizing, we get ( 1 (v 1, v 2 ) = 2, 1 ) 2 Thus P = [ ] and it is easy to verify that A = P DP T where [ ] 2 0 D = The homogeneous equation Consider the following homogeneous equation of the second degree in two real variables: ax 2 + 2hxy + by 2 = 0 (31) Let us denote by S the set of all points (x, y) in the plane which satisfy this equation Our aim is to determine this set 5

6 We will set [ ] a h A = h b and use the notations developed in the preceding section The above equation can be written in matrix form as [ ] [ ] a h x [ x y ] = 0 h b y Writing A = P DP T as in the previous section, let us define [ ] [ ] x x y = P T y Thus x = u 1 x + u 2 y, y = v 1 x + v 2 y Then the equation (31) reduces to [ ] x [ x y ]D = λ 1 x 2 + λ 2 y 2 = 0 (32) We can now easily determine S from (32) y If λ 1 > 0 and λ 2 > 0 or if λ 1 < 0 and λ 2 < 0, then the only solution will be x = y = 0 Then it follows that x = y = 0 and so S = {(0, 0)} Let λ 1 > 0 and λ 2 < 0, or λ 1 < 0 and λ 2 > 0 Then λ 2 λ 1 > 0 and we get from (32) that x = ± λ 2 y λ 1 Thus S consists of a pair of lines passing through the origin given by u 1 x + u 2 y = λ 2 λ 1 (v 1 x + v 2 y), u 1 x + u 2 y = λ 2 λ 1 (v 1 x + v 2 y) If λ 1 = 0 and λ 2 0 or λ 1 0 and λ 2 = 0, we get the solution as y = 0 or x = 0, respectively Thus we have S to consist of a single line given by v 1 x + v 2 y = 0, if λ 1 = 0, λ 2 0, 6

7 and u 1 x + u 2 y = 0, if λ 1 0, λ 2 = 0 The case λ 1 = λ 2 = 0 is excluded since in this case we have a = b = h = 0 and so the equation is vacuous Finally, if λ 1 = λ 2 0, the equation (32) reduces to x 2 y 2 = 0 In this case we get two perpendicular lines x + y = 0 and x y = 0 which are given in the original coordinates by (u 1 x + u 2 y) + (v 1 x + v 2 y) = 0, (u 1 x + u 2 y) (v 1 x + v 2 y) = 0 Thus by computing the eigenvalues and normalized eigenvectors of the matrix A, we can immediately explicitly describe the solution set S The results derived above can be summarized in the following figure 2 lines r lines 2 lines one line λ 2 one line λ 1 > 0, λ 2 > 0 S = {(0, 0)} one line λ 1 2 lines S = {(0, 0)} λ 1 < 0, λ 2 < 0 Figure 1 r lines 2 lines one line 7

8 4 The inhomogeneous equation Let us now consider the inhomogeneous equation ax 2 + 2hxy + by 2 = 1 (41) With the notations established in the preceding sections, this equation reduces to λ 1 x 2 + λ 2 y 2 = 1 (42) If S is the solution set, then we have the following cases Once again the case λ 1 = λ 2 = 0 is excluded since then we have that a = b = h = 0 and the equation (41) is meaningless If λ 1 = λ 2 > 0, ie if a = b > 0 and h = 0, then S is a circle centered at the origin with radius 1 a If λ 1 = λ 2 < 0, ie a = b < 0, h = 0, then S = More generally, if λ 1 0 and λ 2 0, we have S = If λ 1 > 0 and λ 2 > 0, then the equation represents an ellipse with centre at the origin The lengths of the semi-axes of the ellipse are 1 λ1 and 1 λ2 The equations of the respective axes are y = 0 and x = 0, which can be written in the original coordinates as v 1 x + v 2 y = 0, u 1 x + u 2 y = 0 If λ 1 > 0 and λ 2 < 0 or if λ 1 < 0 and λ 2 > 0, then we have a hyperbola The lengths and equations of the axes are given as in the case of the ellipse above If λ 1 = λ 2 0, then the equation (42) reduces to x 2 y 2 = 1 λ 1 which is a rectangular hyperbola with axes given by x = ±y In the original coordinates, this reduces to u 1 x + u 2 y = ±(v 1 x + v 2 y) 8

9 Finally if λ 1 = 0, λ 2 > 0 or if λ 1 > 0, λ 2 = 0, then the equation (42) reduces to λ 2 y 2 = 1 or λ 1 x 2 = 1 respectively In these cases, each time we get a pair of parallel lines given respectively by y = ± 1 λ2, or x = 1 λ1 Thus the equations to the lines in the original coordinates are v 1 x + v 2 y = ± 1 λ2, or u 1 x + u 2 y = ± 1 λ1 Thus, once again, by computing the eigenvalues and eigenvectors of the matrix A, we can explicitly describe the solution set S of (41) In this case also we can summarize the results obtained above in the following figure hyperbola λ 2 pair of l lines hyperbola recthyp ellipse circle ellipse pair of l lines λ 1 hyperbola S = λ 1 0, λ 2 0 recthyp Figure 2 hyperbola 9

10 5 The general equation of the second degree: straight lines Let us now consider the general equation of the second degree in two variables given by ax 2 + 2hxy + by 2 + 2gx + 2fy + c = 0 (51) We will try to completely describe the solution set S of this equation Certain computations will repeatedly occur and so it will be useful for us to do them once and for all Set x = X + α and y = Y + β Then (51) becomes ax 2 + 2hXY + by 2 +2(aα + hβ + g)x + 2(hα + bβ + f)y +(aα 2 + 2hαβ + bβ 2 + 2gα + 2fβ + c) = 0 (52) Notice that the constant term in the last line on the left-hand side of the above equation can also be rewritten as We will also set α(aα + hβ + g) + β(hα + bβ + f) + (gα + fβ + c) (53) = a h g h b f g f c We will denote by A the symmetric matrix associated to the quadratic terms, ie [ ] a h A = h b We will use the notations of the previous sections In particular λ 1 and λ 2 will stand for its two eigenvalues An associated pair of normalized eigenvectors will be denoted respectively by (u 1, u 2 ) and (v 1, v 2 ) as before Notice that ab h 2 = det(a) = λ 1 λ 2 Theorem 51 The general equation of the second degree in two variables given by (51) defines a pair of intersecting lines if, and only if, ab h 2 < 0 and = 0 Proof: Let us assume that the equation (51) represents a pair of intersecting lines and that the point of intersection is (α, β) Then, if we set x = X + 10

11 α, y = Y +β, the equation represents a pair of lines intersecting at the origin in the XY -plane Since (α, β) obviously satisfies the equation, we have aα 2 + 2hαβ + bβ 2 + 2gα + 2fβ + c = 0 (54) Further since the equation in the XY variables must be homogeneous, we must have aα + hβ + g = 0, (55) hα + bβ + f = 0, and (52) reduces to ax 2 + 2hXY + by 2 = 0 Since this represents a pair of lines intersecting at the origin, we deduce that ab h 2 < 0 This then implies that the equations (55) admit a unique solution (α, β) which is the point of intersection of these lines in the xyplane Now it follows from (54), the formulation of the left-hand side of this equation given in (53) and the system of equations in (55) that we also have gα + fβ + c = 0 (56) Thus (α, β, 1) is a non-trivial solution to the system of equations aα + hβ + g = 0, hα + bβ + f = 0, gα + fβ + c = 0 (57) This is possible only if = 0 Conversely, if ab h 2 < 0 and = 0, then choose (α, β) as the unique solution of (55) Then, since = 0, it automatically follows that (56) is also satisfied Then, taking into accont (53), we get that (52) reduces to ax 2 + 2hXY + by 2 = 0 under the transformation x = X + α, y = Y + β Now we know that this represents a pair of straight lines intersecting at the origin in the XY -plane and hence (51) represents a pair of straight lines intersecting at (α, β) This completes the proof 11

12 Remark 51 The unique solution of (55) gives the point of intersection of the two lines From the discussion in Section 2, we can explicitly write down the equations of the two lines They are given by (u 1 (x α) + u 2 (y β)) = λ 2 λ 1 (v 1 (x α) + v 2 (y β)), (u 1 (x α) + u 2 (y β)) = λ 2 λ 1 (v 1 (x α) + v 2 (y β)) Remark 52 If, in addition, we also have that a + b = 0, then the lines will be perpendicular to each other Example 51 Consider the equation x 2 y 2 + x 3y 2 = 0 Then ab h 2 = 1 < 0 Further = = 0 Thus this equation represents a pair of intersecting straight lines The system (55) reduces to α = 0, β 3 2 = 0 Thus the point of intersection is ( 1, 3 ) Now 2 2 [ ] 1 0 A = 0 1 Thus its eigenvalues and eigenvectors are λ 1 = 1, with (u 1, u 2 ) = (1, 0) and λ 2 = 1 with (v 1, v 2 ) = (0, 1) Thus the pair of lines are (x ) = (y ) (x ) = (y ), 12

13 or, equivalently, x y 1 = 0 and x + y + 2 = 0 Notice that since a + b = 0, we also have that these lines are perpendicular to each other Theorem 52 With the notations established above, assume that = 0 and ab h 2 > 0 Then the solution set S of the equation (51) consists of a single point Proof: If ab h 2 > 0, then the system (55) has a unique solution (α, β) Since = 0, this implies that (56) is also satisfied Then the transformation x = X + α, y = Y + β yields ax 2 + 2hXY + by 2 = 0 and we have seen that in this case the solution set consists of only the origin in the XY -plane Thus S = {(α, β)} Let us now turn to the case where ab h 2 = 0 and = 0 In this case the matrix A has λ 1 = 0 and λ 2 0 (If λ 2 = 0 as well, then we saw that a = b = h = 0 and the second degree terms disappear altogether) Let us define x = u 1 X + v 1 Y, y = u 2 X + v 2 Y, or, equivalently, X = u 1 x + u 2 y, Y = v 1 x + v 2 y Then, as in Section 2, the quadratic terms reduce to λ 2 Y 2 The equation (51) now becomes where Now, consider the determinant λ 2 Y 2 + 2GX + 2F Y + c = 0 G = gu 1 + fu 2, and F = gv 1 + fv 2 a h g h b f g f c Developing this by the third row (or third column) and taking into account the fact that ab h 2 = 0, we immediately observe that the determinant 13

14 is independent of the value of c Thus if = 0, it follows that the vectors (a, h, g) and (h, b, f) must be linearly dependent Since (u 1, u 2 ) is an eigenvector corresponding to λ 1 = 0, we have au 1 + hu 2 = 0 and hu 1 + bu 2 = 0 This then implies that G = gu 1 + fu 2 = 0 Thus the equation (51) now reduces to λ 2 Y 2 + 2F Y + c = 0 Completing the square, we get λ 2 ( Y + F λ 2 ) 2 + c F 2 λ 2 = 0 or, equivalently ( Y + F λ 2 ) 2 = 1 λ 2 2 (F 2 λ 2 c) This leads us to the following conclusions If F 2 > λ 2 c, then the solution set S of the equation (51) consists of two parallel lines They are given by v 1 x + v 2 y + gv 1 + fv 2 = ± 1 (gv1 + fv 2 ) λ 2 λ 2 2 λ 2 c If F 2 = λ 2 c, then S consists of a single line given by If F 2 < λ 2 c, then S = Example 52 Consider the equation v 1 x + v 2 y = gv 1 + fv 2 λ 2 2x 2 + 8xy + 8y 2 + 2gx + 2fy + c = 0 Then = 2 4 g 4 8 f g f c 14

15 Then vanishes if, and only if, f = 2g Further, [ ] 2 4 A = 4 8 Its characteristic equation is λ 2 10λ = 0 The eigenvalues are, therefore, λ 1 = 0 and λ 2 = 10 Correspondiong to λ 1 = 0, we get 2x + 4y = 0, and so every eigenvector is a multiple of (2, 1) Thus we can take ( 2 (u 1, u 2 ) = 5, 1 ) 5 Corresponding to λ 2 = 10, we get 2x + 4y = 10x, and so every eigenvector is a constant multiple of (1, 2) so that we can take ( ) 1 2 (v 1, v 2 ) = 5, 5 Thus g = gu 1 + fu 2 = g(u 1 + 2u 2 ) = 0 Similarly, Then F = gv 1 + fv 2 = g(v 1 + 2v 2 ) = 5gv 1 F 2 λ 2 c = 25g 2 v c = 5g 2 10c (i) Let us take g = 2, c = 1 Then F 2 > λ 2 c In this case, the equation is 2x 2 + 8xy + 8y 2 + 4x + 8y + 1 = 0 Dividing throughout by 2, we get (x + 2y + 1) 2 = 1 2 ie x + 2y + 1 = ± 1 2 It can be easily seen that this expression agrees with the abstract equations of the lines given earlier (ii) Let us take g = 2, c = 2 Then F 2 = λ 2 c In this case, the equation is which can be rewritten as 2x 2 + 8xy + 8y 2 + 4x + 8y + 2 = 0 (x + 2y + 1) 2 = 0 15

16 Thus we get a single line x + 2y + 1 = 0 which can be seen to agree with the abstract expression given earlier (iii) Let us take g = 1, c = 1 Then F 2 < λ 2 c In this case, the equation is This can be rewritten as 2x 2 + 8xy + 8y 2 + 2x + 4y + 1 = 0 (x + 2y) 2 + (x + 2y + 1) 2 = 0 It is easy to see that this equation has no solution Thus, all three cases can occur We now summarize the results of this section as follows If = 0 and if ab h 2 < 0, then the equation (51) represents a pair of intersecting straight lines If = 0 and if ab h 2 > 0, then the solution set of the equation (51) consists of a single point If = 0 and if ab h 2 = 0, then the solution set is one of the following: a pair of parallel lines; a single line; the empty set Starting from the coefficients of the equation and the eigenvalues and eigenvectors of the matrix A associated to the quadratic terms, we can determine which of these cases occurs 6 The general equation of the second degree: conics We will consider the equation (51) when 0 Case 1 0 and ab h 2 < 0 Since ab h 2 < 0, there exists a unique solution (α, β) to the system (55) 16

17 However, since 0, we also have that gα + fβ + c 0 Setting x = X + α, y = Y + β, the equation now reads as ax 2 + 2hXY + by 2 + (gα + fβ + c) = 0 Let us set C = gα + fβ + c Then as in Section 3, we can easily see that the solution set is a hyperbola with centre at (α, β) with the lengths of the semi-axes being C and C λ 1 The equations to the axes are given by λ 2 v 1 (x α) + v 2 (y β) = 0, u 1 (x α) + u 2 (y β) = 0, respectively If, in addition, a + b = 0, this will be a rectangular hyperbola Case 2 0 and ab h 2 > 0 Once again, there exists a unique solution (α, β) to the system (55) and setting x = X + α, y = Y + β, the equation reduces to ax 2 + 2hXY + by 2 + C = 0 where C = gα + fβ + c 0 The solution set is either an ellipse or is empty, depending on the signs of C, λ 1 and λ 2 The centre, lengths of the semi-axes and the equations of the axes are exactly as in the previous case Example 61 Consider the equation In this case 5x 2 6xy + 5y x 26y + 29 = 0 = Further ab h 2 > 0 The system (55) reads as 5α 3β + 11 = 0 3α + 5β 13 =

18 which admits the unique solution (α, β) = ( 1, 2) Further gα + fβ + c = 8 Thus, setting x = X + α, y = Y + β, we get 5X 2 6XY + 5Y 2 = 8 The eigenvalues and eigenvectors of the matrix [ 5 ] were computed in Example 21 They are λ 1 = 2, (u 1, u 2 ) = ( ) 1 1 2, 2 and λ 2 = 8, (v 1, v 2 ) = ( 1 2, 1 ) 2 Then setting the equation reduces to or, equivalently, X = u 1 X + u 2 Y Y = v 1 X + v 2 Y, 2X 2 + 8Y 2 = 8 X Y 2 = 1 1 Thus the solution set is an ellipse with centre at ( 1, 2) with the lengths of the semi-major axis being 2 and that of the semi-minor axis being 1 The equation of the major axis is Y = 0 which becomes X = Y, or, x + 1 = y 2, ie x y + 3 = 0 and that of the minor axis is X = 0, which becomes X = Y, or, x + 1 = 2 y ie x + y 1 = 0 18

19 If we change the value of the constant c from 29 to 39, then gα+fβ+c = 2 and in this case the equation reduces to 2X 2 + 8Y 2 = 2 and clearly the solution set is empty Case 3 0 and ab h 2 = 0 Since the matrix A is singular, one of its eigenvalues will be zero Without loss of generality, let λ 1 = 0 and let λ 2 0 Let us make the usual change of coordinates using the normalized eignevectors of A: The equation then transforms to x = u 1 x + v 1 y y = u 2 x + v 2 y λ 2 y 2 + 2Gx + 2F y + c = 0 where F = gu 1 + fu 2 and F = gv 1 + fv 2 Since we already have au 1 + hu 2 = 0, hu 1 + bu 2 = 0 if we also have G = 0, then it will imply that = 0, which is not the case Thus G 0 We complete squares and rewrite the equation as follows: ( λ 2 y + F ) 2 = 2Gx c + F 2 λ 2 λ 2 Now set to get which is a parabola X = x + c 2G F 2 2Gλ 2, Y = y + F λ 2 Y 2 = 2G λ 2 X We summarize the results of this section as follows 19

20 If 0 and if ab h 2 < 0, the equation (51) represents a hyperbola If, in addition, a + b = 0, it represents a rectangular hyperbola If 0 and if ab h 2 > 0, then the equation (51) either represents an ellipse or the solution set is empty If a = b and h = 0, then the equation represents a circle provided the solution set is non-empty If 0 and if ab h 2 parabola = 0, then the equation (51) represents a 20

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

Exercise 1 1. Let A be an n n orthogonal matrix. Then prove that (a) the rows of A form an orthonormal basis of R n. (b) the columns of A form an orthonormal basis of R n. (c) for any two vectors x,y R

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry

1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each)

Math 33 AH : Solution to the Final Exam Honors Linear Algebra and Applications 1. True/False: Circle the correct answer. No justifications are needed in this exercise. (1 point each) (1) If A is an invertible

Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions

Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential

Similarity and Diagonalization. Similar Matrices

MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

Orthogonal Diagonalization of Symmetric Matrices

MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

by the matrix A results in a vector which is a reflection of the given

Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

1 Eigenvalues and Eigenvectors

Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Chapter 1 Vocabulary identity - A statement that equates two equivalent expressions. verbal model- A word equation that represents a real-life problem. algebraic expression - An expression with variables.

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal

The Characteristic Polynomial

Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem

MATHEMATICS (CLASSES XI XII)

MATHEMATICS (CLASSES XI XII) General Guidelines (i) All concepts/identities must be illustrated by situational examples. (ii) The language of word problems must be clear, simple and unambiguous. (iii)

Chapter 6. Orthogonality

6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

Algebra II: Strand 7. Conic Sections; Topic 1. Intersection of a Plane and a Cone; Task 7.1.2

1 TASK 7.1.2: THE CONE AND THE INTERSECTING PLANE Solutions 1. What is the equation of a cone in the 3-dimensional coordinate system? x 2 + y 2 = z 2 2. Describe the different ways that a plane could intersect

[1] Diagonal factorization

8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:

Identifying second degree equations

Chapter 7 Identifing second degree equations 7.1 The eigenvalue method In this section we appl eigenvalue methods to determine the geometrical nature of the second degree equation a 2 + 2h + b 2 + 2g +

Examination paper for TMA4115 Matematikk 3

Department of Mathematical Sciences Examination paper for TMA45 Matematikk 3 Academic contact during examination: Antoine Julien a, Alexander Schmeding b, Gereon Quick c Phone: a 73 59 77 82, b 40 53 99

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

Lecture 14: Section 3.3

Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in

Linear Algebra Review. Vectors

Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length

1 2 3 1 1 2 x = + x 2 + x 4 1 0 1

(d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which

13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms

Introduction to Matrix Algebra

Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

ISOMETRIES OF R n KEITH CONRAD

ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x

1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

ELEMENTS OF VECTOR ALGEBRA

ELEMENTS OF VECTOR ALGEBRA A.1. VECTORS AND SCALAR QUANTITIES We have now proposed sets of basic dimensions and secondary dimensions to describe certain aspects of nature, but more than just dimensions

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number

University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0

WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE?

WHICH LINEAR-FRACTIONAL TRANSFORMATIONS INDUCE ROTATIONS OF THE SPHERE? JOEL H. SHAPIRO Abstract. These notes supplement the discussion of linear fractional mappings presented in a beginning graduate course

x 2 + y 2 = 1 y 1 = x 2 + 2x y = x 2 + 2x + 1

Implicit Functions Defining Implicit Functions Up until now in this course, we have only talked about functions, which assign to every real number x in their domain exactly one real number f(x). The graphs

The Matrix Elements of a 3 3 Orthogonal Matrix Revisited

Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation

Recall that two vectors in are perpendicular or orthogonal provided that their dot

Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal

Vector Spaces II: Finite Dimensional Linear Algebra 1

John Nachbar September 2, 2014 Vector Spaces II: Finite Dimensional Linear Algebra 1 1 Definitions and Basic Theorems. For basic properties and notation for R N, see the notes Vector Spaces I. Definition

Lecture 2. Observables

Lecture 2 Observables 13 14 LECTURE 2. OBSERVABLES 2.1 Observing observables We have seen at the end of the previous lecture that each dynamical variable is associated to a linear operator Ô, and its expectation

Solutions to Math 51 First Exam January 29, 2015

Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not

DETERMINANTS. b 2. x 2

DETERMINANTS 1 Systems of two equations in two unknowns A system of two equations in two unknowns has the form a 11 x 1 + a 12 x 2 = b 1 a 21 x 1 + a 22 x 2 = b 2 This can be written more concisely in

Notes on Determinant

ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

Multiplicity. Chapter 6

Chapter 6 Multiplicity The fundamental theorem of algebra says that any polynomial of degree n 0 has exactly n roots in the complex numbers if we count with multiplicity. The zeros of a polynomial are

Determine whether the following lines intersect, are parallel, or skew. L 1 : x = 6t y = 1 + 9t z = 3t. x = 1 + 2s y = 4 3s z = s

Homework Solutions 5/20 10.5.17 Determine whether the following lines intersect, are parallel, or skew. L 1 : L 2 : x = 6t y = 1 + 9t z = 3t x = 1 + 2s y = 4 3s z = s A vector parallel to L 1 is 6, 9,

Vector and Matrix Norms

Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

Inner product. Definition of inner product

Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product

(January 14, 2009) End k (V ) End k (V/W )

(January 14, 29) [16.1] Let p be the smallest prime dividing the order of a finite group G. Show that a subgroup H of G of index p is necessarily normal. Let G act on cosets gh of H by left multiplication.

Solving Systems of Linear Equations

LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL

SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics

NOTES on LINEAR ALGEBRA 1

School of Economics, Management and Statistics University of Bologna Academic Year 205/6 NOTES on LINEAR ALGEBRA for the students of Stats and Maths This is a modified version of the notes by Prof Laura

JUST THE MATHS UNIT NUMBER 8.5. VECTORS 5 (Vector equations of straight lines) A.J.Hobson

JUST THE MATHS UNIT NUMBER 8.5 VECTORS 5 (Vector equations of straight lines) by A.J.Hobson 8.5.1 Introduction 8.5. The straight line passing through a given point and parallel to a given vector 8.5.3

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they

Roots and Coefficients of a Quadratic Equation Summary

Roots and Coefficients of a Quadratic Equation Summary For a quadratic equation with roots α and β: Sum of roots = α + β = and Product of roots = αβ = Symmetrical functions of α and β include: x = and

5.3 The Cross Product in R 3

53 The Cross Product in R 3 Definition 531 Let u = [u 1, u 2, u 3 ] and v = [v 1, v 2, v 3 ] Then the vector given by [u 2 v 3 u 3 v 2, u 3 v 1 u 1 v 3, u 1 v 2 u 2 v 1 ] is called the cross product (or

Presentation 3: Eigenvalues and Eigenvectors of a Matrix

Colleen Kirksey, Beth Van Schoyck, Dennis Bowers MATH 280: Problem Solving November 18, 2011 Presentation 3: Eigenvalues and Eigenvectors of a Matrix Order of Presentation: 1. Definitions of Eigenvalues

11.1. Objectives. Component Form of a Vector. Component Form of a Vector. Component Form of a Vector. Vectors and the Geometry of Space

11 Vectors and the Geometry of Space 11.1 Vectors in the Plane Copyright Cengage Learning. All rights reserved. Copyright Cengage Learning. All rights reserved. 2 Objectives! Write the component form of

Systems of Linear Equations

Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

Inner Product Spaces

Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

Linear Algebra: Vectors

A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector

Matrix Norms. Tom Lyche. September 28, Centre of Mathematics for Applications, Department of Informatics, University of Oslo

Matrix Norms Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 28, 2009 Matrix Norms We consider matrix norms on (C m,n, C). All results holds for

THREE DIMENSIONAL GEOMETRY

Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,

4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows

Content. Chapter 4 Functions 61 4.1 Basic concepts on real functions 62. Credits 11

Content Credits 11 Chapter 1 Arithmetic Refresher 13 1.1 Algebra 14 Real Numbers 14 Real Polynomials 19 1.2 Equations in one variable 21 Linear Equations 21 Quadratic Equations 22 1.3 Exercises 28 Chapter

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?

Solutions to Linear Algebra Practice Problems

Solutions to Linear Algebra Practice Problems. Find all solutions to the following systems of linear equations. (a) x x + x 5 x x x + x + x 5 (b) x + x + x x + x + x x + x + 8x Answer: (a) We create the

Matrix Representations of Linear Transformations and Changes of Coordinates

Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under

APPLICATIONS. are symmetric, but. are not.

CHAPTER III APPLICATIONS Real Symmetric Matrices The most common matrices we meet in applications are symmetric, that is, they are square matrices which are equal to their transposes In symbols, A t =

160 CHAPTER 4. VECTOR SPACES

160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results

The basic unit in matrix algebra is a matrix, generally expressed as: a 11 a 12. a 13 A = a 21 a 22 a 23

(copyright by Scott M Lynch, February 2003) Brief Matrix Algebra Review (Soc 504) Matrix algebra is a form of mathematics that allows compact notation for, and mathematical manipulation of, high-dimensional

Lecture L3 - Vectors, Matrices and Coordinate Transformations

S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between

October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix

Linear Algebra & Properties of the Covariance Matrix October 3rd, 2012 Estimation of r and C Let rn 1, rn, t..., rn T be the historical return rates on the n th asset. rn 1 rṇ 2 r n =. r T n n = 1, 2,...,

The Phase Plane. Phase portraits; type and stability classifications of equilibrium solutions of systems of differential equations

The Phase Plane Phase portraits; type and stability classifications of equilibrium solutions of systems of differential equations Phase Portraits of Linear Systems Consider a systems of linear differential

Math Review Large Print (18 point) Edition Chapter 2: Algebra

GRADUATE RECORD EXAMINATIONS Math Review Large Print (18 point) Edition Chapter : Algebra Copyright 010 by Educational Testing Service. All rights reserved. ETS, the ETS logo, GRADUATE RECORD EXAMINATIONS,

PYTHAGOREAN TRIPLES KEITH CONRAD

PYTHAGOREAN TRIPLES KEITH CONRAD 1. Introduction A Pythagorean triple is a triple of positive integers (a, b, c) where a + b = c. Examples include (3, 4, 5), (5, 1, 13), and (8, 15, 17). Below is an ancient

Eigenvalues and Eigenvectors

Chapter 6 Eigenvalues and Eigenvectors 6. Introduction to Eigenvalues Linear equations Ax D b come from steady state problems. Eigenvalues have their greatest importance in dynamic problems. The solution

C 1 x(t) = e ta C = e C n. 2! A2 + t3

Matrix Exponential Fundamental Matrix Solution Objective: Solve dt A x with an n n constant coefficient matrix A x (t) Here the unknown is the vector function x(t) x n (t) General Solution Formula in Matrix

Lecture 2: Homogeneous Coordinates, Lines and Conics

Lecture 2: Homogeneous Coordinates, Lines and Conics 1 Homogeneous Coordinates In Lecture 1 we derived the camera equations λx = P X, (1) where x = (x 1, x 2, 1), X = (X 1, X 2, X 3, 1) and P is a 3 4

The Inverse of a Matrix

The Inverse of a Matrix 7.4 Introduction In number arithmetic every number a ( 0) has a reciprocal b written as a or such that a ba = ab =. Some, but not all, square matrices have inverses. If a square

Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday.

Math 312, Fall 2012 Jerry L. Kazdan Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday. In addition to the problems below, you should also know how to solve

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone

Math Placement Test Study Guide. 2. The test consists entirely of multiple choice questions, each with five choices.

Math Placement Test Study Guide General Characteristics of the Test 1. All items are to be completed by all students. The items are roughly ordered from elementary to advanced. The expectation is that

4. MATRICES Matrices

4. MATRICES 170 4. Matrices 4.1. Definitions. Definition 4.1.1. A matrix is a rectangular array of numbers. A matrix with m rows and n columns is said to have dimension m n and may be represented as follows:

Introduction to Matrix Algebra I

Appendix A Introduction to Matrix Algebra I Today we will begin the course with a discussion of matrix algebra. Why are we studying this? We will use matrix algebra to derive the linear regression model

MATH 551 - APPLIED MATRIX THEORY

MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

String Art Mathematics: An Introduction to Geometry Expressions and Math Illustrations Stephen Arnold Compass Learning Technologies

String Art Mathematics: An Introduction to Geometry Expressions and Math Illustrations Stephen Arnold Compass Learning Technologies Introduction How do you create string art on a computer? In this lesson

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression

The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression The SVD is the most generally applicable of the orthogonal-diagonal-orthogonal type matrix decompositions Every

1 Introduction to Matrices

1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

Math 497C Sep 9, Curves and Surfaces Fall 2004, PSU

Math 497C Sep 9, 2004 1 Curves and Surfaces Fall 2004, PSU Lecture Notes 2 15 sometries of the Euclidean Space Let M 1 and M 2 be a pair of metric space and d 1 and d 2 be their respective metrics We say

4.6 Null Space, Column Space, Row Space

NULL SPACE, COLUMN SPACE, ROW SPACE Null Space, Column Space, Row Space In applications of linear algebra, subspaces of R n typically arise in one of two situations: ) as the set of solutions of a linear

Eigenvalues and eigenvectors of a matrix

Eigenvalues and eigenvectors of a matrix Definition: If A is an n n matrix and there exists a real number λ and a non-zero column vector V such that AV = λv then λ is called an eigenvalue of A and V is

EL-9650/9600c/9450/9400 Handbook Vol. 1

Graphing Calculator EL-9650/9600c/9450/9400 Handbook Vol. Algebra EL-9650 EL-9450 Contents. Linear Equations - Slope and Intercept of Linear Equations -2 Parallel and Perpendicular Lines 2. Quadratic Equations