The Matrix Elements of a 3 3 Orthogonal Matrix Revisited

Save this PDF as:

Size: px
Start display at page:

Transcription

1 Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation of the explicit form for most general 3 3 orthogonal matrix. A proper rotation matrix with determinant 1, denoted by R(ˆn, θ), represents a counterclockwise rotation by an angle θ about a fixed axis ˆn. An improper rotation with determinant 1, denoted by R(ˆn, θ) represents a reflection through a plane that passes through the origin and is perpendicular to a unit vector ˆn (called the normal to the reflection plane) together with a counterclockwise rotation by θ about ˆn. For example, the matrix representation of the counterclockwise rotation by an angle θ about a fixed z-axis is given by [cf. eq. (7.18) on p. 129 of Boas]: cosθ sin θ 0 R(k, θ) sin θ cosθ 0. (1) Similarly, the matrix representation of a reflection through the x y plane together with a counterclockwise rotation by an angle θ about a fixed z-axis (which points along the normal to the reflection plane) is given by [cf. eq. (7.19) on p. 129 of Boas]: cosθ sin θ 0 R(k, θ) sin θ cosθ 0. (2) In these notes, I shall derive the general expressions for the matrix elements of R(ˆn, θ) and R(ˆn, θ). These expressions have already been obtained in the class handout cited above. In contrast, I will provide here a much simpler derivation of the explicit forms for R(ˆn, θ) and R(ˆn, θ), based on the techniques of tensor algebra. 2. A derivation of the Rodriguez formula The matrix elements of R(ˆn, θ) will be denoted by R ij. Since R(ˆn, θ) describes a rotation by an angle θ about an axis ˆn, the formula for R ij that we seek will depend on θ and on the coordinates of ˆn = (n 1, n 2, n 3 ) with respect to a fixed Cartesian coordinate system. Note that since ˆn is a unit vector, it follows that: n n n 2 3 = 1. (3) Using the techniques of tensor algebra, we can derive the formula for R ij in the following way. We can regard R ij as the components of a second-rank tensor (see 1

2 Appendix A). Likewise, the n i are components of a vector (equivalently, a first-rank tensor). Two other important quantities for the analysis are the invariant tensors δ ij (the Kronecker delta) and ǫ ijk (the Levi-Civita tensor). If we invoke the covariance of Cartesian tensor equations, then one must be able to express R ij in terms of a second-rank tensor composed of n i, δ ij and ǫ ijk, as there are no other tensors in the problem that could provide a source of indices. Thus, the form of the formula for R ij must be: R ij = aδ ij + bn i n j + cǫ ijk n k, (4) where there is an implicit sum over the index k in the third term of eq. (4). 1 The numbers a, b and c are real scalar quantities. As such, a, b and c are functions of θ, since the rotation angle is the only relevant scalar quantity in this problem. 2 If we also allow for transformations between right-handed and left-handed orthonormal coordinate systems, then R ij and δ ij are true second-rank tensors and ǫ ijk is a third-rank pseudotensor. Thus, to ensure that eq. (4) is covariant with respect to transformations between two bases that are related by either a proper or an improper rotation, we conclude that a and b are true scalars, and the product cˆn is a pseudovector. 3 We now propose to deduce conditions that are satisfied by a, b and c. The first condition is obtained by noting that R(ˆn, θ)ˆn = ˆn. (5) This is clearly true, since R(ˆn, θ), when acting on a vector, rotates the vector around the axis ˆn, whereas any vector parallel to the axis of rotation is invariant under the action of R(ˆn, θ). In terms of components R ij n j = n i. (6) To determine the consequence of this equation, we insert eq. (4) into eq. (6) and make use of eq. (3). Noting that 4 δ ij n j = n i, n j n j = 1 ǫ ijk n j n k = 0, (7) 1 We follow the Einstein summation convention in these notes. That is, there is an implicit sum over any pair of repeated indices in the present and all subsequent formulae. 2 One can also construct a scalar by taking the dot product of ˆn ˆn, but this quantity is equal to 1 [cf. eq. (3)], since ˆn is a unit vector. 3 Under inversion of the coordinate system, θ θ and ˆn ˆn. However, since 0 θ π (by convention), we must then use eq. (12) to flip the signs of both θ and ˆn to represent the rotation R(ˆn, θ) in the new coordinate system. Hence, the signs of θ and ˆn effectively do not change under the inversion of the coordinate system. That is, θ is a true scalar and ˆn is a pseudovector, in which case c is also a true scalar. In a different convention where π θ π (which we do not adopt in these notes), θ is a pseudoscalar and ˆn is a true vector, in which case c is also a pseudoscalar. Independent of these conventions, the product cˆn is a pseudovector as asserted in the text above. 4 In the third equation of eq. (7), there is an implicit sum over j and k. Since ǫ ijk = ǫ jik, when the sum ǫ ijk n j n k is carried out, we find that for every positive term, there is an identical negative term to cancel it. The total sum is therefore equal to zero. This is an example of a very general rule. Namely, one always finds that the product of two tensor quantities, one symmetric under the interchange of a pair of summed indices and one antisymmetric under the interchange of a pair of summed indices, is equal to zero when summed over the two indices. In the present case, n j n k is symmetric under the interchange of j and k, whereas ǫ ijk is antisymmetric under the interchange of j and k. Hence their product, summed over j and k, is equal to zero. 2

3 it follows immediately that n i (a + b) = n i. Hence, a + b = 1. (8) Since the formula for R ij given by eq. (4) must be completely general, it must hold for any special case. In particular, consider the case where ˆn = k. In this case, eqs. (1) and (4) yields: R(k, θ) 11 = cosθ = a, R(k, θ) 12 = sin θ = c ǫ 123 n 3 = c. (9) Using eqs. (8) and (9) we conclude that, a = cos θ, b = 1 cosθ, c = sin θ. (10) Inserting these results into eq. (4) yields the Rodriguez formula: R ij (ˆn, θ) = cosθ δ ij + (1 cosθ)n i n j sin θ ǫ ijk n k (11) We can write R(ˆn, θ) explicitly in 3 3 matrix form, although eq. (11) is more compact and convenient. Indeed, one can check that eq. (11) is equivalent to eq. (17) of the previous class handout, Three-Dimensional Proper and Improper Rotation Matrices. The general rotation matrix R(ˆn, θ) given in eq. (11) satisfies the following two relations: [R(ˆn, θ)] 1 = R(ˆn, θ) = R( ˆn, θ), (12) R(ˆn, 2π θ) = R( ˆn, θ), (13) which implies that any three-dimensional rotation can be described by a counterclockwise rotation by θ about an arbitrary axis ˆn, where 0 θ π. In this convention, the overall sign of ˆn is meaningful for 0 < θ < π. In the case of θ = π, we have R ij (ˆn, π) = R ij ( ˆn, π) = 2n i n j δ ij, (14) which means that R(ˆn, π) and R( ˆn, π) represent the same rotation. Finally, if θ = 0 then R ij (ˆn, 0) = δ ij is the identity operator, which is independent of the direction of ˆn. 3. The matrix elements of an improper rotation matrix An improper rotation matrix is an orthogonal matrix, R, such that det R = 1. The most general three-dimensional improper rotation is of the form: R(ˆn, θ) R(ˆn, θ)r(ˆn) = R(ˆn)R(ˆn, θ), (15) where R(ˆn) is called a reflection matrix. In particular, R(ˆn) = R( ˆn) represents a mirror reflection through a plane passing through the origin that is normal to the unit 3

4 vector ˆn. 5 Note that the improper rotation defined in eq. (15) does not depend on the order in which the proper rotation and reflection are applied. It follows from eq. (15) that R(ˆn) = R( ˆn) = R(ˆn, 0), (16) after using R(ˆn, 0) = I. Note that all reflection matrices are orthogonal matrices with detr(ˆn) = 1, with the property that: [R(ˆn)] 2 = I. The matrix elements of R(ˆn, θ) will be denoted by R ij. The derivation of an explicit form for R(ˆn, θ) follows closely the derivation of R(ˆn, θ) given in Section 1. In particular, we can also express R ij in terms of a second-rank tensor composed of n i, δ ij and ǫ ijk, since there are no other tensors in the problem that could provide a source of indices. Thus, the form of the formula for R ij must be: R ij = aδ ij + bn i n j + cǫ ijk n k, (17) where the coefficients of a, b and c need not be the same as those that appear in eq. (4). In this case, a, b and c can be determined as follows. The first condition is obtained by noting that R(ˆn, θ)ˆn = ˆn. This equation is true since any vector that is parallel to ˆn (which points along the normal to the reflection plane) is inverted (whereas the associated rotation about ˆn has no further effect). In terms of components R ij n j = n i. (18) To determine the consequence of this equation, we insert eq. (17) into eq. (18) and make use of eq. (3). using eq. (7), it follows immediately that n i (a+b) = n i. Hence, a + b = 1. (19) Since the formula for R ij given by eq. (17) must be completely general, it must hold for any special case. In particular, consider the case where ˆn = k. In this case, eqs. (2) and (17) yields: R(k, θ) 11 = cosθ = a, R(k, θ) 12 = sin θ = c ǫ 123 n 3 = c. (20) Using eqs. (19) and (20) we conclude that, a = cosθ, b = 1 cosθ, c = sin θ. (21) 5 Geometrically, it is clear that a mirror reflection is not sensitive to the sign of the normal to the reflection plane ˆn. 4

5 Inserting these results into eq. (17) yields the analog of the Rodriguez formula for improper rotation matrices: R ij (ˆn, θ) = cosθ δ ij (1 + cosθ)n i n j sin θ ǫ ijk n k (22) We can write R(ˆn, θ) explicitly in 3 3 matrix form, although eq. (22) is more compact and convenient. Indeed, one can check that eq. (22) is equivalent to eq. (47) of the previous class handout, Three-Dimensional Proper and Improper Rotation Matrices. The general improper rotation matrix R(ˆn, θ) given in eq. (22) satisfies the following two relations: [R(ˆn, θ)] 1 = R(ˆn, θ) = R( ˆn, θ), (23) R(ˆn, 2π θ) = R( ˆn, θ), (24) which implies that any three-dimensional improper rotation can be described by a reflection through a plane that passes through the origin and is perpendicular to ˆn, together with a counterclockwise rotation by θ about ˆn, where 0 θ π. In this convention, the overall sign of ˆn is meaningful for 0 < θ < π. In the case of θ = 0, the improper rotation matrix R(ˆn, 0) = R( ˆn, 0) = δ ij 2n i n j, corresponds to the reflection matrix R(ˆn) defined in eq. (16). Indeed, R(ˆn, 0) and R( ˆn, 0) represent the same reflection. Finally, if θ = π then R ij (ˆn, π) = δ ij is the inversion operator, which is independent of the direction of ˆn. 4. Determining the properties of a general 3 3 orthogonal matrix The results obtained in eqs. (11) and (22) can be expressed as a single equation. Consider a general 3 3 orthogonal matrix R, corresponding to either a proper or improper rotation. Then its matrix elements are given by: where R ij (ˆn, θ) = cosθ δ ij + (ε cosθ)n i n j sin θ ǫ ijk n k, (25) ε detr(ˆn, θ). (26) That is, ε = 1 for a proper rotation and ε = 1 for an improper rotation. Using eq. (25), one can derive expressions for the unit vector ˆn and the angle θ in terms of the matrix elements of R. With some tensor algebra manipulations involving the Levi-Civita tensor, we can quickly obtain the desired results. First, we compute the trace of R(ˆn, θ). In particular, using eq. (25) it follows that: 6 Tr R(ˆn, θ) R ii = ε + 2 cosθ. (27) 6 Since we are employing the Einstein summation convention, the quantities R ii Tr R and δ ii TrI each involve an implicit sum over i, and thus define the trace of R and I, respectively. 5

6 In deriving this result, we used the fact that δ ii = Tr I = 3 (since the indices run over i = 1, 2, 3 in three-dimensional space) and ǫ iik = 0 (the latter is a consequence of the fact that the Levi-Civita tensor is totally antisymmetric under the interchange of any two indices). By convention, 0 θ π, which implies that sin θ 0. Thus, cos θ = 1 2 (TrR ε) and sin θ = (1 cos2 θ) 1/2 = 1 2 (3 ε TrR)(1 + ε TrR), (28) where cosθ is determined from eq. (27) and we have used ε 2 = 1. All that remains is to determine the unit vector ˆn. Let us multiply eq. (11) by ǫ ijm and sum over i and j. Noting that 7 it follows that ǫ ijm δ ij = ǫ ijm n i n j = 0, ǫ ijk ǫ ijm = 2δ km, (29) 2n m sin θ = R ij ǫ ijm. (30) If R is a symmetric matrix (i.e. R ij = R ji ), then R ij ǫ ijm = 0 automatically since ǫ ijk is antisymmetric under the interchange of the indices i and j. In this case sin θ = 0 and we must seek other means to determine ˆn. If sin θ 0, then one can divide both sides of eq. (30) by sin θ. Using eq. (28), we obtain: More explicitly, ˆn = n m = R ijǫ ijm 2 sin θ = R ij ǫ ijm, sin θ 0. (31) (3 ε TrR)(1 + ε TrR) ( ) 1 R 32 R 23, R 13 R 31, R 21 R 12, ε TrR 1, 3. (3 ε TrR)(1 + ε TrR) (32) In Appendix B, we verify that ˆn as given by eq. (31) is a vector of unit length [as required by eq. (3)]. The overall sign of ˆn is fixed by eq. (31) due to our convention in which sin θ 0. If we multiply eq. (30) by n m and sum over m, then sin θ = 1 2 ǫ ijmr ij n m, (33) after using n m n m = 1. This provides an additional check on the determination of the rotation angle. Alternatively, we can define a matrix S whose matrix elements are given by: S jk R jk + R kj + (ε TrR)δ jk (34) = 2(ε cosθ)n j n k = (3ε TrR)n j n k, 7 As discussed in footnote 4, the identity ǫ ijm n i n j = 0 arises because ǫ ijm is antisymmetric and n i n j is symmetric under the interchange of i and j. Summing over the repeated indices i and j then yields an equal number of positive and negative terms that cancel exactly. To derive the second identity of eq. (29), one can set j = l in eq. (38) and then sum over the repeated index l. Relabeling the indices of the resulting sum then gives the identity ǫ ijk ǫ ijm = 2δ km. 6

7 after using eq. (25) for R jk. Hence, 8 n j n k = S jk, TrR 3ε. (35) 3ε TrR To determine ˆn up to an overall sign, we simply set j = k (no sum) in eq. (35), which fixes the value of n 2 j. If sin θ 0, the overall sign of ˆn is fixed by eq. (30). As noted above, if R is a symmetric matrix (i.e. R ij = R ji ), then sin θ = 0 and ˆn cannot be determined from eq. (31). In this case, eq. (27) determines whether cosθ = +1 or cosθ = 1. If cos θ = ε, then R ij = εδ ij, in which case S = 0 and the axis ˆn is undefined. For cosθ = ε, one can determine ˆn up to an overall sign using eq. (35) as noted above. In this case, the ambiguity in the overall sign of ˆn is immaterial, in light of eq. (14). To summarize, eqs. (28), (32) and (35) provide a simple algorithm for determining the unit vector ˆn and the rotation angle θ for any proper or improper rotation matrix R(ˆn, θ) εi. Finally, one additional property of three-dimensional proper rotation matrices is especially noteworthy: R(ˆn, θ) = PR(ˆn, θ)p 1, for ˆn = P ˆn, where P is a proper rotation matrix. (36) A proof that employs the methods of tensor algebra is given in Appendix C. Appendix A: Matrix elements of matrices correspond to the components of second rank tensors In the class handout entitled, Vector coordinates, matrix elements and changes of basis, we examined how the matrix elements of linear operators change under a change of basis. Consider the matrix elements of a linear operator with respect to two different orthonormal bases, B = {ê 1, ê 2, ê 3 } and B = {ê 1, ê 2, ê 3 }. Then, using the Einstein summation convention, ê j = P ijê i, where P is an orthogonal matrix. Given any linear operator A with matrix elements a ij with respect to the basis B, the matrix elements a ij with respect to the basis B are given by a kl = (P 1 ) ki a ij P jl = P ik a ij P jl, where we have used the fact that P 1 = P T in the second step above. Finally, identifying P = R 1, where R is also an orthogonal matrix, it follows that a kl = R kir lj a ij, which we recognize as the transformation law for the components of a second rank Cartesian tensor. 8 Eq. (34) yields Tr S = 3ε TrR. One can then use eq. (35) to verify that ˆn is a unit vector. 7

8 Appendix B: Verifying that ˆn obtained from eq. (31) is a unit vector We first need some preliminary results. Using the results from the handout entitled, The Characteristic Polynomial, the characteristic equation of an arbitrary 3 3 matrix R is given by: p(λ) = [ λ 3 λ 2 Tr R + c 2 λ det R ], where [ c 2 = 1 2 (Tr R) 2 Tr(R 2 ) ]. For an orthogonal matrix, ε det R = ±1. Hence, p(λ) = [ λ 3 λ 2 Tr R λ [ (Tr R) 2 Tr(R 2 ) ] ε ]. We now employ the Cayley-Hamilton theorem, which states that a matrix satisfies its own characteristic equation, i.e. p(r) = 0. Hence, R 3 R 2 Tr R R [ (Tr R) 2 Tr(R 2 ) ] εi = 0. Multiplying the above equation by R 1, and using the fact that R 1 = R T for an orthogonal matrix, R 2 R Tr R I [ (Tr R) 2 Tr(R 2 ) ] εr T = 0. Finally, we take the trace of the above equation. Using Tr(R T ) = Tr R, we can solve for Tr(R 2 ). Using Tr I = 3, the end result is given by: Tr(R 2 ) = (Tr R) 2 2ε TrR, (37) which is satisfied by all 3 3 orthogonal matrices. We now verify that ˆn as determined from eq. (31) is a unit vector. For convenience, we repeat eq. (31) here: n m = 1 R ij ǫ ijm 2 sin θ = R ij ǫ ijm (3 ε Rii )(1 + ε R ii ), sin θ 0, where R ii Tr R. We evaluate ˆn ˆn = n m n m as follows: R ij ǫ ijm R kl ǫ klm n m n m = (3 ε R ii )(1 + ε R ii ) = R ijr kl (δ ik δ jl δ il δ jk ) (3 ε R ii )(1 + ε R ii ) = R ijr ij R ij R ji (3 ε R ii )(1 + ε R ii ), after making use of the identity given in eq. (5.8) on p. 510 of Boas, The numerator of the above expression is equal to: ǫ ijm ǫ klm = δ ik δ jl δ il δ jk. (38) R ij R ij R ij R ji = Tr(R T R) Tr(R 2 ) = Tr I Tr(R 2 ) = 3 Tr(R 2 ) = 3 (TrR) 2 + 2ε TrR = (3 ε R ii )(1 + ε R ii ), (39) after using eq. (37) for Tr(R 2 ). Hence, employing eq. (39) in the expression for n m n m above yields ˆn ˆn = n m n m = R ijr ij R ij R ji (3 ε R ii )(1 + ε R ii ) = 1, and the proof is complete. 8

9 Appendix C: Proof of the theorem given in eq. (36) In this appendix, we prove the following theorem for proper three-dimensional rotation matrices, R(ˆn, θ) = PR(ˆn, θ)p 1, for ˆn = P ˆn, where P is a proper rotation matrix. (40) To prove eq. (40), we first compute the angle of the rotation PR(ˆn, θ)p 1 using eq. (28) with ε = 1. Since Tr[PR(ˆn, θ)p 1 ] = Tr R(ˆn, θ) using the cyclicity of the trace, it follows that the angles of rotation corresponding to PR(ˆn, θ)p 1 and R(ˆn, θ) coincide and are both equal to θ. To compute the corresponding axis of rotation ˆn, we employ eq. (30), 2n m sin θ = (PR P 1 ) ij ǫ ijm, (41) where R R(ˆn, θ). Since P is a rotation matrix, we have P 1 = P T, or equivalently (P 1 ) lj = P jl. Hence, we can rewrite eq. (41) as: 2n m sin θ = P ik R kl P jlǫ ijm. (42) Multiplying both sides of eq. (42) by P mn and using the definition of the determinant of a 3 3 matrix, P ik P jl P mn ǫ ijm = (det P)ǫ kln, it then follows that: 2P mn n m sin θ = R kl ǫ kln. (43) after noting that det P = 1 (since P is a proper rotation matrix). Finally, we again use eq. (30) which yields 2n n sin θ = R kl ǫ kln. (44) Assuming that sin θ 0, we can subtract eqs. (43) and (44) and divide out by 2 sinθ. Using (P T ) nm = P mn, the end result is: Since PP T = P T P = I, we conclude that ˆn P Tˆn = 0. ˆn = P ˆn. (45) The case of sin θ = 0 must be treated separately. Using eq. (5), one can determine the axis of rotation ˆn of the rotation matrix PR(ˆn, θ)p 1 up to an overall sign. Since R(ˆn, θ)ˆn = ˆn, the following eigenvalue equation is obtained: PR(ˆn, θ)p 1 (P ˆn ) = PR(ˆn, θ)ˆn = P ˆn. (46) That is, P ˆn is an eigenvector of PR(ˆn, θ)p 1 with eigenvalue +1. It then follows that P ˆn is the normalized eigenvector of PR(ˆn, θ)p 1 up to an overall undetermined sign. For sin θ 0, the overall sign is fixed and is positive by eq. (45). If sinθ = 0, then there are two cases to consider. If θ = 0, then R(ˆn, 0) = R(ˆn, 0) = I and eq. (40) is trivially satisfied. If θ = π, then eq. (14) implies that the unit vector parallel to the rotation axis is only defined up to an overall sign. Hence, eq. (45) is valid even in the case of sin θ = 0, and the proof is complete. 9

1 Scalars, Vectors and Tensors

DEPARTMENT OF PHYSICS INDIAN INSTITUTE OF TECHNOLOGY, MADRAS PH350 Classical Physics Handout 1 8.8.2009 1 Scalars, Vectors and Tensors In physics, we are interested in obtaining laws (in the form of mathematical

Three-Dimensional Proper and Improper Rotation Matrices

Physics 116A Winter 011 Three-Dimensional Proper and Improper Rotation Matrices 1. Proper and improper rotation matrices ArealorthogonalmatrixRisamatrixwhoseelementsarerealnumbersandsatisfies R 1 = R T

Index Notation for Vector Calculus

Index Notation for Vector Calculus by Ilan Ben-Yaacov and Francesc Roig Copyright c 2006 Index notation, also commonly known as subscript notation or tensor notation, is an extremely useful tool for performing

The Characteristic Polynomial

Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem

A Primer on Index Notation

A Primer on John Crimaldi August 28, 2006 1. Index versus Index notation (a.k.a. Cartesian notation) is a powerful tool for manipulating multidimensional equations. However, there are times when the more

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

Vectors and Index Notation

Vectors and Index Notation Stephen R. Addison January 12, 2004 1 Basic Vector Review 1.1 Unit Vectors We will denote a unit vector with a superscript caret, thus â denotes a unit vector. â â = 1 If x is

Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product

Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product Geometrical definition Properties Expression in components. Definition in components Properties Geometrical expression.

Index notation in 3D. 1 Why index notation?

Index notation in 3D 1 Why index notation? Vectors are objects that have properties that are independent of the coordinate system that they are written in. Vector notation is advantageous since it is elegant

Geometry of Vectors. 1 Cartesian Coordinates. Carlo Tomasi

Geometry of Vectors Carlo Tomasi This note explores the geometric meaning of norm, inner product, orthogonality, and projection for vectors. For vectors in three-dimensional space, we also examine the

Figure 1.1 Vector A and Vector F

CHAPTER I VECTOR QUANTITIES Quantities are anything which can be measured, and stated with number. Quantities in physics are divided into two types; scalar and vector quantities. Scalar quantities have

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?

Similarity and Diagonalization. Similar Matrices

MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

Matrix Algebra LECTURE 1. Simultaneous Equations Consider a system of m linear equations in n unknowns: y 1 = a 11 x 1 + a 12 x 2 + +a 1n x n,

LECTURE 1 Matrix Algebra Simultaneous Equations Consider a system of m linear equations in n unknowns: y 1 a 11 x 1 + a 12 x 2 + +a 1n x n, (1) y 2 a 21 x 1 + a 22 x 2 + +a 2n x n, y m a m1 x 1 +a m2 x

CBE 6333, R. Levicky 1. Tensor Notation.

CBE 6333, R. Levicky 1 Tensor Notation. Engineers and scientists find it useful to have a general terminology to indicate how many directions are associated with a physical quantity such as temperature

Physics 235 Chapter 1. Chapter 1 Matrices, Vectors, and Vector Calculus

Chapter 1 Matrices, Vectors, and Vector Calculus In this chapter, we will focus on the mathematical tools required for the course. The main concepts that will be covered are: Coordinate transformations

13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation

Chapter 6. Orthogonality

6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

[1] Diagonal factorization

8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:

Summary of week 8 (Lectures 22, 23 and 24)

WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry

1 Spherical Kinematics

ME 115(a): Notes on Rotations 1 Spherical Kinematics Motions of a 3-dimensional rigid body where one point of the body remains fixed are termed spherical motions. A spherical displacement is a rigid body

Lecture Notes 1: Matrix Algebra Part B: Determinants and Inverses

University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 57 Lecture Notes 1: Matrix Algebra Part B: Determinants and Inverses Peter J. Hammond email: p.j.hammond@warwick.ac.uk Autumn 2012,

Tensors on a vector space

APPENDIX B Tensors on a vector space In this Appendix, we gather mathematical definitions and results pertaining to tensors. The purpose is mostly to introduce the modern, geometrical view on tensors,

Dot product and vector projections (Sect. 12.3) There are two main ways to introduce the dot product

Dot product and vector projections (Sect. 12.3) Two definitions for the dot product. Geometric definition of dot product. Orthogonal vectors. Dot product and orthogonal projections. Properties of the dot

A matrix over a field F is a rectangular array of elements from F. The symbol

Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted

Lecture L3 - Vectors, Matrices and Coordinate Transformations

S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between

State of Stress at Point

State of Stress at Point Einstein Notation The basic idea of Einstein notation is that a covector and a vector can form a scalar: This is typically written as an explicit sum: According to this convention,

University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal

Brief Review of Tensors

Appendix A Brief Review of Tensors A1 Introductory Remarks In the study of particle mechanics and the mechanics of solid rigid bodies vector notation provides a convenient means for describing many physical

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

Primer on Index Notation

Massachusetts Institute of Technology Department of Physics Physics 8.07 Fall 2004 Primer on Index Notation c 2003-2004 Edmund Bertschinger. All rights reserved. 1 Introduction Equations involving vector

9 Multiplication of Vectors: The Scalar or Dot Product

Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation

How is a vector rotated?

How is a vector rotated? V. Balakrishnan Department of Physics, Indian Institute of Technology, Madras 600 036 Appeared in Resonance, Vol. 4, No. 10, pp. 61-68 (1999) Introduction In an earlier series

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

1 Vectors: Geometric Approach

c F. Waleffe, 2008/09/01 Vectors These are compact lecture notes for Math 321 at UW-Madison. Read them carefully, ideally before the lecture, and complete with your own class notes and pictures. Skipping

A vector is a directed line segment used to represent a vector quantity.

Chapters and 6 Introduction to Vectors A vector quantity has direction and magnitude. There are many examples of vector quantities in the natural world, such as force, velocity, and acceleration. A vector

Inner Product Spaces

Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

UNIT 2 MATRICES - I 2.0 INTRODUCTION. Structure

UNIT 2 MATRICES - I Matrices - I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress

THREE DIMENSIONAL GEOMETRY

Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,

Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions

Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential

Chapter 5 Polar Coordinates; Vectors 5.1 Polar coordinates 1. Pole and polar axis

Chapter 5 Polar Coordinates; Vectors 5.1 Polar coordinates 1. Pole and polar axis 2. Polar coordinates A point P in a polar coordinate system is represented by an ordered pair of numbers (r, θ). If r >

13.4 THE CROSS PRODUCT

710 Chapter Thirteen A FUNDAMENTAL TOOL: VECTORS 62. Use the following steps and the results of Problems 59 60 to show (without trigonometry) that the geometric and algebraic definitions of the dot product

CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with

3 Orthogonal Vectors and Matrices

3 Orthogonal Vectors and Matrices The linear algebra portion of this course focuses on three matrix factorizations: QR factorization, singular valued decomposition (SVD), and LU factorization The first

Linear Algebra: Vectors

A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector

5.3 The Cross Product in R 3

53 The Cross Product in R 3 Definition 531 Let u = [u 1, u 2, u 3 ] and v = [v 1, v 2, v 3 ] Then the vector given by [u 2 v 3 u 3 v 2, u 3 v 1 u 1 v 3, u 1 v 2 u 2 v 1 ] is called the cross product (or

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

CHAPTER 12 MOLECULAR SYMMETRY

CHAPTER 12 MOLECULAR SYMMETRY In many cases, the symmetry of a molecule provides a great deal of information about its quantum states, even without a detailed solution of the Schrödinger equation. A geometrical

1. LINEAR EQUATIONS. A linear equation in n unknowns x 1, x 2,, x n is an equation of the form

1. LINEAR EQUATIONS A linear equation in n unknowns x 1, x 2,, x n is an equation of the form a 1 x 1 + a 2 x 2 + + a n x n = b, where a 1, a 2,..., a n, b are given real numbers. For example, with x and

1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

DETERMINANTS. b 2. x 2

DETERMINANTS 1 Systems of two equations in two unknowns A system of two equations in two unknowns has the form a 11 x 1 + a 12 x 2 = b 1 a 21 x 1 + a 22 x 2 = b 2 This can be written more concisely in

Lectures on Vector Calculus

Lectures on Vector Calculus Paul Renteln Department of Physics California State University San Bernardino, CA 92407 March, 2009; Revised March, 2011 c Paul Renteln, 2009, 2011 ii Contents 1 Vector Algebra

Unified Lecture # 4 Vectors

Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,

v 1 v 3 u v = (( 1)4 (3)2, [1(4) ( 2)2], 1(3) ( 2)( 1)) = ( 10, 8, 1) (d) u (v w) = (u w)v (u v)w (Relationship between dot and cross product)

0.1 Cross Product The dot product of two vectors is a scalar, a number in R. Next we will define the cross product of two vectors in 3-space. This time the outcome will be a vector in 3-space. Definition

Vector has a magnitude and a direction. Scalar has a magnitude

Vector has a magnitude and a direction Scalar has a magnitude Vector has a magnitude and a direction Scalar has a magnitude a brick on a table Vector has a magnitude and a direction Scalar has a magnitude

Recall that two vectors in are perpendicular or orthogonal provided that their dot

Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal

9 MATRICES AND TRANSFORMATIONS

9 MATRICES AND TRANSFORMATIONS Chapter 9 Matrices and Transformations Objectives After studying this chapter you should be able to handle matrix (and vector) algebra with confidence, and understand the

Lecture 3: Coordinate Systems and Transformations

Lecture 3: Coordinate Systems and Transformations Topics: 1. Coordinate systems and frames 2. Change of frames 3. Affine transformations 4. Rotation, translation, scaling, and shear 5. Rotation about an

How to add sine functions of different amplitude and phase

Physics 5B Winter 2009 How to add sine functions of different amplitude and phase In these notes I will show you how to add two sinusoidal waves each of different amplitude and phase to get a third sinusoidal

Introduction to Matrix Algebra

Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

Cross product 1 Chapter 7 Cross product We are getting ready to study integration in several variables. Until now we have been doing only differential calculus. One outcome of this study will be our ability

VECTOR CALCULUS: USEFUL STUFF Revision of Basic Vectors

Prof. S.M. Tobias Jan 2009 VECTOR CALCULUS: USEFUL STUFF Revision of Basic Vectors A scalar is a physical quantity with magnitude only A vector is a physical quantity with magnitude and direction A unit

1 Chapter 13. VECTORS IN THREE DIMENSIONAL SPACE Let s begin with some names and notation for things: R is the set (collection) of real numbers. We write x R to mean that x is a real number. A real number

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two

geometric transforms

geometric transforms 1 linear algebra review 2 matrices matrix and vector notation use column for vectors m 11 =[ ] M = [ m ij ] m 21 m 12 m 22 =[ ] v v 1 v = [ ] T v 1 v 2 2 3 matrix operations addition

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

Lecture 14: Section 3.3

Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar

On the general equation of the second degree

On the general equation of the second degree S Kesavan The Institute of Mathematical Sciences, CIT Campus, Taramani, Chennai - 600 113 e-mail:kesh@imscresin Abstract We give a unified treatment of the

Section 9.1 Vectors in Two Dimensions

Section 9.1 Vectors in Two Dimensions Geometric Description of Vectors A vector in the plane is a line segment with an assigned direction. We sketch a vector as shown in the first Figure below with an

Vector Representation of Rotations

Vector Representation of Rotations Carlo Tomasi The vector representation of rotation introduced below is based on Euler s theorem, and has three parameters. The conversion from a rotation vector to a

Rotation Matrices and Homogeneous Transformations

Rotation Matrices and Homogeneous Transformations A coordinate frame in an n-dimensional space is defined by n mutually orthogonal unit vectors. In particular, for a two-dimensional (2D) space, i.e., n

Diagonal, Symmetric and Triangular Matrices

Contents 1 Diagonal, Symmetric Triangular Matrices 2 Diagonal Matrices 2.1 Products, Powers Inverses of Diagonal Matrices 2.1.1 Theorem (Powers of Matrices) 2.2 Multiplying Matrices on the Left Right by

9.4. The Scalar Product. Introduction. Prerequisites. Learning Style. Learning Outcomes

The Scalar Product 9.4 Introduction There are two kinds of multiplication involving vectors. The first is known as the scalar product or dot product. This is so-called because when the scalar product of

Lecture 2. Observables

Lecture 2 Observables 13 14 LECTURE 2. OBSERVABLES 2.1 Observing observables We have seen at the end of the previous lecture that each dynamical variable is associated to a linear operator Ô, and its expectation

Linear Algebra I. Ronald van Luijk, 2012

Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.

MAT188H1S Lec0101 Burbulla

Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

Math 315: Linear Algebra Solutions to Midterm Exam I

Math 35: Linear Algebra s to Midterm Exam I # Consider the following two systems of linear equations (I) ax + by = k cx + dy = l (II) ax + by = 0 cx + dy = 0 (a) Prove: If x = x, y = y and x = x 2, y =

AP Physics - Vector Algrebra Tutorial

AP Physics - Vector Algrebra Tutorial Thomas Jefferson High School for Science and Technology AP Physics Team Summer 2013 1 CONTENTS CONTENTS Contents 1 Scalars and Vectors 3 2 Rectangular and Polar Form

28 CHAPTER 1. VECTORS AND THE GEOMETRY OF SPACE. v x. u y v z u z v y u y u z. v y v z

28 CHAPTER 1. VECTORS AND THE GEOMETRY OF SPACE 1.4 Cross Product 1.4.1 Definitions The cross product is the second multiplication operation between vectors we will study. The goal behind the definition

Notes on Determinant

ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL

SOLUTIONS OF THEORETICAL EXERCISES selected from INTRODUCTORY LINEAR ALGEBRA WITH APPLICATIONS B. KOLMAN, D. R. HILL Eighth Edition, Prentice Hall, 2005. Dr. Grigore CĂLUGĂREANU Department of Mathematics

Systems of Linear Equations

Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

Vector and Tensor Algebra (including Column and Matrix Notation)

Vector and Tensor Algebra (including Column and Matrix Notation) 2 Vectors and tensors In mechanics and other fields of physics, quantities are represented by vectors and tensors. Essential manipulations

MATH 304 Linear Algebra Lecture 4: Matrix multiplication. Diagonal matrices. Inverse matrix.

MATH 304 Linear Algebra Lecture 4: Matrix multiplication. Diagonal matrices. Inverse matrix. Matrices Definition. An m-by-n matrix is a rectangular array of numbers that has m rows and n columns: a 11

Structure of the Root Spaces for Simple Lie Algebras

Structure of the Root Spaces for Simple Lie Algebras I. Introduction A Cartan subalgebra, H, of a Lie algebra, G, is a subalgebra, H G, such that a. H is nilpotent, i.e., there is some n such that (H)

Vector Algebra II: Scalar and Vector Products

Chapter 2 Vector Algebra II: Scalar and Vector Products We saw in the previous chapter how vector quantities may be added and subtracted. In this chapter we consider the products of vectors and define

Lecture 1: Schur s Unitary Triangularization Theorem

Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections

GCE Mathematics (6360) Further Pure unit 4 (MFP4) Textbook

Version 36 klm GCE Mathematics (636) Further Pure unit 4 (MFP4) Textbook The Assessment and Qualifications Alliance (AQA) is a company limited by guarantee registered in England and Wales 364473 and a

Least-Squares Intersection of Lines

Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a

VECTOR ALGEBRA. 10.1.1 A quantity that has magnitude as well as direction is called a vector. is given by a and is represented by a.

VECTOR ALGEBRA Chapter 10 101 Overview 1011 A quantity that has magnitude as well as direction is called a vector 101 The unit vector in the direction of a a is given y a and is represented y a 101 Position