Notes on Groups S O(3) and S U(2)


 Bryce Page
 2 years ago
 Views:
Transcription
1 Notes on Groups S O(3) and S U() Ivan G. Avramidi New Mexico Institute of Mining and Technology October, 011 Author: IGA File: su.tex; Date: November 9, 011; Time: 10:07
2 1 1 Group S O(3) 1.1 Algebra so(3) The generators X i = X i, i = 1,, 3, of the algebra so(3) are 3 3 real antisymmetric matrices defined by or X 1 = , X = , X 3 = , (1.1) (X i ) k j = ε k i j = ε ik j (1.) Raising and lowering a vector index does not change anything; it is done just for convenience of notation. These matrices have a number of properties and where Therefore, they satisfy the algebra The Casimir operator is defined by X i X j = E ji Iδ i j, (1.3) ε i jk X k = E i j E ji (1.4) (E ji ) kl = δ k j δl i (1.5) [X i, X j ] = ε k i jx k (1.6) X = X i X i (1.7) It is equal to and commutes with all generators X = I (1.8) [X i, X ] = 0. (1.9) su.tex; November 9, 011; 10:07; p. 1
3 where A general element of the group S O(3) in canonical coordinates y i has the form exp (X(y)) = I cos r + P(1 cos r) + X(y) sin r r (1.10) = P + Π cos r + X(y) sin r, r (1.11) r = y i y i, θ i = yi r X(y) = X i y i, and the matrices Π and P are defined by In particular, for r = π we have (1.1) P i j = θ i θ j (1.13) Π i j = δ i j θ i θ j (1.14) exp (X(y)) = P Π (1.15) Notice that and, therefore, tr P = 1, tr Π = (1.16) tr exp (X(y)) = 1 + cos r. (1.17) Also, det sinh X(y) X(y) ( ) sin r = (1.18) r 1. Representations of S O(3) Let X i be the generators of S O(3) in a general irreducible representation satisfying the algebra [X i, X j ] = ε i jk X k (1.19) They are determined by symmetric traceless tensors of type ( j, j) (with j a positive integer) (X i ) l 1...l j k1...k j = jε (l 1 i(k1 δ l k δ l j) k j ) (1.0) Then X = j( j + 1)I (1.1) su.tex; November 9, 011; 10:07; p.
4 3 where Another basis of generators is (J +, J, J 3 ), where I l 1...l jk1...k j = δ (l 1 (k 1 δ l j) k j ) (1.) J + = ix 1 + X (1.3) J = ix 1 X (1.4) J 3 = ix 3 (1.5) so that J + = J, J 3 = J 3. (1.6) They satisfy the commutation relations [J +, J ] = J 3, [J 3, J + ] = J +, [J 3, J ] = J. (1.7) The Casimir operator in this basis is X = J + J J 3 + J 3 = J J + J 3 J 3 (1.8) From the commutation relations we have J 3 J + = J + (J 3 + 1) (1.9) J 3 J = J (J 3 1) (1.30) J J + = X J 3 J 3 (1.31) J + J = X J 3 + J 3 (1.3) Since J 3 commutes with X they can be diagonalized simultaneously. Since they are both Hermitian, they have real eigenvalues. Notice also that since X i are antihermitian, then X 0. (1.33) Let λ, m be the basis in which both operators are diagonal, that is, X λ, m = λ λ, m (1.34) J 3 λ, m = m λ, m (1.35) There are many ways to show that the eigenvalues m of the operator J 3 are integers. Now, since X 1 + X = X X 3 = X + J 3 0 (1.36) su.tex; November 9, 011; 10:07; p. 3
5 4 then the eigenvalues of J 3 are bounded by a maximal integer j, m j (1.37) that is, m takes ( j + 1) values j, j + 1,, 1, 0, 1,, j 1, j (1.38) From the commutation relations above we have J + λ, m = const λ, m + 1 (1.39) J λ, m = const λ, m 1 (1.40) Let λ, 0 be the eigenvector corresponding to the eigenvalue m = 0, that is J 3 λ, 0 = 0. (1.41) Then for positive m > 0 λ, m = a λm J+ m λ, 0 (1.4) λ, m = a λ, m J m λ, m (1.43) Then J + λ, j = 0 (1.44) J λ, j = 0 (1.45) Therefore, 0 = J J + λ, j = ( X J 3 J 3) λ, j (1.46) 0 = J + J λ, j = ( X J 3 + J 3) λ, j (1.47) Thus, ( λ j j ) λ, j = 0 (1.48) So, the eigenvalues of the Casimir operator X are labeled by a nonnegative integer j 0, λ j = j( j + 1). (1.49) su.tex; November 9, 011; 10:07; p. 4
6 5 These eigenvalues have the multiplicity d j = j + 1. (1.50) In particular, for j = 1 we get λ 1 = as expected before. From now on we will denote the basis in which the operators X and J 3 are diagonal by λ j, m = j. (1.51) m The matrix elements of the generators J ± can be computed as follows. First of all, j m J j 3 = δ m mm m (1.5) We have and This gives j j( j + 1) = m J j j + m 1 m 1 J j + m m (1.53) m j m J j j + = m 1 m 1 J j m j m J j + m j m J j m (1.54) = δ m,m 1 ( j + m)( j m + 1) (1.55) = δ m,m+1 ( j m)( j + m + 1) (1.56) The matrix elements of the generators X i are j m X j i 1 = δ m m,m 1 ( j + m)( j m + 1) i +δ m,m+1 ( j m)( j + m + 1) (1.57) j m X j 1 = δ m m,m 1 ( j + m)( j m + 1) 1 δ m,m+1 ( j m)( j + m + 1) (1.58) j m X j 3 = δ m mm im (1.59) su.tex; November 9, 011; 10:07; p. 5
7 6 Thus, irreducible representations of S O(3) are labeled by a nonnegative integer j 0. The matrices of the representation j in canonical coordinates y i will be denoted by j D j mm (y) = m exp [ X(y) ] j (1.60) m where X(y) = X i y i. The characters of the irreducible representations are χ j (y) = j Dmm(y) j = m= j Group S U() j e imr = 1 + m= j j cos(mr) (1.61) m=1.1 Algebra su() Pauli matrices σ i = σ i, i = 1,, 3, are defined by 0 1 σ 1 = 1 0, σ 0 i = i 0, σ 3 = They have the following properties (.1) σ i = σ i (.) σ i = I (.3) det σ i = 1 (.4) tr σ i = 0, (.5) σ 1 σ = iσ 3 (.6) σ σ 3 = iσ 1 (.7) σ 3 σ 1 = iσ (.8) σ 1 σ σ 3 = ii (.9) (.10) which can be written in the form σ i σ j = δ i j I + iε i jk σ k. (.11) su.tex; November 9, 011; 10:07; p. 6
8 7 They satisfy the following commutation relations The traces of products are Also, there holds [σ i, σ j ] = iε i jk σ k (.1) tr σ i σ j = δ i j (.13) tr σ i σ j σ k = iε i jk (.14) σ i σ i = 3I. (.15) The Pauli matrices satisfy the anticommutation relations σ i σ j + σ j σ i = δ i j I (.16) Therefore, Pauli matrices form a representation of Clifford algebra in 3 dimensions and are, in fact, Dirac matrices σ i = (σ i B ), where A, B = 1,, in 3 dimen A sions. The spinor metric E = (E AB ) is defined by 0 1 E = 1 0. (.17) Notice that and the inverse metric E 1 = (E AB ) is Then E = iσ (.18) E 1 = E (.19) σ T i = Eσ i E 1 (.0) This allows to define the matrices Eσ i = (σ i AB ) with covariant indices Eσ 1 = 1 0 σ 3 = 0 1, (.1) Eσ = i 0 ii = 0 i, (.) Eσ 3 = 0 1 σ 1 = 1 0. (.3) su.tex; November 9, 011; 10:07; p. 7
9 8 Notice that the matrices Eσ i are all symmetric (Eσ i ) T = Eσ i (.4) Pauli matrices satisfy the following completeness relation which implies This gives In a more symmetric form σ i A B σ i C D = δ A Dδ C B δ A Bδ C D (.5) σ i (A (B σ i C) D) = δ (A (Bδ C) D) (.6) σ i AB σ i CD = E AD E CB E AB E CD (.7) σ i AB σ i CD = E AD E CB + E BD E CA (.8) A spinor is a two dimensional complex vector ψ = (ψ A ), A = 1, in the spinor space C. We use the metric E to lower the index to get the covariant components which defines the cospinor ψ = (ψ A ) We also define the conjugated spinor ψ = ( ψ A ) by ψ A = E AB ψ B (.9) ψ = Eψ (.30) ψ A = ψ A (.31) On the complex spinor space there are two invariant bilinear forms ψ, ϕ = ψϕ = ψ A ϕ A (.3) and Note that and ψϕ = ψ A ϕ A = E AB ψ B ϕ A = ψ 1 ϕ ψ ϕ 1 (.33) ψ, ψ = ψψ = ψ A ψ A = ψ 1 + ψ (.34) ψψ = ψ A ψ A = 0 (.35) su.tex; November 9, 011; 10:07; p. 8
10 9 The matrices X i = i σ i (.36) are the generators of the Lie algebra su() with the commutation relations [X i, X j ] = ε i jk X k (.37) The Casimir operator is It commutes with all generators X = X i X i (.38) [X i, X ] = 0. (.39) There holds X = 3 4 I (.40) Notice that tr X i X j = 1 δ i j (.41) X = j( j + 1)I (.4) with j = 1, which determines the fundamental representation of S U(). An arbitrary element of the group S U() in canonical coordinates y i has the form ( i ) exp σ jy j = I cos r + iσ jθ j sin r, (.43) This can be written in the form exp (X(y)) = I cos(r/) + X(y) sin(r/). (.44) r/ where X(y) = X i y i. In particular, for r = π exp (X(y)) = I. (.45) Obviously, tr exp (X(y)) = cos(r/) (.46) su.tex; November 9, 011; 10:07; p. 9
11 10. Representations of S U() We define symmetric contravariant spinors ψ A 1...A j of rank j. Each index here can be lowered by the metric E AB. The number of independent components of such spinor is precisely j 1 = ( j + 1) (.47) i=0 Let X i be the generators of S U() in a general irreducible representation satisfying the algebra [X i, X j ] = ε i jk X k (.48) They are determined by symmetric spinors of type ( j, j) Then (X k ) A 1...A j B1...B j = i jσ k (A 1(B1 δ A B δ A j) B j ) (.49) X = j( j + 1)I (.50) where I A 1...A j B 1...B j = δ (A 1 (B 1 δ A j) B j ) (.51) Another basis of generators is J + = ix 1 + X (.5) J = ix 1 X (.53) J 3 = ix 3 (.54) They have all the properties of the generators of S O(3) discussed above except that the operator J 3 has integer eigenvalues m. One can show that they can be either integers or halfintegers taking ( j + 1) values j, j + 1,..., j 1, j, m j (.55) with j being a positive integer or halfinteger. If j is integer, then all eigenvalues m of J 3 are integers including 0. If j is a halfinteger, then all eigenvalues m are halfintegers and 0 is excluded. We choose a basis in which the operators X and J 3 are diagonal. Let ψ = ψ A 1...A j be a symmetric spinor of rank j. Let e 1 = (e A 1 ) and e = (e A ) be the basis cospinors defined by e 1 A = δ A 1, e A = δ A. (.56) su.tex; November 9, 011; 10:07; p. 10
12 11 Let e j,m A 1...A j+m B 1...B j m = e 1 (A1 e 1 A j+m e B1 e B j m ) = δ (A 1 1 δ A j+m 1 δ B 1 δb j m) (.57) Then this is an eigenspinor of the operator J 3 with the eigenvalue m Then the spinors J 3 e j,m = me j,m (.58) j ( ) 1/ ( j)! = e j,m (.59) m ( j + m)!( j m)! form an orthonormal basis in the space of symmetric spinors of rank j. The matrix elements of the generators X i are given by exactly the same formulas j m X j i 1 = δ m m,m 1 ( j + m)( j m + 1) i +δ m,m+1 ( j m)( j + m + 1) (.60) j m X j 1 = δ m m,m 1 ( j + m)( j m + 1) 1 δ m,m+1 ( j m)( j + m + 1) (.61) j m X j 3 = δ m mm im (.6) but now j and m are either both integers or halfintegers. Thus, irreducible representations of S U() are labeled by a nonnegative halfinteger j 0. The matrices of the representation j in canonical coordinates y i are j D j mm (y) = m exp [ X(y) ] j (.63) m where X(y) = X i y i. The representations of S U() with integer j are at the same time representations of S O(3). However, representations with halfinteger j do not give representations of S O(3), since for r = π D j mm (y) = δ mm. (.64) su.tex; November 9, 011; 10:07; p. 11
13 1 The characters of the irreducible representations are: for integer j χ j (y) = and for halfinteger j χ j (y) = 1 + j j Dmm(y) j = m= j k= 1 j D j 1 +k, 1 +k(y) = 1 + j j e imr = 1 + m= j k= 1 j exp j cos(mr) (.65) m=1 [ ( ) ] 1 i + k r = Now, let X i and Y j be the generators of two representations and let Then G i generate the product representation X Y and j k=1 [( ) ] 1 cos + k r (.66) G i = X i I Y + I X Y i (.67) exp[g(y)] = exp[x(y)] exp[y(y)] (.68) Therefore, the character of the product representation factorizes χ G (y) = χ X (y)χ Y (y) (.69).3 Double Covering Homomorphism S U() S O(3) Recall that the matrices Eσ i = (σ i AB ) are symmetric and the matrix E = (E AB ) is antisymmetric. Let ψ AB be a symmetric spinor. Then one can form a vector A i = σ i BA ψ AB = tr (Eσ i ψ). (.70) Conversely, given a vector A i we get a symmetric spinor of rank by ψ AB = 1 Ai σ i AB (.71) or ψ = 1 Ai σ i E 1 (.7) More generally, given a symmetric spinor of rank j we can associate to it a symmetric tensor of rank j by A i1...i j = σ i1 A 1 B 1 σ i j A j B j ψ A 1...A j B 1...B j (.73) su.tex; November 9, 011; 10:07; p. 1
14 13 One can show that this tensor is traceless. Indeed by using the eq. (??) and the fact that the spinor is symmetric we see that the tensor A is traceless. The inverse transformation is defined by ψ A1...A j B 1...B j = 1 j Ai 1...i j σ i1 A 1 B 1 σ i j A j B j (.74) The double covering homomorphism Λ : S U() S O(3) (.75) is defined as follows. Let U S U(). Then the matrices Uσ i U 1 satisfy all the properties of the Pauli matrices and are therefore in the algebra su(). Therefore, Uσ i U 1 = Λ j i(u)σ j. (.76) That is, Λ i j(u) = 1 tr σi Uσ j U 1 (.77) The matrix Λ(U) depends on the matrix U and is, in fact, in S O(3). For infinitesimal transformations U = exp(x k y k ) = I + i σ ky k + (.78) the matrix Λ is Λ i j = δ i j + ε i jk y k + = [ exp(x k y k ) ] i j (.79) It is easy to see that So, in short where Λ(I) = Λ( I) = I. (.80) Λ(exp(X i y i )) = exp(λ(x i )y i ) (.81) [Λ(σ k )] i j = iε i k j (.8) 3 Invariant Vector Fields Let y i be the canonical coordinates on the group S U() ranging over ( π, π). The position of the indices on the coordinates will be irrelevant, that is, y i = y i. We introduce the radial coordinate r = y = y i y i, so that the geodesic distance su.tex; November 9, 011; 10:07; p. 13
15 14 from the origin is equal to r, and the angular coordinates (the coordinates on S ) θ i = ŷ i = y i /r. Let us consider the fundamental representation of the group S U() with generators T a = iσ a, where σ a are the Pauli matrices satisfying the algebra Let T(y) = T i y i. Then where (y, p) = y i p i, (y p) k = ε i jk y i p j and Therefore, where ŷ is a unit vector. Next, we compute [T a, T b ] = ε c abt c. (3.83) T(y)T(p) = (y, p) T(y p) (3.84) [T(y)] = y I (3.85) exp[rt(ŷ)] = cos r + T(ŷ) sin r (3.86) exp[t(f(r ˆq, ρ ˆp))] = cos r cos ρ (ˆq, ˆp) sin r sin ρ + cos r sin ρt( ˆp) + cos ρ sin rt(ˆq) sin r sin ρt(ˆq ˆp) (3.87) This defines the group multiplication map F(q, p) in canonical coordinates. This map has a number of important properties. In particular, F(0, p) = F(p, 0) = p and F(p, p) = 0. Another obvious but very useful property of the group map is that if q = F(ω, p), then ω = F(q, p) and p = F( ω, q). Also, there is the associativity property and the inverse property F(ω, F(p, q)) = F(F(ω, p), q) (3.88) F( p, ω) = F(ω, p). (3.89) This map defines all the properties of the group. We introduce the matrices L α b(x) = y b Fα (y, x), (3.90) y=0 R α b(x) = y b Fα (x, y), (3.91) y=0 su.tex; November 9, 011; 10:07; p. 14
16 15 Let X and Y be the inverses of the matrices L and R, that is, Also let so that LX = XL = RY = YR = I. (3.9) D = XR, D 1 = YL (3.93) L = RD 1, X = DY (3.94) Let C a be the matrices of the adjoint representation defined by and C = C(x) be the matrix defined by Then one can show that X = exp(c) I C L = It is easy to see also that (C a ) b c = ε b ac (3.95) C = C a x a. (3.96), Y = I exp( C) C, (3.97) C exp(c) I, R = C I exp( C). (3.98) x a = L a bx b = R a bx b = X a bx b = Y a bx b. (3.99) and and as well as X = Y T, L = R T (3.100) ( ) sinh(c/) XY = YX =, C/ (3.101) ( ) C/ LR = RL =, sinh(c/) (3.10) D = XR = RX = exp(c), D T = YL = LY = exp( C). (3.103) su.tex; November 9, 011; 10:07; p. 15
17 16 Note that where r = x and with θ i = x i /r. Therefore, C = 4r Π, (3.104) Π i j = δ i j θ i θ j (3.105) C n = ( 1) n (r) n Π, C n+1 = ( 1) n (r) n C (3.106) Therefore, the eigenvalues of the matrix C are (ir, ir, 0), and for any analytic function f (C) = f (0)(I Π) + Π 1 and, therefore, This enables one to compute Then one can show that [ f (ir) + f ( ir) ] + 1 4ir C [ f (ir) f ( ir ] tr f (C) = f (0) + f (ir) + f ( ir) (3.107) det f (C) = f (0) f (ir) f ( ir) (3.108) Y = I Π + sin r r X = I Π + sin r r cos rπ 1 sin r cos rπ + 1 C (3.109) r sin r C (3.110) r R = I Π + r cot rπ + 1 C (3.111) L = I Π + r cot rπ 1 C (3.11) D = I Π + Π cos(r) + sin(r) C (3.113) r x µ Fα (x, y) = L α γ(f(x, y))x γ µ(x), (3.114) x µ Fα (y, x) = R α γ(f(y, x))y γ µ(x). (3.115) These matrices satisfy important identities µ X a ν ν X a µ = ε a bcx b µx c ν, (3.116) µ Y a ν ν Y a µ = ε a bcy b µy c ν. (3.117) su.tex; November 9, 011; 10:07; p. 16
18 17 and Let us define the oneforms They satisfy important identities Next, we define the vector fields. L µ a µ L ν b L µ b µ L ν a = ε c abl γ c, (3.118) R µ a µ R ν b R µ b µ R ν a = ε c abr ν c. (3.119) σ a = X a µ(x)dx µ, σ a + = Y a µ(x)dx µ. dσ a = ε a bcσ b σ c, (3.10) dσ a + = ε a bcσ b + σ c +. (3.11) K a = L µ a(x) µ, K + a = R µ a(x) µ. These are the rightinvariant and the leftinvariant vector fields on the group. They are related by K a = D a b K + b (3.1) They form two representations of the group S U() and satisfy the algebra [K + a, K + b ] = εc abk + c (3.13) [K a, K b ] = εc abk c (3.14) [K + a, K b ] = 0 (3.15) That is, the leftinvariant vector fields commute with the rightinvariant ones. Now, let T a be the generators of some representation of the group S U() satisfying the same algebra. Then, of course, exp[t(ω)] exp[t(p)] = exp[t(f(ω, p))] (3.16) First, one can derive a useful commutation formula exp[t(ω)]t b exp[ T(ω)] = D a bt a, (3.17) where the matrix D = (D a b) is defined by D = exp C. More generally, exp[t(ω)] exp[t(p)] exp[ T(ω)] = exp[t(d(ω)p)]. (3.18) su.tex; November 9, 011; 10:07; p. 17
19 18 Next, one can show that This means that exp[ T(ω)] ω i exp[t(ω)] = Ya it a, (3.19) ω i exp[t(ω)] = Ya i exp[t(ω)]t a = Y i a T a exp[t(ω)], (3.130) The Killing vectors have important properties K + a exp[t(ω)] = (exp[t(ω)])t a (3.131) K a exp[t(ω)] = T a (exp[t(ω)]) (3.13) This immediately leads to further important equations and exp[k + (p)] exp[t(ω)] = exp[t(ω)] exp[t(p)] (3.133) exp[k (q)] exp[t(ω)] = exp[t(q)] exp[t(ω)] (3.134) where K ± (p) = p a K ± a. More generally, exp[k + (p) + K (q)] exp[t(ω)] = exp[t(q)] exp[t(ω)] exp[t(p)] (3.135) and, therefore, exp[k + (p) + K (q)] exp[t(ω)] = exp[t(q)] exp[t(p)] = exp[t(f(q, p))] ω=0 (3.136) In particular, for the adjoint representation this formulas take the form DC b = D a bc a D (3.137) K + a D = DC a (3.138) K a D = C a D (3.139) exp[k + (p)]d(x) = D(x)D(p) (3.140) exp[k (q)]d(x) = D(q)D(x) (3.141) exp[k + (p) + K (q)]d(x) = D(q)D(x)D(p) (3.14) su.tex; November 9, 011; 10:07; p. 18
20 19 Then for a scalar function f (ω) the action of the rightinvariant and leftinvariant vector fields is simply exp[k + (p)] f (ω) = f (F(ω, p)), exp[k (q)] f (ω) = f (F(q, ω)). (3.143) Since the action of these vector fields commutes, we have more generally Therefore, that is, exp[k + (p) + K (q)] f (ω) = f (F(q, F(ω, p))). (3.144) exp[k + (p) + K (q)] f (ω) = f (F(q, p)). (3.145) ω=0 The biinvariant metric is defined by g αβ = δ ab Y a αy b β = δ ab X a αx b β (3.146) g αβ = δ ab L α al β b = δ ab R α ar β b, (3.147) ( ) sinh[c/] g = = I Π + sin r Π. (3.148) C/ r The Riemannian volume elements of the metric g is defined as usual where dvol = g 1/ (x)dx 1 dx dx 3 (3.149) g 1/ = (det g µν ) 1/ = sin r r. (3.150) Thus the biinvariant volume element of the group is dvol (x) = sin r r dx = sin r dr dθ (3.151) where dx = dx 1 dx dx 3, and dθ is the volume element on S. The invariance of the volume element means that for any fixed p dvol (x) = dvol (F(x, p)) = dvol (F(p, x)) (3.15) Notice also, that F(p, q) is nothing but the geodesic distance between the points p and q. su.tex; November 9, 011; 10:07; p. 19
21 0 We will always use the orthonormal basis of the rightinvariant vector fields K + a and oneforms σ a + and denote the covariant derivative with respect to K + a simply by a. The LeviCivita connection of the biinvariant metric in the rightinvariant basis is defined by a K + b = εc bak + c (3.153) so that the coefficients of the affine connection are so that for an arbitrary vector ω a bc = σ a +( c K + b ) = εa bc. (3.154) a V = [ (K + a V b ) + ε b cav c] K + b (3.155) Then where a K d = Mb dak + b (3.156) M b da = K + a D d b + ε b cad d c = ε b cad d c + ε b cad d c = ε b cad d c (3.157) so that Now, the curvature tensor is The Ricci curvature tensor and the scalar curvature are The scalar Laplacian is One can show that Further, a K d = εb cad d c K + b (3.158) R a bcd = ε f cdε a f b. (3.159) R ab = δ ab, R = 6. (3.160) 0 = δ ab K a K b = δab K + a K + b. (3.161) 0 ω sin ω = ω sin ω (3.16) 0 exp[t(ω)] = (exp[t(ω)])t = T exp[t(ω)] (3.163) su.tex; November 9, 011; 10:07; p. 0
22 1 where T = T a T a. This means that exp[t 0 ] exp[t(ω)] = exp(tt ) exp[t(ω)] (3.164) The Killing vector fields K a ± are divergence free, which means that they are antiselfadjoint with respect to the invariant measure dvol (ω). that is, Thus the Laplacian 0 is selfadjoint with respect to the same measure as well. Let be the total connection on a vector bundle V realizing the representation G. Let us define the oneforms A = σ a +G a = G a Y a µdx µ. (3.165) and F = da + A A. Then, by using the above properties of the rightinvariant oneforms we compute where F = 1 F abσ a + σ b + = 1 F aby a µy b νdx µ dx ν (3.166) F ab = ε c abg c (3.167) Thus A is the YangMills curvature on the vector bundle V with covariantly constant curvature. The covariant derivative of a section of the bundle V is then (recall that a = K + a ) a ϕ = (K + a + G a )ϕ (3.168) and the covariant derivative of a section of the endomorphism bundle End(V) along the rightinvariant basis is then Therefore, we find a Q = K + a Q + [G a, Q] (3.169) a G b = ε c bag c + [G a, G b ] = 0. (3.170) Then the derivatives along the leftinvariant vector fields are K a ϕ = (K a + B a )ϕ (3.171) where K a Q = K a Q + [B a, Q] (3.17) B a = D a b G b (3.173) su.tex; November 9, 011; 10:07; p. 1
23 The Laplacian takes the form = a a = (K + a + G a )(K + a + G a ) = 0 + G a K + a + G (3.174) where G = G i G i. We want to rewrite the Laplacian in terms of Casimir operators of some representations of the group S U(). The covariant derivatives a do not form a representation of the algebra S U(). The operators that do are the covariant Lie derivatives. The covariant Lie derivatives along a Killing vector ξ of sections of this vector bundle are defined by L ξ = ξ 1 σa +( b ξ)ε bc ag c (3.175) By denoting the Lie derivatives along the Killing vectors K ± a by K ± a this gives for the rightinvariant and the leftinvariant bases K + a = L K + a = a + G a = K + a + G a. (3.176) K a = L K a = K a B a = K a. (3.177) It is easy to see that these operators form the same algebra S U() S U() [K + a, K + b ] = εc abk + c (3.178) [K a, K b ] = εc abk c (3.179) [K + a, K b ] = 0 (3.180) Now, the Laplacian is given now by the sums of the Casimir operators = K G (3.181) where K = 1 K K (3.18) and K ± = K ± a K ± a. 4 Heat Kernel on S 3 Our goal is to evaluate the heat kernel diagonal. Since it is constant we can evaluate it at any point, say, at the origin. We use the geodesic coordinates y i defined su.tex; November 9, 011; 10:07; p.
24 3 above. That is why, we will evaluate the heat kernel when one point is fixed at the origin. We will denote it simply by U(t; y). We obviously have exp(t ) = exp ( tg ) exp ( tk ). (4.1) Therefore, the heat kernel is equal to U(t, y) = exp ( tg ) ( t ) Ψ, y (4.) where Ψ(t, y) = exp(tk )δ(y) is the heat kernel of the operator K. 4.1 Scalar Heat Kernel Let Φ(t, ω) = (4πt) 3/ e t n= ω + πn sin ω ) ( ω + πn) exp ( 4t One can show by direct calculation that this function satisfies the equation (4.3) t Φ = 0 Φ (4.4) with the initial condition Φ(0, ω) = δ S 3(ω) (4.5) Therefore, this is the scalar heat kernel on the group S U() and, therefore, on S 3. Now, let us compute the integral Ψ(t) = dvol (ω)φ(t, ω) exp[t(ω)] (4.6) S U() Obviously, Ψ(0) = 1. Next, we have t Ψ = dvol (ω) exp[t(ω)] 0 Φ(t, ω) (4.7) S U() Now, by integrating by parts we get t Ψ = DωΦ(t, ω) 0 exp[t(ω)] (4.8) S U() which gives t Ψ = ΨT (4.9) su.tex; November 9, 011; 10:07; p. 3
25 4 Thus, we obtain a very important equation exp(tt ) = dvol (ω) Φ(t, ω) exp[t(ω)] (4.10) S U() This is true for any representation of S U(). We derive two corollaries. First, we get exp(tt ) = dvol (ω) Φ(t, ω) exp[t(p)] exp[t(ω)] exp[ T(p)] (4.11) S U() which means that for any p or Also, we see that exp[(t + s)t ] = Φ(t, ω) = Φ(t, F( p, F(ω, p)). (4.1) Φ(t, F(p, q)) = Φ(t, F(q, p)). (4.13) S 3 S 3 dvol (q)dvol (p)φ(t, q)φ(s, p) exp[t(q)] exp[t(p)] (4.14) By changing the variables z = F(q, p) and p p we obtain exp[(t + s)t ] = dvol (z) dvol (p)φ(t, F(p, z))φ(s, p) exp[t(z)] which means that S U() S U() S U() (4.15) dvol (p)φ(t, F(p, z))φ(s, p) = Φ(t + s, z) (4.16) In particular, for z = 0 and s = t this becomes dvol (p)φ(t, p)φ(t, p) = Φ(t, 0) (4.17) S U() su.tex; November 9, 011; 10:07; p. 4
Chapter 9 Unitary Groups and SU(N)
Chapter 9 Unitary Groups and SU(N) The irreducible representations of SO(3) are appropriate for describing the degeneracies of states of quantum mechanical systems which have rotational symmetry in three
More informationUNIT 2 MATRICES  I 2.0 INTRODUCTION. Structure
UNIT 2 MATRICES  I Matrices  I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress
More information1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
More informationThe Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
More informationTensors on a vector space
APPENDIX B Tensors on a vector space In this Appendix, we gather mathematical definitions and results pertaining to tensors. The purpose is mostly to introduce the modern, geometrical view on tensors,
More informationMath 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
More informationProblem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday.
Math 312, Fall 2012 Jerry L. Kazdan Problem Set 5 Due: In class Thursday, Oct. 18 Late papers will be accepted until 1:00 PM Friday. In addition to the problems below, you should also know how to solve
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 19967 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationINVARIANT METRICS WITH NONNEGATIVE CURVATURE ON COMPACT LIE GROUPS
INVARIANT METRICS WITH NONNEGATIVE CURVATURE ON COMPACT LIE GROUPS NATHAN BROWN, RACHEL FINCK, MATTHEW SPENCER, KRISTOPHER TAPP, AND ZHONGTAO WU Abstract. We classify the leftinvariant metrics with nonnegative
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationInvariant Metrics with Nonnegative Curvature on Compact Lie Groups
Canad. Math. Bull. Vol. 50 (1), 2007 pp. 24 34 Invariant Metrics with Nonnegative Curvature on Compact Lie Groups Nathan Brown, Rachel Finck, Matthew Spencer, Kristopher Tapp and Zhongtao Wu Abstract.
More informationarxiv:math/0010057v1 [math.dg] 5 Oct 2000
arxiv:math/0010057v1 [math.dg] 5 Oct 2000 HEAT KERNEL ASYMPTOTICS FOR LAPLACE TYPE OPERATORS AND MATRIX KDV HIERARCHY IOSIF POLTEROVICH Preliminary version Abstract. We study the heat kernel asymptotics
More informationStructure of the Root Spaces for Simple Lie Algebras
Structure of the Root Spaces for Simple Lie Algebras I. Introduction A Cartan subalgebra, H, of a Lie algebra, G, is a subalgebra, H G, such that a. H is nilpotent, i.e., there is some n such that (H)
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationThe threedimensional rotations are defined as the linear transformations of the vector x = (x 1, x 2, x 3 ) x i = R ij x j, (2.1) x 2 = x 2. (2.
2 The rotation group In this Chapter we give a short account of the main properties of the threedimensional rotation group SO(3) and of its universal covering group SU(2). The group SO(3) is an important
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationx1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.
Cross product 1 Chapter 7 Cross product We are getting ready to study integration in several variables. Until now we have been doing only differential calculus. One outcome of this study will be our ability
More informationDecember 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in twodimensional space (1) 2x y = 3 describes a line in twodimensional space The coefficients of x and y in the equation
More informationSection 6.1  Inner Products and Norms
Section 6.1  Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More informationCONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation
Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in
More informationMATH36001 Background Material 2015
MATH3600 Background Material 205 Matrix Algebra Matrices and Vectors An ordered array of mn elements a ij (i =,, m; j =,, n) written in the form a a 2 a n A = a 2 a 22 a 2n a m a m2 a mn is said to be
More informationDATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
More informationThe Matrix Elements of a 3 3 Orthogonal Matrix Revisited
Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, ThreeDimensional Proper and Improper Rotation Matrices, I provided a derivation
More informationTHREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
More information13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in threespace, we write a vector in terms
More informationEigenvalues and Eigenvectors
Chapter 6 Eigenvalues and Eigenvectors 6. Introduction to Eigenvalues Linear equations Ax D b come from steady state problems. Eigenvalues have their greatest importance in dynamic problems. The solution
More informationIntroduction to Matrix Algebra I
Appendix A Introduction to Matrix Algebra I Today we will begin the course with a discussion of matrix algebra. Why are we studying this? We will use matrix algebra to derive the linear regression model
More information9 MATRICES AND TRANSFORMATIONS
9 MATRICES AND TRANSFORMATIONS Chapter 9 Matrices and Transformations Objectives After studying this chapter you should be able to handle matrix (and vector) algebra with confidence, and understand the
More information1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More information4. MATRICES Matrices
4. MATRICES 170 4. Matrices 4.1. Definitions. Definition 4.1.1. A matrix is a rectangular array of numbers. A matrix with m rows and n columns is said to have dimension m n and may be represented as follows:
More informationA Primer on Index Notation
A Primer on John Crimaldi August 28, 2006 1. Index versus Index notation (a.k.a. Cartesian notation) is a powerful tool for manipulating multidimensional equations. However, there are times when the more
More informationequations Karl Lundengård December 3, 2012 MAA704: Matrix functions and matrix equations Matrix functions Matrix equations Matrix equations, cont d
and and Karl Lundengård December 3, 2012 Solving General, Contents of todays lecture and (Kroenecker product) Solving General, Some useful from calculus and : f (x) = x n, x C, n Z + : f (x) = n x, x R,
More informationMATH PROBLEMS, WITH SOLUTIONS
MATH PROBLEMS, WITH SOLUTIONS OVIDIU MUNTEANU These are free online notes that I wrote to assist students that wish to test their math skills with some problems that go beyond the usual curriculum. These
More informationThe Solution of Linear Simultaneous Equations
Appendix A The Solution of Linear Simultaneous Equations Circuit analysis frequently involves the solution of linear simultaneous equations. Our purpose here is to review the use of determinants to solve
More informationMATH 240 Fall, Chapter 1: Linear Equations and Matrices
MATH 240 Fall, 2007 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 9th Ed. written by Prof. J. Beachy Sections 1.1 1.5, 2.1 2.3, 4.2 4.9, 3.1 3.5, 5.3 5.5, 6.1 6.3, 6.5, 7.1 7.3 DEFINITIONS
More informationApplied Linear Algebra I Review page 1
Applied Linear Algebra Review 1 I. Determinants A. Definition of a determinant 1. Using sum a. Permutations i. Sign of a permutation ii. Cycle 2. Uniqueness of the determinant function in terms of properties
More informationLecture 2 Matrix Operations
Lecture 2 Matrix Operations transpose, sum & difference, scalar multiplication matrix multiplication, matrixvector product matrix inverse 2 1 Matrix transpose transpose of m n matrix A, denoted A T or
More informationLinear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices
MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two
More informationEC9A0: Presessional Advanced Mathematics Course
University of Warwick, EC9A0: Presessional Advanced Mathematics Course Peter J. Hammond & Pablo F. Beker 1 of 55 EC9A0: Presessional Advanced Mathematics Course Slides 1: Matrix Algebra Peter J. Hammond
More informationMathematics Notes for Class 12 chapter 3. Matrices
1 P a g e Mathematics Notes for Class 12 chapter 3. Matrices A matrix is a rectangular arrangement of numbers (real or complex) which may be represented as matrix is enclosed by [ ] or ( ) or Compact form
More informationLinear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University
Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone
More informationThe cover SU(2) SO(3) and related topics
The cover SU(2) SO(3) and related topics Iordan Ganev December 2011 Abstract The subgroup U of unit quaternions is isomorphic to SU(2) and is a double cover of SO(3). This allows a simple computation of
More informationThe Greatest Common Factor; Factoring by Grouping
296 CHAPTER 5 Factoring and Applications 5.1 The Greatest Common Factor; Factoring by Grouping OBJECTIVES 1 Find the greatest common factor of a list of terms. 2 Factor out the greatest common factor.
More informationChapter 7. Matrices. Definition. An m n matrix is an array of numbers set out in m rows and n columns. Examples. ( 1 1 5 2 0 6
Chapter 7 Matrices Definition An m n matrix is an array of numbers set out in m rows and n columns Examples (i ( 1 1 5 2 0 6 has 2 rows and 3 columns and so it is a 2 3 matrix (ii 1 0 7 1 2 3 3 1 is a
More informationSolutions to Linear Algebra Practice Problems
Solutions to Linear Algebra Practice Problems. Find all solutions to the following systems of linear equations. (a) x x + x 5 x x x + x + x 5 (b) x + x + x x + x + x x + x + 8x Answer: (a) We create the
More informationFacts About Eigenvalues
Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v
More informationNotes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 201112 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
More informationDEFINITION 5.1.1 A complex number is a matrix of the form. x y. , y x
Chapter 5 COMPLEX NUMBERS 5.1 Constructing the complex numbers One way of introducing the field C of complex numbers is via the arithmetic of matrices. DEFINITION 5.1.1 A complex number is a matrix of
More informationMAT188H1S Lec0101 Burbulla
Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u
More informationThe basic unit in matrix algebra is a matrix, generally expressed as: a 11 a 12. a 13 A = a 21 a 22 a 23
(copyright by Scott M Lynch, February 2003) Brief Matrix Algebra Review (Soc 504) Matrix algebra is a form of mathematics that allows compact notation for, and mathematical manipulation of, highdimensional
More informationUniversity of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
More informationMath 550 Notes. Chapter 7. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010
Math 550 Notes Chapter 7 Jesse Crawford Department of Mathematics Tarleton State University Fall 2010 (Tarleton State University) Math 550 Chapter 7 Fall 2010 1 / 34 Outline 1 SelfAdjoint and Normal Operators
More informationSection 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
More informationC 1 x(t) = e ta C = e C n. 2! A2 + t3
Matrix Exponential Fundamental Matrix Solution Objective: Solve dt A x with an n n constant coefficient matrix A x (t) Here the unknown is the vector function x(t) x n (t) General Solution Formula in Matrix
More informationInner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 34 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
More informationFinite dimensional C algebras
Finite dimensional C algebras S. Sundar September 14, 2012 Throughout H, K stand for finite dimensional Hilbert spaces. 1 Spectral theorem for selfadjoint opertors Let A B(H) and let {ξ 1, ξ 2,, ξ n
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationTopic 1: Matrices and Systems of Linear Equations.
Topic 1: Matrices and Systems of Linear Equations Let us start with a review of some linear algebra concepts we have already learned, such as matrices, determinants, etc Also, we shall review the method
More information1 Introduction. 2 Matrices: Definition. Matrix Algebra. Hervé Abdi Lynne J. Williams
In Neil Salkind (Ed.), Encyclopedia of Research Design. Thousand Oaks, CA: Sage. 00 Matrix Algebra Hervé Abdi Lynne J. Williams Introduction Sylvester developed the modern concept of matrices in the 9th
More informationEigenvalues, Eigenvectors, Matrix Factoring, and Principal Components
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationIntroduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra  1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
More informationFactoring of Prime Ideals in Extensions
Chapter 4 Factoring of Prime Ideals in Extensions 4. Lifting of Prime Ideals Recall the basic AKLB setup: A is a Dedekind domain with fraction field K, L is a finite, separable extension of K of degree
More informationQuantum Mechanics and Representation Theory
Quantum Mechanics and Representation Theory Peter Woit Columbia University Texas Tech, November 21 2013 Peter Woit (Columbia University) Quantum Mechanics and Representation Theory November 2013 1 / 30
More information1 Scalars, Vectors and Tensors
DEPARTMENT OF PHYSICS INDIAN INSTITUTE OF TECHNOLOGY, MADRAS PH350 Classical Physics Handout 1 8.8.2009 1 Scalars, Vectors and Tensors In physics, we are interested in obtaining laws (in the form of mathematical
More information1 Short Introduction to Time Series
ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The
More informationBILINEAR FORMS KEITH CONRAD
BILINEAR FORMS KEITH CONRAD The geometry of R n is controlled algebraically by the dot product. We will abstract the dot product on R n to a bilinear form on a vector space and study algebraic and geometric
More informationA QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS
A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors
More informationAdding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors
1 Chapter 13. VECTORS IN THREE DIMENSIONAL SPACE Let s begin with some names and notation for things: R is the set (collection) of real numbers. We write x R to mean that x is a real number. A real number
More information2. Introduction to quantum mechanics
2. Introduction to quantum mechanics 2.1 Linear algebra Dirac notation Complex conjugate Vector/ket Dual vector/bra Inner product/bracket Tensor product Complex conj. matrix Transpose of matrix Hermitian
More informationSemiSimple Lie Algebras and Their Representations
i SemiSimple Lie Algebras and Their Representations Robert N. Cahn Lawrence Berkeley Laboratory University of California Berkeley, California 1984 THE BENJAMIN/CUMMINGS PUBLISHING COMPANY Advanced Book
More information2.1: MATRIX OPERATIONS
.: MATRIX OPERATIONS What are diagonal entries and the main diagonal of a matrix? What is a diagonal matrix? When are matrices equal? Scalar Multiplication 45 Matrix Addition Theorem (pg 0) Let A, B, and
More informationMATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar
More informationSection 8.8. 1. The given line has equations. x = 3 + t(13 3) = 3 + 10t, y = 2 + t(3 + 2) = 2 + 5t, z = 7 + t( 8 7) = 7 15t.
. The given line has equations Section 8.8 x + t( ) + 0t, y + t( + ) + t, z 7 + t( 8 7) 7 t. The line meets the plane y 0 in the point (x, 0, z), where 0 + t, or t /. The corresponding values for x and
More informationQuotient Rings and Field Extensions
Chapter 5 Quotient Rings and Field Extensions In this chapter we describe a method for producing field extension of a given field. If F is a field, then a field extension is a field K that contains F.
More informationSolution to Homework 2
Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if
More informationInverses. Stephen Boyd. EE103 Stanford University. October 27, 2015
Inverses Stephen Boyd EE103 Stanford University October 27, 2015 Outline Left and right inverses Inverse Solving linear equations Examples Pseudoinverse Left and right inverses 2 Left inverses a number
More information10. Graph Matrices Incidence Matrix
10 Graph Matrices Since a graph is completely determined by specifying either its adjacency structure or its incidence structure, these specifications provide far more efficient ways of representing a
More informationLinear Algebra: Matrices
B Linear Algebra: Matrices B 1 Appendix B: LINEAR ALGEBRA: MATRICES TABLE OF CONTENTS Page B.1. Matrices B 3 B.1.1. Concept................... B 3 B.1.2. Real and Complex Matrices............ B 3 B.1.3.
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationMATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix.
MATH 304 Linear Algebra Lecture 8: Inverse matrix (continued). Elementary matrices. Transpose of a matrix. Inverse matrix Definition. Let A be an n n matrix. The inverse of A is an n n matrix, denoted
More informationNOTES ON MINIMAL SURFACES
NOTES ON MINIMAL SURFACES DANNY CALEGARI Abstract. These are notes on minimal surfaces, with an emphasis on the classical theory and its connection to complex analysis, and the topological applications
More information(a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular.
Theorem.7.: (Properties of Triangular Matrices) (a) The transpose of a lower triangular matrix is upper triangular, and the transpose of an upper triangular matrix is lower triangular. (b) The product
More informationCONNECTIONS ON PRINCIPAL GBUNDLES
CONNECTIONS ON PRINCIPAL GBUNDLES RAHUL SHAH Abstract. We will describe connections on principal Gbundles via two perspectives: that of distributions and that of connection 1forms. We will show that
More informationMatrix Norms. Tom Lyche. September 28, Centre of Mathematics for Applications, Department of Informatics, University of Oslo
Matrix Norms Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 28, 2009 Matrix Norms We consider matrix norms on (C m,n, C). All results holds for
More informationLecture 2: Essential quantum mechanics
Department of Physical Sciences, University of Helsinki http://theory.physics.helsinki.fi/ kvanttilaskenta/ p. 1/46 Quantum information and computing Lecture 2: Essential quantum mechanics JaniPetri Martikainen
More informationInverses and powers: Rules of Matrix Arithmetic
Contents 1 Inverses and powers: Rules of Matrix Arithmetic 1.1 What about division of matrices? 1.2 Properties of the Inverse of a Matrix 1.2.1 Theorem (Uniqueness of Inverse) 1.2.2 Inverse Test 1.2.3
More informationLecture Notes for Ph219/CS219: Quantum Information and Computation Chapter 2. John Preskill California Institute of Technology
Lecture Notes for Ph19/CS19: Quantum Information and Computation Chapter John Preskill California Institute of Technology Updated July 015 Contents Foundations I: States and Ensembles 3.1 Axioms of quantum
More informationContinuous Groups, Lie Groups, and Lie Algebras
Chapter 7 Continuous Groups, Lie Groups, and Lie Algebras Zeno was concerned with three problems... These are the problem of the infinitesimal, the infinite, and continuity... Bertrand Russell The groups
More information1.4. Removing Brackets. Introduction. Prerequisites. Learning Outcomes. Learning Style
Removing Brackets 1. Introduction In order to simplify an expression which contains brackets it is often necessary to rewrite the expression in an equivalent form but without any brackets. This process
More informationExam 1 Sample Question SOLUTIONS. y = 2x
Exam Sample Question SOLUTIONS. Eliminate the parameter to find a Cartesian equation for the curve: x e t, y e t. SOLUTION: You might look at the coordinates and notice that If you don t see it, we can
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationThe Hadamard Product
The Hadamard Product Elizabeth Million April 12, 2007 1 Introduction and Basic Results As inexperienced mathematicians we may have once thought that the natural definition for matrix multiplication would
More informationLinear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
More informationThe Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression
The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression The SVD is the most generally applicable of the orthogonaldiagonalorthogonal type matrix decompositions Every
More informationSimilar matrices and Jordan form
Similar matrices and Jordan form We ve nearly covered the entire heart of linear algebra once we ve finished singular value decompositions we ll have seen all the most central topics. A T A is positive
More informationLecture 18  Clifford Algebras and Spin groups
Lecture 18  Clifford Algebras and Spin groups April 5, 2013 Reference: Lawson and Michelsohn, Spin Geometry. 1 Universal Property If V is a vector space over R or C, let q be any quadratic form, meaning
More informationPUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5.
PUTNAM TRAINING POLYNOMIALS (Last updated: November 17, 2015) Remark. This is a list of exercises on polynomials. Miguel A. Lerma Exercises 1. Find a polynomial with integral coefficients whose zeros include
More information3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More information