C. J. Papachristou MATHEMATICAL HANDBOOK

Similar documents
Mathematics Course 111: Algebra I Part IV: Vector Spaces

1 Introduction to Matrices

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

Matrix Representations of Linear Transformations and Changes of Coordinates

Vector and Matrix Norms

NOTES ON LINEAR TRANSFORMATIONS

Abstract Algebra Cheat Sheet

1 Sets and Set Notation.

13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

by the matrix A results in a vector which is a reflection of the given

Similarity and Diagonalization. Similar Matrices

9 MATRICES AND TRANSFORMATIONS

University of Lille I PC first year list of exercises n 7. Review

( ) which must be a vector

The Characteristic Polynomial

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

GROUP ALGEBRAS. ANDREI YAFAEV

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATH PROBLEMS, WITH SOLUTIONS

MA106 Linear Algebra lecture notes

Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)

Introduction to Matrix Algebra

Orthogonal Diagonalization of Symmetric Matrices

Chapter 13: Basic ring theory

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

Linear Algebra Review. Vectors

4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION

Inner Product Spaces and Orthogonality

Name: Section Registered In:

Rotation Matrices and Homogeneous Transformations

NOV /II. 1. Let f(z) = sin z, z C. Then f(z) : 3. Let the sequence {a n } be given. (A) is bounded in the complex plane

1 VECTOR SPACES AND SUBSPACES

4.5 Linear Dependence and Linear Independence

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

1 Homework 1. [p 0 q i+j p i 1 q j+1 ] + [p i q j ] + [p i+1 q j p i+j q 0 ]

Continuous Groups, Lie Groups, and Lie Algebras

Chapter 17. Orthogonal Matrices and Symmetries of Space

Lecture 2 Matrix Operations

Chapter 20. Vector Spaces and Bases

x = + x 2 + x

A =

Lecture L3 - Vectors, Matrices and Coordinate Transformations

Notes on Symmetric Matrices

LINEAR ALGEBRA W W L CHEN

Linear Algebra Notes

Using row reduction to calculate the inverse and the determinant of a square matrix

T ( a i x i ) = a i T (x i ).

LINEAR ALGEBRA. September 23, 2010

IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction

GROUPS ACTING ON A SET

3. Let A and B be two n n orthogonal matrices. Then prove that AB and BA are both orthogonal matrices. Prove a similar result for unitary matrices.

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Applied Linear Algebra I Review page 1

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Notes on Determinant

Finite dimensional C -algebras

Nilpotent Lie and Leibniz Algebras

Geometric Transformations

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

MAT188H1S Lec0101 Burbulla

Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University

Section Inner Products and Norms

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Projective Geometry: A Short Introduction. Lecture Notes Edmond Boyer

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

Vector Spaces. Chapter R 2 through R n

6 Commutators and the derived series. [x,y] = xyx 1 y 1.

Numerical Analysis Lecture Notes

Linearly Independent Sets and Linearly Dependent Sets

Mathematics Review for MS Finance Students

Brief Introduction to Vectors and Matrices

Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj

Systems of Linear Equations

BILINEAR FORMS KEITH CONRAD

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

MATH1231 Algebra, 2015 Chapter 7: Linear maps

Notes on Linear Algebra. Peter J. Cameron

Linear Algebra: Determinants, Inverses, Rank

Introduction to Matrices for Engineers

Group Theory. Contents

MATH APPLIED MATRIX THEORY

How To Prove The Dirichlet Unit Theorem

Recall that two vectors in are perpendicular or orthogonal provided that their dot

1 Determinants and the Solvability of Linear Systems

8 Square matrices continued: Determinants

MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.

Solutions to TOPICS IN ALGEBRA I.N. HERSTEIN. Part II: Group Theory

3. INNER PRODUCT SPACES

Unified Lecture # 4 Vectors

Math 312 Homework 1 Solutions

Vector Spaces 4.4 Spanning and Independence

4. FIRST STEPS IN THE THEORY 4.1. A

Notes on Algebraic Structures. Peter J. Cameron

BANACH AND HILBERT SPACE REVIEW

Eigenvalues and Eigenvectors

26. Determinants I. 1. Prehistory

Recall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.

18.06 Problem Set 4 Solution Due Wednesday, 11 March 2009 at 4 pm in Total: 175 points.

Transcription:

C. J. Papachristou MATHEMATICAL HANDBOOK

& I. sin A cos A 1 sin x cos x 1 tan x, cot x cos x sin x tan x sin ( A B) sin A cos B cos Asin B cos( A B) cos A cos B sin A sin B tan A tan B cot Acot B1 tan ( A B), cot ( A B) 1 tan Atan B cot B cot A sin A sin Acos A cos A cos A sin A cos A 1 1 sin tan A cot A1 tan A, cot A 1 tan A cot A A A B A B sin A sin B sin cos A B A B sin A sin B sin cos A B A B cos A cos B cos cos A B B A cos A cos B sin sin 1 sin Asin B [cos( A B) cos( A B)] 1 cos Acos B [cos( A B) cos( A B)] 1 sin Acos B [sin ( A B) sin ( A B)] sin ( A) sin A, cos( A) cos A tan ( A) tan A, cot ( A) cot A sin ( A) cos A, cos( A) sin A sin ( A) sin A, cos( A) cos A 1

sin cos tan cot 0 0 1 0 /6 = 30 1 3 3 3 /4 = 45 1 1 /3 = 60 3 1 3 3 3 / = 90 1 0 0 = 180 0 1 0 3 II. sin x sin xk x(k1) ( k 0, 1,, ) cos x cos xk xk ( k 0, 1,, ) tan x tan x k ( k 0, 1,, ) cot x cot x k ( k 0, 1,, ) sin xsin x k x ( k 1) ( k 0, 1,, ) x(k1) cos xcos x(k1) ( k 0, 1,, )

III. a b b c a c a b b a a b a b ab 1 1 0 a b a b a b c d a c b d 0 a b 0 c d a c b d a a a a a a 3 n n 0 1, 1, 1 a a a a a a 3 n n 1, 1, 1 n n n n 0 a b a b, a b IV.. :,, 3

V. a a, a 0 a, a 0 a 0 a a a a a a x x ( 0) x a 0 x a xa a b a b a b ab a b k k a a ( k Z) a b a b ( b 0) 4

VI. x 0 1 ( x 0) x x x x x 1 x x x x x ln(1) 0 e ln ln ( ), e ( ) ln( ) ln ln ln ln ln ln 1 ln ln k ln k ln ( k) 5

VII. x +1=0. x R.,, - i i = 1,, i 1,, x= ± i. x y, z = x+ i y : z = x+ i y (x, y) x= Re z, y= Im z -., z = 0 x = 0 y = 0., y = 0, o z. z = x+ i y, z x i y z., z. z = (x + y ) 1/ z z : z = 3+ i, z 3 i z z 13. :, z :, z = x+ i y, z, z,. z z z z Re z x, Im z y i 6

z 1 = x 1 + i y 1, z = x + i y. -, z 1 + z = (x 1 + x ) + i (y 1 + y ) z 1 - z = (x 1 - x ) + i (y 1 - y ) :, z 1 = z, x 1 = x y 1 = y., i = 1, z 1 z = (x 1 x y 1 y ) + i (x 1 y + x y 1 ), z 1 = z = x+ i y z = z x i y, : z z x y z z 1 / z (z 0), : z z z z z ( x i y )( x i y ) x x y y x y x y i z z z z x y x y x y 1 1 1 1 1 1 1 1 1, z = x+ i y, 1 z z x i y x y i z z z z x y x y x y : z z z z, z z z z 1 1 1 1 z z z z 1 1, z z z z 1 1 z z, z z z, z1 z z1 z n n z z, z z z z 1 1 : : z 1 = 3 i z = + i. z1 z, z / z. z1 z 1 7

y y z x i y r O x x z = x+ i y (x, y) (x, y)., -. x y,,. x = r cos, y = r sin, : r = z = ( x + y ) 1/ tan y x z = x+ i y = r (cos + i sin) z. z r (cos i sin ). z 1 = r 1 (cos 1 + i sin 1 ) z = r (cos + i sin )., z z r r [cos( ) i sin( )] 1 1 1 1 z z r [cos( 1) i sin( 1)] r 1 1, z = r (cos + i sin) : 1 1 1 1 z (cos i sin ) [cos( ) i sin( )] z r r :, z z 1 = 1. 8

e i cos isin (, ). i i ( ) e e i i cos ( ) sin ( ) cos sin, i i e e cos sin 1. : e 1/ e e i i i z = r (cos + i sin), r z, z i r e : 1 1 1 z e e, z r e z r r z r z1 z r1 r e, e z r 1 i i ( ) i i( 1 ) 1 1 i ( 1 ) i i 1 z1 r1 e, z r e.. : z i, z = r =, : z i cos i sin e e 4 4 i ( / 4) i / 4 9

z = r (cos + i sin) = i r e, r = z. n n in n z r e r (cos n i sin n ) ( n 0, 1,, ), z = cos + i sin = e i (r=1), de Moivre: (cos i sin ) n (cos n i sin n ), z 0, z 0 = 1 z n = 1/z n. z = r (cos + i sin), r = z, n- c c n = z. n : c z. n- n n k k ck r cos i sin, k 0,1,,,( n 1) n n : z = 1. 4, - c c 4 = 1. : z = 1 (cos 0 + i sin 0) (, r = 1, = 0), k k k k ck cos i sin cos i sin, k 0,1,,3 4 4 : c 0 = 1, c 1 = i, c = 1, c 3 = i : z = i. i, c c = i. : z = 1 [cos (/) + i sin(/)] (, r = 1, = /) ( / ) k ( / ) k ck cos i sin, k 0,1 c0 cos( / 4) i sin ( / 4) (1 i) c1 cos(5 / 4) i sin (5 / 4) (1 i) 10

ALGEBRA: SOME BASIC CONCEPTS Sets Subset: XY <=> ( xx => xy ) ; X = Y <=> XY and YX Proper subset: XY <=> XY and XY Union of sets: XY = { x / xx or xy } Intersection of sets: X Y = { x / xx and xy } Disjoint sets: X Y = Difference of sets: X Y = { x / xx and xy } Complement of a subset: X Y ; X \ Y = X Y Cartesian product: X Y = {(x, y) / xx and yy } Mapping: f : XY ; ( xx ) y = f (x)y Domain / range of f : D ( f ) = X, R ( f ) = f ( X ) = { f ( x) / xx } Y ; f is defined in X and has values in Y ; y= f ( x) is the image of x under f Composite mapping: f : XY, g : YZ ; f o g : XZ ; ( xx ) g( f (x)) Z Injective (1-1) mapping: Surjective (onto) mapping: Bijective mapping: f (x 1 ) = f (x ) <=> x 1 = x, or x 1 x <=> f (x 1 ) f (x ) f ( X ) = Y f is both injective and surjective => invertible Identity mapping: f id : XX ; f id (x) = x, xx Internal operation on X: X XX ; [(x, y) X X ] z X External operation on X: A XX ; [(a, x) A X ] y = a x X

Groups A group is a set G, together with an internal operation G GG ; (x, y) z = x y, such that: 1. The operation is associative: x (y z) = (x y) z. eg (identity) : x e = e x = x, xg 3. xg, x 1 G (inverse): x 1 x = x x 1 = e A group G is abelian or commutative if x y = y x, x, yg. A subset SG is a subgroup of G if S is itself a group (clearly, then, S contains the identity e of G, as well as the inverse of every element of S ). Vector space over R Let V={x, y, z,...}, and let a, b, c,... R. Consider an internal operation + and an external operation on V : + : V VV ; x+y = z : R VV ; ax = y Then, V is a vector space over R iff 1. V is a commutative group with respect to +. The identity element is denoted 0, while the inverse of x is denoted x.. The operation satisfies the following: a(bx)= (ab)x (a+b)x = ax + bx a(x+y) = ax + ay 1x = x, 0x = 0 A set {x 1, x,..., x k } of elements of V is linearly independent iff the equation 1 c 1 x 1 + c x +... + c k x k = 0 can only be satisfied for c 1 = c =... = c k = 0 ; otherwise, the set is linearly dependent. The dimension dimv of V is the largest number of vectors in V that constitute a linearly independent set. If dimv=n, then any system {e 1, e,..., e n } of n linearly independent elements is a basis for V, and any xv can be uniquely expressed as x = c 1 e 1 + c e +... + c n e n. A subset SV is a subspace of V if S is itself a vector space under the operations (+) and (). In particular, the sum x+y of any two elements of S, as well as the scalar multiple ax and the inverse x of any element x of S, must belong to S. Clearly, this set must contain the identity 0 of V. If S is a subspace of V, then dim SdimV. In particular, S coincides with V iff dim S = dimv. 1 The symbol () will often be omitted in the sequel.

Functionals A functional on a vector space V is a mapping : V R ; (xv) t = (x)r. The functional is linear if (ax+by)=a(x)+b(y). The collection of all linear functionals on V is called the dual space of V, denoted V*. It is itself a vector space over R, and dimv*= dimv. Algebras A real algebra A is a vector space over R equipped with a binary operation ( ) : A AA ; (x y) = z, such that, for a, br, (ax + by z) = a(x z) + b(y z) (x ay + bz) = a(x y) + b(x z) An algebra is commutative if, for any two elements x, y, (x y) = (y x) ; it is associative if, for any x, y, z, (x (y z)) = ((x y) z). Example: The set 0 (R n ) of all functions on R n is a commutative, associative algebra. The multiplication operation ( ) : 0 (R n ) 0 (R n ) 0 (R n ) is defined by ( f g )(x 1,..., x n ) = f (x 1,..., x n ) g(x 1,..., x n ). Example: The set of all nn matrices is an associative, non-commutative algebra. The binary operation ( ) is matrix multiplication. A subspace S of A is a subalgebra of A if S is itself an algebra under the same binary operation ( ). In particular, S must be closed under this operation; i.e., (x y)s for any x, y in S. We write: SA. A subalgebra SA is an ideal of A iff (x y)s and (y x)s, for any xs, ya. Modules Note first that R is an associative, commutative algebra under the usual operations of addition and multiplication. Thus, a vector space over R is a vector space over an associative, commutative algebra. More generally, a module M over A is a vector space over an associative but (generally) non-commutative algebra. In particular, the external operation () on M is defined by : A MM ; ax = y (aa ; x, y M ). Example: The collection of all n-dimensional column matrices, with A taken to be the algebra of nn matrices, and with matrix multiplication as the external operation. 3

Vector fields A vector field V on R n is a map from a domain of R n into R n : V : R n UR n ; [x (x 1,..., x n )U ] V(x) (V 1 (x k ),...,V n (x k ))R n. The vector x represents a point in U, with coordinates (x 1,..., x n ). The functions V i (x k ) (i=1,...,n) are the components of V in the coordinate system (x k ). Given two vector fields U and V, we can construct a new vector field W=U+V such that W(x)=U(x)+V(x). The components of W are the sums of the respective components of U and V. Given a vector field V and a constant ar, we can construct a new vector field Z=aV such that Z(x)= av(x). The components of Z are scalar multiples (by a) of those of V. It follows from the above that the collection of all vector fields on R n is a vector space over R. More generally, given a vector field V and a function f 0 (R n ), we can construct a new vector field Z= f V such that Z(x)= f (x)v(x). Given that 0 (R n ) is an associative algebra, we conclude that the collection of all vector fields on R n is a module over 0 (R n ) (in this particular case, the algebra 0 (R n ) is commutative). A note on linear independence: Let {V 1,...,V r }{V a } be a collection of vector fields on R n. (a) The set {V a } is linearly dependent over R (linearly dependent with constant coefficients) iff there exist real constants c 1,...,c r, not all zero, such that c 1 V 1 (x) + + c r V r (x) = 0, xr n. If the above relation is satisfied only for c 1 = = c r = 0, the set {V a } is linearly independent over R. (b) The set {V a } is linearly dependent over 0 (R n ) iff there exist functions f 1 (x k ),..., f r (x k ), not all identically zero over R n, such that f 1 (x k ) V 1 (x) + + f r (x k ) V r (x) = 0, x(x k )R n. If this relation is satisfied only for f 1 (x k )== f r (x k ) 0, the set {V a } is linearly independent over 0 (R n ). There can be at most n elements in a linearly independent system over 0 (R n ). These elements form a basis {e 1,..., e n }{e k } for the module of all vector fields on R n. An element of this module, i.e. an arbitrary vector field V, is written as a linear combination of the {e k } with coefficients V k 0 (R n ). Thus, at any point x(x k )R n, 4

V(x) = V 1 (x k ) e 1 +...+ V n (x k ) e n (V 1 (x k ),...,V n (x k )). In particular, in the basis {e k }, e 1 (1,0,0,...,0), e (0,1,0,...,0),, e n (0,0,...,0,1). Example: Let n=3, i.e., R n =R 3. Call {e 1, e, e 3 } {i, j, k}. Let V be a vector field on R 3. Then, at any point x (x, y, z)r 3, V(x) = V x (x, y, z) i + V y (x, y, z) j + V z (x, y, z) k (V x, V y, V z ). Now, consider the six vector fields V 1 = i, V = j, V 3 = k, V 4 = x j y i, V 5 = y k z j, V 6 = z i x k. Clearly, the {V 1, V, V 3 } are linearly independent over 0 (R 3 ), since they constitute the basis {i, j, k}. On the other hand, the V 4, V 5, V 6 are separately linearly dependent on the {V 1, V, V 3 } over 0 (R 3 ). Moreover, the set {V 4, V 5, V 6 } is also linearly dependent over 0 (R 3 ), since zv 4 + xv 5 + yv 6 = 0. Thus, the set {V 1,, V 6 } is linearly dependent over 0 (R 3 ). On the other hand, the system {V 1,, V 6 } is linearly independent over R, since the equation c 1 V 1 + + c 6 V 6 = 0, with c 1,...,c 6 R (constant coefficients), can only be satisfied for c 1 =...= c 6 = 0. In general, there is an infinite number of linearly independent vector fields on R n over R, but only n linearly independent fields over 0 (R n ). Derivation on an algebra Let L be an operation on an algebra A (an operator on A): L : AA ; (xa) y = L x A. L is a derivation on A iff, x, y A and a, br, L (ax+by) = al(x) + bl(y) L (x y) = (L x y) + (x L y) (linearity) (Leibniz rule) Example: Let A= 0 (R n )={ f (x 1,, x n )}, and let L be the linear operator L = 1 (x k ) /x 1 + + n (x k ) /x n i (x k ) /x i, where the i (x k ) are given functions. As can be shown, L [ f (x k ) g(x k )] = [L f (x k )] g(x k ) + f (x k ) L g(x k ). Hence, L is a derivation on 0 (R n ). 5

Lie algebra An algebra over R is a (real) Lie algebra with binary operation [, ]: (Lie bracket) iff this operation satisfies the properties: [ax + by, z] = a [x, z] + b [ y, z] [x, y] = [ y, x] (antisymmetry) [x, [ y, z]] + [ y, [z, x]] + [z, [x, y]] = 0 (Jacobi identity) (where x, y, z and a, br). Note that, by the antisymmetry of the Lie bracket, the first and third properties are written, alternatively, [x, ay + bz] = a [x, y] + b [x, z], [[x, y], z] + [[ y, z], x] + [[z, x], y] = 0. A Lie algebra is a non-associative algebra, since, as follows by the above properties, [x, [ y, z]] [[x, y], z]. Example: The algebra of nn matrices, with [A, B]=ABBA (commutator). Example: The algebra of all vectors in R 3, with [a, b] = ab (vector product). Lie algebra of derivations Consider the algebra A= 0 (R n )={ f (x 1,, x n )}. Consider also the set D(A) of linear operators on A, of the form L = i (x k ) /x i (sum on i = 1,,, n). These first-order differential operators are derivations on A (the Leibniz rule is satisfied). Now, given two such operators L 1, L, we construct the linear operator (Lie bracket of L 1 and L ), as follows: [L 1, L ] = L 1 L L L 1 ; [L 1, L ] f (x k ) = L 1 (L f (x k )) L (L 1 f (x k )). It can be shown that [L 1, L ] is a first-order differential operator (a derivation), hence is a member of D(A). (This is not the case with second-order operators like L 1 L!) Moreover, the Lie bracket of operators satisfies all the properties of the Lie bracket of a general Lie algebra (such as antisymmetry and Jacobi identity). It follows that the set D(A) of derivations on 0 (R n ) is a Lie algebra, with binary operation defined as the Lie bracket of operators. 6

Direct sum of subspaces Let V be a vector space over a field K (where K may be R or C), of dimension dimv=n. Let S 1, S be disjoint (i.e., S 1 S ={0}) subspaces of V. We say that V is the direct sum of S 1 and S if each vector of V can be uniquely represented as the sum of a vector of S 1 and a vector of S. We write: V= S 1 S. In terms of dimensions, dimv=dim S 1 +dim S. We similarly define the vector sum of three subspaces of V, each of which is disjoint from the direct sum of the other two (i.e., S 1 (S S 3 )={0}, etc.). Homomorphism of vector spaces Let V, W be vector spaces over a field K. A mapping :VW is said to be a linear mapping or homomorphism if it preserves linear operations, i.e., (x+y) = (x) + ( y), (kx) = k(x), x, yv and kk. A homomorphism which is also bijective (1-1) is called an isomorphism. The set of vectors xv mapping under into the zero of W is called the kernel of the homomorphism : Ker = { xv : (x) = 0}. Note that (0)=0, for any homomorphism (clearly, the two zeros refer to different vector spaces). Thus, in general, 0Ker. If Ker={0}, then the homomorphism is also an isomorphism of V onto a subspace of W. If, moreover, dimv=dimw, then the map :VW is itself an isomorphism. In this case, Im=W, where, in general, Im (image of the homomorphism) is the collection of images of all vectors of V under the map. The algebra of linear operators Let V be a vector space over a field K. A linear operator A on V is a homomorphism A : VV. Thus, A(x+y) = A(x) + A(y), A(kx) = k A(x), x, yv and kk. The sum A+B and the scalar multiplication ka (kk) are linear operators defined by (A+B) x = A x + B x, (ka) x = k (A x). Under these operations, the set Op(V) of all linear operators on V is a vector space. The zero element of that space is a zero operator 0 such that 0x=0, xv. 7

Since A and B are mappings, their composition may be defined. This is regarded as their product AB: (AB) xa(bx), xv. Note that AB is a linear operator on V, hence belongs to Op(V). In general, operator products are non-commutative: ABBA. However, they are associative and distributive over addition: (AB)C = A (BC) ABC, A (B+C) = AB+AC. The identity operator E is the mapping of Op(V) which leaves every element of V fixed: E x = x. Thus, AE=EA=A. Operators of the form ke (kk), called scalar operators, are commutative with all operators. In fact, any operator commutative with every operator of Op(V) is a scalar operator. It follows from the above that the set Op(V) of all linear operators on a given vector space V is an algebra. This algebra is associative but (generally) non-commutative. An operator A is said to be invertible if it represents a bijective (1-1) mapping, i.e., if it is an isomorphism of V onto itself. In this case, an inverse operator A 1 exists such that AA 1 = A 1 A=E. Practically this means that, if A maps xv onto yv, then A 1 maps y back onto x. For an invertible operator A, Ker A={0} and Im A=V. Matrix representation of a linear operator Let A be a linear operator on V. Let {e i } ( i=1,..., n) be a basis of V. Let A e k = e i A i k (sum on i) where the A ik are real or complex, depending on whether V is a vector space over R or C. The n n matrix A=[A i k] is called the matrix of the operator A in the basis {e i }. Now, let x=x i e i (sum on i) be a vector in V, and let y=a x. If y= y i e i, then, by the linearity of A, y i = A i k x k (sun on k). In matrix form, [ y ] n 1 = [ A ] n n [ x ] n 1. Next, let A, B be linear operators on V. Define their product C=AB by C x = (AB) xa (Bx), xv. Then, for any basis {e i }, C e k = A (B e k ) = e i A i j B j k e i C i k => 8

C i k = A i j B j k or, in matrix form, That is, C = A B. the matrix of the product of two operators is the product of the matrices of these operators, in any basis of V. Consider now a change of basis defined by the transition matrix T= [T i k]: e k = e i T i k. The inverse transformation is e k = e i (T 1 ) i k. Under this basis change, the matrix A of an operator A transforms as A = T 1 A T (similarity transformation). Under basis transformations, the trace and the determinant of A remain unchanged: tra = tra, deta = deta. An operator A is said to be nonsingular (singular) if deta0 (deta=0). Note that this is a basis-independent property. Any nonsingular operator is invertible, i.e., there exists an inverse operator A 1 Op(V) such that A A 1 = A 1 A=E. Since an invertible operator represents a bijective mapping (i.e., both 1-1 and onto), it follows that KerA={0} and ImA=V. If A is invertible (nonsingular), then, for any basis {e i } (i=1,..., n) of V, the vectors {Ae i } are linearly independent and hence also constitute a basis. Invariant subspaces and eigenvectors Let V be an n-dimensional vector space over a field K, and let A be a linear operator on V. The subspace S of V is said to be invariant under A if, for every vector x of S, the vector Ax again belongs to S. Symbolically, ASS. A vector x0 is said to be an eigenvector of A if it generates a one-dimensional invariant subspace of V under A. This means that an element K exists, such that A x = x. The element is called an eigenvalue of A, to which eigenvalue the eigenvector x belongs. Note that, trivially, the null vector 0 is an eigenvector of A, belonging to any 9

eigenvalue. The set of all eigenvectors of A, belonging to a given, is a subspace of V called the proper subspace belonging to. It can be shown that the eigenvalues of A are basis-independent quantities. Indeed, let A=[A ik ] be the (n n) matrix representation of A in some basis {e i } of V, and let x=x i e i be an eigenvector belonging to. We denote by X=[x i ] the column vector representing x in that basis. The eigenvalue equation for A is written, in matrix form, This is written A ik x k = x i or A X = X. (A 1 n ) X = 0. This equation constitutes a linear homogeneous system for X=[x i ], which has a nontrivial solution iff det (A 1 n ) = 0. This polynomial equation determines the eigenvalues i (i=1,...,n) (not necessarily all different from each-other) of A. Since the determinant of the matrix representation of an operator [in particular, of the operator (A E) for any given ] is a basisindependent quantity, it follows that, if the above equation is satisfied for a certain in a certain basis (where A is represented by the matrix A), it will also be satisfied for the same in any other basis (where A is represented by another matrix, say, A ). We conclude that the eigenvalues of an operator are a property of the operator itself and do not depend on the choice of basis of V. If we can find n linearly independent eigenvectors {x i } of A, belonging to the corresponding eigenvalues i, we can use these vectors to define a basis for V. In this basis, the matrix representation of A has a particularly simple diagonal form: A = diag ( 1,..., n ). Using this expression, and the fact that the quantities tra, deta and i are invariant under basis transformations, we conclude that, in any basis of V, tra = 1 + +...+ n, deta = 1... n. We note, in particular, that all eigenvalues of an invertible (nonsingular) operator are nonzero. Indeed, if even one is zero, then deta=0 and A is singular. An operator A is called nilpotent if A m =0 for some natural number m>1. The smallest such value of m is called the degree of nilpotency, and it cannot exceed n. All eigenvalues of a nilpotent operator are zero. Thus, such an operator is singular (noninvertible). An operator A is called idempotent (or projection operator) if A =A. It follows that A m =A, for any natural number m. The eigenvalues of an idempotent operator can take the values 0 or 1. 10

BASIC MATRIX PROPERTIES T T T T T T ( A B) A B ; ( AB) B A T * * T ( AB) A B ; ( AB) B A where M ( M ) ( M ) T T * ( ka) ka ; ( ka) k A ( kc) 1 1 1 T 1 1 T 1 1 ( AB) B A ; ( A ) ( A ) ; ( A ) ( A ) T T T [ A, B] [ B, A ] ; [ A, B] [ B, A ] 1 1 A adja (det A 0) det A 1 a b 1 d b c d ad bc c a tr( AB) tra trb T * tra tra ; tra ( tra) tr( AB) tr( BA), tr( ABC) tr( BCA) tr( CAB), etc. tr[ A, B] 0 T 1 * det A det A ; det A (det A) det( AB) det( BA) det A det B det( A ) 1/ det A n det( ca) c det A cc, A gl( n, C) If any row or column of A is multiplied by c, then so is det A. [ A, B] [ B, A] AB BA [ A, B C] [ A, B] [ A, C] ; [ A B, C] [ A, C] [ B, C] [ A, BC] [ A, B] C B[ A, C] ; [ AB, C] A[ B, C] [ A, C] B [ A, [ B, C]] [ B, [ C, A]] [ C, [ A, B]] 0 [[ A, B], C] [[ B, C], A] [[ C, A], B] 0 Let A A( t) [ a ( t)], B B( t) [ b ( t)], be (n n) matrices. The derivative i j of A (similarly, of B) is the (n n) matrix da/dt, with elements da d a i j ( t ). dt i j dt The integral of A (similarly, of B) is the (n n) matrix defined by A( t) dt ai j ( t) dt. i j i j 11

d da db d da db ( A B) ; ( AB) B A dt dt dt dt dt dt d da db [ A, B], B A, dt dt dt d da ( A ) A A dt dt d ( A ) A ( da) A da d tr ( tra) dt dt 1 1 1 1 1 1 Let A A( x, y). Call A / x x A Ax, etc.: 1 1 1 1 x A Ay y A Ax A Ax A Ay ( ) ( ) [, ] 0 1 1 1 1 x Ay A y Ax A Ax A Ay A ( ) ( ) [, ] 0 1 1 1 1 1 1 x y y x y x x y A( A A ) A ( A A ) A ( A A ) A ( A A ) A A A e exp A 1 A n! A B e B 1 e n0 1 * T T ; ; ; * 1 A A A A A A A A e e e e e when [ A, B] 0 In general, BAB A B B A AB n e e e e e e e e A B C e e e where 1 1 C A B [ A, B] [ A, [ A, B]] [ B, [ B, A]] 1 By definition, log B A B e. A tra tr (log B) det e e det B e tr (log B) log(det B) det (1 A) 1 tr A, for infinitesimal A A 1 1 x x x x x tr ( A A ) tr ( A A ) tr (log A) [ tr (log A)] [log(det A)] 1