Explicit inverses of some tridiagonal matrices
|
|
- Grant Flowers
- 7 years ago
- Views:
Transcription
1 Linear Algebra and its Applications 325 (2001) 7 21 wwwelseviercom/locate/laa Explicit inverses of some tridiagonal matrices CM da Fonseca, J Petronilho Depto de Matematica, Faculdade de Ciencias e Technologia, Univ Coimbra, Apartado 3008, 3000 Coimbra, Portugal Received 28 April 1999; accepted 26 June 2000 Submitted by G Heinig Abstract We give explicit inverses of tridiagonal 2-Toeplitz and 3-Toeplitz matrices which generalize some well-known results concerning the inverse of a tridiagonal Toeplitz matrix 2001 Elsevier Science Inc All rights reserved AMS classification: 33C45; 42C05 Keywords: Invertibility of matrices; r-toeplitz matrices; Orthogonal polynomials 1 Introduction In recent years the invertibility of nonsingular tridiagonal or block tridiagonal matrices has been quite investigated in different fields of applied linear algebra (for historical notes see [8]) Several numerical methods, more or less efficient, have risen in order to give expressions of the entries of the inverse of this kind of matrices Though, explicit inverses are known only in a few cases, in particular when the tridiagonal matrix is symmetric with constant diagonals and subject to some restrictions (cf [3,8,10]) Furthermore, Lewis [5] gave a different way to compute other explicit inverses of nonsymmetric tridiagonals matrices In [1], Gover defines a tridiagonal 2-Toeplitz matrix of order n, as a matrix A = (a ij ),a ij = 0if i j > 1anda ij = a kl if (i, j) (k, l) mod 2, ie, Corresponding author addresses: cmf@matucpt (CM da Fonseca), josep@matucpt (J Petronilho) /01/$ - see front matter 2001 Elsevier Science Inc All rights reserved PII:S (00)
2 8 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 a 1 b 1 c 1 a 2 b 2 c 2 a 1 b 1 A = c 1 a 2 b 2 (1) c 2 a 1 n n Similarly, a tridiagonal 3-Toeplitz matrix (n n) is of the form a 1 b 1 c 1 a 2 b 2 c 2 a 3 b 3 c 3 a 1 b 1 B = c 1 a 2 b 2 (2) c 2 a 3 b 3 c 3 a 1 n n Making use of the theory of orthogonal polynomials, we will give the explicit inverse of tridiagonal 2-Toeplitz and 3-Toeplitz matrices, based on recent results from [1,6,7] As a corollary, we will obtain the explicit inverse of a tridiagonal Toeplitz matrix, ie, a tridiagonal matrix with constant diagonals (not necessarily symmetric) Different proofs of the results by Kamps [3,4] involving the sum of all the entries of the inverse can be simplified Throughout this paper, it is assumed that all the matrices are invertible 2 Inverse of a tridiagonal matrix Let T be an n n real nonsingular tridiagonal matrix a 1 b 1 0 c 1 a 2 b 2 T = c 2 (3) bn 1 0 c n 1 a n Denote T 1 = (α ij ) In [5] Lewis proved the following result Lemma 21 Let T be the matrix (3) and assume that b l /= 0 for l = 1,,n 1 Then α ji = γ ij α ij when i<j,
3 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) γ ij = j 1 l=i c l b l Of course this reduces to α ij = α ji, when T is symmetric Using this lemma, one can prove that there exist two finite sequences {u i } and {v i } (i = 1,,n 1) such that u 1 v 1 u 1 v 2 u 1 v 3 u 1 v n γ 12 u 1 v 2 u 2 v 2 u 2 v 3 u 2 v n T 1 = γ 13 u 1 v 3 γ 23 u 2 v 3 u 3 v 3 u 3 v n (4) γ 1,n u 1 v n γ 2,n u 2 v n γ 3,n u 3 v n u n v n Since {u i } and {v i } are only defined up to a multiplicative constant, we make u 1 = 1 The following result shows how to compute {u i } and {v i } The determination of these sequences is very similar to the particular case studied in [8], the reader may check for details Lemma 22 (i) The {v i } (i = 1,,n 1) can be computed from v 1 = 1 d 1, v k = b k 1 d k v k 1, k = 2,,n, d n = a n, d i = a i b ic i, i = n 1,,1 d i+1 (ii) The {u i } (i = 1,,n 1) can be computed from u n = 1, δ n v n δ 1 = a 1, u k = b k δ k u k+1, k = n 1,,1, δ i = a i b i 1c i 1 δ i 1, i = 2,,n 1 Notice that for centro-symmetric matrices, ie, a ij = a n+1 i,n+1 j, we have v i = u n+1 i The following proposition is known in the literature We give a proof based on the lecture of [8] Theorem 21 Let T be the n n tridiagonal matrix defined in (3) and assume that b l /= 0 for l = 1,,n 1Then
4 10 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 ( 1) i+j d j+1 d n b i b j 1 if i j, δ i δ n (T 1 ) ij = ( 1) i+j d i+1 d n c j c i 1 if i>j δ j δ n (with the convention that empty product equals 1) Proof Let us assume that i jthen But (T 1 ) ij = u i v j = ( 1) n i b i b n 1 δ i δ n v n ( 1) j 1 b 1 b j 1 d 1 d j v n = ( 1) n 1 b 1 b n 1 d 1 d n Therefore (T 1 ) ij =( 1) n i b i b n 1 δ i δ n ( 1) n 1 d 1 d n b 1 b n 1 ( 1) j 1 b 1 b j 1 d 1 d j =( 1) i+j b i b j 1 d j+1 d n δ i δ n If we set δ i = θ i, θ i 1 θ 0 = 1, θ 1 = a 1, (5) we get the recurrence relation θ i = a i θ i 1 b i 1 c i 1 θ i 2 ; and if we now put d i = φ i, φ n+1 = 1, φ n = a n, (6) φ i+1 we get the recurrence relation φ i = a i φ i+1 b i c i φ i+2 As a consequence, we will achieve ( 1) i+j b i b j 1 θ i 1 φ j+1 /θ n if i j, (T 1 ) ij = (7) ( 1) i+j c j c i 1 θ j 1 φ i+1 /θ n if i>j These are the general relations given by Usmani [9], which we will use throghout this paper Notice that θ n = δ 1 δ n = det T
5 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) Chebyshev polynomials of the second kind In the next it is useful to consider the set of polynomials {U n } n 0, such that each U n is of degree exactly n, satisfying the recurrence relation U n+1 (x) = 2xU n (x) U n 1 (x), n 1 with initial conditions U 0 = 1 and U 1 = 2x These polynomials are called Chebyshev polynomials of the second kind and they have the form sin(n + 1)θ U n (x) =, cos θ = x, sin θ when x < 1 When x > 1, they have the form sinh(n + 1)θ U n (x) =, cosh θ = x sinh θ In particular U n (±1) = (±1) n (n + 1) In any case, if x /= 1, then U n (x) = rn+1 + rn+1, r + r r ± = x ± x 2 1 are the two solutions of the quadratic equation r 2 2xr + 1 = 0 The following proposition is a slight extension of Lemma 25 of Meurant [8] Lemma 31 Fix a positive integer number n and let a and b be nonzero real numbers such that U i (a/2b) /= 0 for all i = 1, 2,,n Consider the recurrence relation α 1 = a, α i = a b2, i = 2,,n α i 1 Under these conditions, if a =±2b, then α i = Otherwise ±b(i + 1) i α i = b ri+1 r+ i ri + ri+1,
6 12 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 r ± = a ± a 2 4b 2 2b are the two solutions of the quadratic equation r 2 (a/b)r + 1 = 0 Proof Let us assume that a/=±2b (otherwise it is trivial) and set α i = β i /β i 1 Then β i aβ i 1 + b 2 β i 2 = 0, β 0 = 1, β 1 = a, hence β i = b i U i (a/2b) and the conclusion follows 4 Inverse of a tridiagonal 2-Toeplitz matrix In this section we will determine the explicit inverse of a matrix of type (1) By Theorem 21, this is equivalent to determine δ i and d i, ie, by (7) and transformations (5) and (6), θ i and φ i So, to evaluate θ i,wehave θ 0 = 1, θ 1 = a 1, θ 2i = a 2 θ 2i 1 b 1 c 1 θ 2i 2, θ 2i+1 = a 1 θ 2i b 2 c 2 θ 2i 1 Making the change z i = ( 1) i θ i, we get the relations z 0 = 1, z 1 = a 1, z 2i = a 2 z 2i 1 b 1 c 1 z 2i 2, z 2i+1 = a 1 z 2i b 2 c 2 z 2i 1 According to the results in [6], z i = Q i (0), Q i is a polynomial of degree exactly i, defined by the recurrence relation Q i+1 (x) = (x β i )Q i (x) γ i Q i 1 (x), (8) { β 2j = a 1, β 2j+1 = a 2, and { γ2j = b 2 c 2, γ 2j+1 = b 1 c 1, (9) and initial conditions Q 0 (x) = 1andQ 1 (x) = x β 0
7 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) The solution of the recurrence relation (8) with coefficients (9) is Q 2i (x) = P i [π 2 (x)], Q 2i+1 (x) = (x a 1 )Pi [π 2 (x)], (10) π 2 (x) = (x a 1 )(x a 2 ), (11) ( ) i x Pi (x) = b1 c 1 b 2 c 2 b1 b 2 c 1 c 2 Ui 2, (12) b 1 b 2 c 1 c 2 and ( ) [ i x b1 c 1 b 2 c 2 P i (x)= b1 b 2 c 1 c 2 U i 2 b 1 b 2 c 1 c ] 2 x b1 c 1 b 2 c 2 + βu i 1 2, (13) b 1 b 2 c 1 c 2 with β 2 = b 2 c 2 /b 1 c 1 We may conclude that θ i = ( 1) i Q i (0) Now, to evaluate φ i we put ψ i = φ n+1 i, for i = 1,,n, and we have to distinguish the cases when the order n of the matrix (1) is odd and when it is even First let us suppose that n is odd Then we have ψ 0 = 1, ψ 1 = a 1, ψ 2i = a 2 ψ 2i 1 b 2 c 2 ψ 2i 2, ψ 2i+1 = a 1 ψ 2i b 1 c 1 ψ 2i 1 Following the same steps as in the previous case, we obtain ψ i = ( 1) i Q i (0), ie, φ i = ( 1) i Q n+1 i (0), Q i (x) is the same as (10) but with β 2 = b 1 c 1 /b 2 c 2 in (13) Notice that if b 1 c 1 = b 2 c 2,thenθ i = φ n+1 i If n is even, we get ψ 0 = 1, ψ 1 = a 2, ψ 2i = a 1 ψ 2i 1 b 1 c 1 ψ 2i 2, ψ 2i+1 = a 2 ψ 2i b 2 c 2 ψ 2i 1 Therefore ψ i = ( 1) i Q i (0),
8 14 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 ie, φ i = ( 1) i+1 Q n+1 i (0), Q 2i (x) = P i [π 2 (x)], Q 2i+1 (x) = (x a 2 )Pi [π 2 (x)], with π 2 (x), P i (x) and Pi (x) the same as in (11), (13) and (12), respectively Observe that if a 1 = a 2, then θ i = φ n+1 i We have determined completely the θ i s and φ i s of (7 ) thus the inverse of the matrix in the case of a tridiagonal 2-Toeplitz matrix (1) Theorem 41 Let A be the tridiagonal matrix (1), with a 1 a 2 /= 0 and b 1 b 2 c 1 c 2 > 0 Put π 2 (x):=(x a 1 )(x a 2 ), β := b 2 c 2 /b 1 c 1 and let {Q i ( ; α, γ )} be the sequence of polynomials defined by ( ) [ i π2 (x) b 1 c 1 b 2 c 2 Q 2i (x; α, γ )= b1 b 2 c 1 c 2 U i 2 b 1 b 2 c 1 c ] 2 π2 (x) b 1 c 1 b 2 c 2 +γu i 1 2 b 1 b 2 c 1 c 2 ( ) i π2 (x) b 1 c 1 b 2 c 2 Q 2i+1 (x; α, γ ) = (x α) b1 b 2 c 1 c 2 Ui 2, b 1 b 2 c 1 c 2 α and γ are some parameters Under these conditions, ( 1) i+j b (A 1 p (j i)/2 i bq (j i+1)/2 i θ i 1 φ j+1 /θ n if i j, ) ij = ( 1) i+j c (j i)/2 cq (j i+1)/2 j θ j 1 φ i+1 /θ n if i>j, p j (14) p l = (3 ( 1) l )/2, q l = (3 + ( 1) l )/2, z denotes the greater integer less or equal to the real number z, θ i = ( 1) i Q i (0; a 1,β), (15) and ( 1) i Q n+1 i (0; a 1, 1/β) φ i = ( 1) i+1 Q n+1 i (0; a 2,β) if n is odd, if n is even Remark As we already noticed, under the conditions of Theorem 41, if n is odd and b 1 c 1 = b 2 c 2, then θ i = φ n+1 i ;andifn is even and a 1 = a 2, then also θ i = φ n+1 i
9 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) As a first application, let us consider a classical problem on the inverse of a tridiagonal Toeplitz matrix Suppose that a 1 = a 2 = a, b 1 = b 2 = b and c 1 = c 2 = c in (1), with a/= 0andbc > 0 Therefore, we have the n n tridiagonal Toeplitz matrix a b 0 c a b T = c a (16) b 0 c a According to (14), ( 1) i+j b j i θ i 1 φ j+1 /θ n if i j, (T 1 ) ij = ( 1) i+j c i j θ j 1 φ i+1 /θ n if i>j It is well-known and we can easily check using elementary trigonometry that ) U 2i+1 (x) = 2xU i (2x 2 1 Since β = 1and ( a 2 ) ( 2bc U i = U i 2bc from (15) we get ( ) ( i a θ i = bc Ui a ) = bc bc a a U 2i+1 2, bc ) 2, bc and taking into account the above remark the φ i s can be computed using the θ i s, giving ( ) n+1 i a φ i = θ n+1 i = bc Un+1 i 2 bc Therefore, the next result comes immediately Corollary 41 Let T be the matrix in (16) and define d :=a/2 bc The inverse is given by ( 1) i+j b j i U i 1 (d)u n j (d) ( bc ) j i+1 if i j, U n (d) (T 1 ) ij = ( 1) i+j c i j U j 1 (d)u n i (d) ( bc ) i j+1 if i>j U n (d)
10 16 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 This result is well-known It can be deduced, eg, using results in the book of Heinig and Rost [2, p 28] The next corollary is Theorem 31 of Kamps [3] Corollary 42 Let Σ be the n-square matrix a b 0 b a b Σ = b a (17) b 0 b a The inverse is given by ( 1) i+j 1 b (Σ 1 ) ij = ( 1) i+j 1 b U i 1 (a/2b)u n j (a/2b) U n (a/2b) U j 1 (a/2b)u n i (a/2b) U n (a/2b) if i j, if i>j We consider now a second application In [3,4], the matrices of type (17), with a>0, b/= 0anda>2 b, arose as the covariance matrix of one-dependent random variables Y 1,,Y n, with same expectation Let us consider the least squares estimator ˆµ opt = 1t Σ 1 Y 1 t Σ 1 1, 1 = (1,,1) and Y = (Y 1,,Y n ) t, which estimates the parameter µ equal to the common expectation of the Y i s, with variance V 1 ˆµ opt = 1 t Σ 1 1 According to Kamps, the estimator ˆµ opt is the best unbiased estimator based on Y So, the sum of all the entries of the inverse of Σ, 1 t Σ 1 1, has an important role in the determination of this estimator and therefore in the computation of the variance V ( ˆµ opt ) Kamps used in [3] some sum and product formulas involving different kinds of Chebyshev polynomials We will prove the same results in a more concise way Corollary 43 The sum s i of the ith row (or column) of Σ 1, for i = 1,,n is given by s i = 1 + b(σ 1i + σ 1,n i+1 ), a + 2b σ ij = (Σ 1 ) ij, when i j Proof By Corollary 42, for i j,
11 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) ( σ ij = ( 1) i+j 1 r i + r ) i r n j+1 + r n j+1 ( ), b (r + r ) r+ n+1 rn+1 r ± = a ± a 2 4b 2 2b Thus i 1 n s i = σ ji + σ ij j=1 j=1 j=i By the sum of a geometric progression and the fact that r + r = 1, we have i 1 ( r+ i+1 )( and σ ji = b (σ 1i + σ ii ) a + 2b + 1 a + 2b n σ ij = b σ 1,n i+1 σ ii a + 2b j=i 1 a + 2b (r + r ) ( r i + r i ) ( r n i ) r+ n i+1 r n i+1 ( ) r+ n+1 rn+1 + rn i ( (r + r ) r+ n+1 rn+1 ) ) Finally ( r+ i+1 ri+1 = (r + r ) r+ n i+1 r n i+1 ( r+ n+1 rn+1 ) r+ i ri r+ n i rn i ) The previous theorem says that s i depends only on the first row of Σ 1 In fact (as the following result says) the sum of all entries of Σ 1 depends only on σ 11 and σ 1n Corollary 44 1 t Σ 1 1 = n + 2bs 1 a + 2b 5 Inverse of a tridiagonal 3-Toeplitz matrix In an analogous way of the 2-Toeplitz matrices, we may ask about the inverse of a tridiagonal 3-Toeplitz matrix To solve this case, we have to compute θ i and φ i in (7)
12 18 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 For θ i,wehave θ 0 = 1, θ 1 = a 1, θ 3i = a 3 θ 3i 1 b 2 c 2 θ 3i 2, θ 3i+1 = a 1 θ 3i b 3 c 3 θ 3i 1, θ 3i+2 = a 2 θ 3i+1 b 1 c 1 θ 3i Making the change z i = ( 1) i θ i, we get the relations z 0 = 1, z 1 = a 1, z 3i = a 3 z 3i 1 b 2 c 2 z 3i 2, z 3i+1 = a 1 z 3i b 3 c 3 z 3i 1, z 3i+2 = a 2 z 3i+1 b 1 c 1 z 3i According to the results in [7], z i = Q i (0), Q i is a polynomial of degree exactly i, defined according to Q i+1 (x) = (x β i )Q i (x) γ i Q i 1 (x), β 3j = a 1, γ 3j = b 3 c 3, β 3j+1 = a 2, and γ 3j+1 = b 1 c 1, β 3j+2 = a 3, γ 3j+2 = b 2 c 2 The solution of the recurrence relation (8) with coefficients (18) is (18) Q 3i (x) = P i [π 3 (x)] + b 3 c 3 (x a 2 ) P i 1 [π 3 (x)], Q 3i+1 (x) = (x a 1 )P i [π 3 (x)] + b 1 c 1 b 3 c 3 P i 1 [π 3 (x)], Q 3i+2 (x) = [(x a 1 )(x a 2 ) b 1 c 1 ] P i [π 3 (x)], x a π 3 (x) = b 1 c 1 x a 2 1 b 3 c 3 b 2 c 2 x a 3 and P i (x) = (2 ) [ ] i x b1 c 1 b 2 c 2 b 3 c 3 b 1 b 2 b 3 c 1 c 2 c 3 U i 2 b 1 b 2 b 3 c 1 c 2 c 3
13 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) We may conclude that θ i = ( 1) i Q i (0) In order to compute the φ i s, we define ψ i = φ n+1 i for i = 1,,n In this situation, we have to distinguish three cases If n 0 (mod 3), then ψ 0 = 1, ψ 1 = a 3, ψ 3i = a 1 ψ 3i 1 b 1 c 1 ψ 3i 2, ψ 3i+1 = a 3 ψ 3i b 3 c 3 ψ 3i 1, ψ 3i+2 = a 2 ψ 3i+1 b 2 c 2 ψ 3i If n 1(mod3),then ψ 0 = 1, ψ 1 = a 1, ψ 3i = a 2 ψ 3i 1 b 2 c 2 ψ 3i 2, ψ 3i+1 = a 1 ψ 3i b 1 c 1 ψ 3i 1, ψ 3i+2 = a 3 ψ 3i+1 b 3 c 3 ψ 3i, and if n 2(mod3),then ψ 0 = 1, ψ 1 = a 2, ψ 3i = a 3 ψ 3i 1 b 3 c 3 ψ 3i 2, ψ 3i+1 = a 2 ψ 3i b 2 c 2 ψ 3i 1, ψ 3i+2 = a 1 ψ 3i+1 b 1 c 1 ψ 3i It is clear that these three recurrence relations can be solved by a similar process as for the computation of the θ i s (mutatis mutandis) Following in this way, we can state the next proposition Theorem 51 Let B be the tridiagonal matrix (2), with a 1 a 2 a 3 /= 0 and b 1 b 2 b 3 c 1 c 2 c 3 > 0Let π 3 α β γ ; x x a 1 1 := α x b 1 γ β x c, P i α β γ ; x :=2 αβγ ( [ ]) 1 U i 2 π 3 αβγ α β γ ; x (a + b + c),
14 20 CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) 7 21 Q 3i α β γ ; x := P i α β γ ; x +γ(x b) P i 1 α β γ ; x, Q 3i+1 α β γ ; x := (x a)p i α β γ ; x +αγ P i 1 α β γ ; x, Q 3i+2 α β γ ; x := [(x a)(x b) α] P i α β γ ; x, a, b, c, α, β and γ are some parameters, subject to the restriction αβγ > 0 Under these conditions, ( 1) i+j β i β i+1 β j θ i 1 φ j+1 /θ n if i j, (B 1 ) ij = (19) ( 1) i+j γ j γ j+1 γ i θ j 1 φ i+1 /θ n if i>j, β 3l+s+1 = b s+1, γ 3l+s+1 = c s+1 (s = 0, 1, 2; l = 0, 1, 2,), and the θ i s and φ i s are explicitly given by θ i = ( 1) i a1 a Q 2 a 3 i ; 0 (20) b 1 c 1 b 2 c 2 b 3 c 3 and ( 1) n+1 i a3 a Q 2 a 1 n+1 i ; 0 if n 0, (mod 3), b 2 c 2 b 1 c 1 b 3 c 3 φ i = ( 1) n+1 i a1 a Q 3 a 2 n+1 i ; 0 b 3 c 3 b 2 c 2 b 1 c 1 ( 1) n+1 i a2 a Q 1 a 3 n+1 i ; 0 b 1 c 1 b 3 c 3 b 2 c 2 if n 1, (mod 3), if n 2,(mod 3) Acknowledgments This work was supported by CMUC (Centro de Matemática da Universidade de Coimbra) The authors thank the editor as well as the unknown referees for their helpful remarks which helped to improve the final version of the paper
15 References CM da Fonseca, J Petronilho / Linear Algebra and its Applications 325 (2001) [1] M Gover, The eigenproblem of a tridiagonal 2-Toeplitz matrix, Linear Algebra Appl 197/198 (1994) [2] G Heinig, K Rost, Algebraic Methods for Toeplitz-like Matrices and Operators, vol 13 Birkhäuser, OT, 1984 [3] U Kamps, Chebyshev polynomials and least squares estimation based on one-dependent random variables, Linear Algebra Appl 112 (1989) [4] U Kamps, Relative efficiency of estimates and the use of Chebyshev polynomials in a model of pairwise overlapping samples, Linear Algebra Appl 127 (1990) [5] JW Lewis, Inversion of tridiagonal matrices, Numer Math 38 (1982) [6] F Marcellán, J Petronilho, Eigenproblems for tridiagonal 2-Toeplitz matrices and quadratic polynomials mappings, Linear Algebra Appl 260 (1997) [7] F Marcellán, J Petronilho, Orthogonal polynomials and cubic polynomial mappings I, Communications in the Analytic Theory of Continued Fractions 8 (2000) [8] G Meurant, A review on the inverse of symmetric tridiagonal and block tridiagonal matrices, SIAM J Matrix Anal Appl 13 (3) (1992) [9] R Usmani, Inversion of a tridiagonal Jacobi matrix, Linear Algebra Appl 212/213 (1994) [10] P Schlegel, The explicit inverse of a tridiagonal matrix, Math Computation 24 (111) (1970) 665
a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationSolving Linear Systems, Continued and The Inverse of a Matrix
, Continued and The of a Matrix Calculus III Summer 2013, Session II Monday, July 15, 2013 Agenda 1. The rank of a matrix 2. The inverse of a square matrix Gaussian Gaussian solves a linear system by reducing
More informationPYTHAGOREAN TRIPLES KEITH CONRAD
PYTHAGOREAN TRIPLES KEITH CONRAD 1. Introduction A Pythagorean triple is a triple of positive integers (a, b, c) where a + b = c. Examples include (3, 4, 5), (5, 1, 13), and (8, 15, 17). Below is an ancient
More informationA note on companion matrices
Linear Algebra and its Applications 372 (2003) 325 33 www.elsevier.com/locate/laa A note on companion matrices Miroslav Fiedler Academy of Sciences of the Czech Republic Institute of Computer Science Pod
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationI. GROUPS: BASIC DEFINITIONS AND EXAMPLES
I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationFACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z
FACTORING POLYNOMIALS IN THE RING OF FORMAL POWER SERIES OVER Z DANIEL BIRMAJER, JUAN B GIL, AND MICHAEL WEINER Abstract We consider polynomials with integer coefficients and discuss their factorization
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationChapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
More informationNotes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
More informationLectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain
Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal
More informationT ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
More informationMA106 Linear Algebra lecture notes
MA106 Linear Algebra lecture notes Lecturers: Martin Bright and Daan Krammer Warwick, January 2011 Contents 1 Number systems and fields 3 1.1 Axioms for number systems......................... 3 2 Vector
More informationThe Matrix Elements of a 3 3 Orthogonal Matrix Revisited
Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
More informationLinear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
More informationMATRICES WITH DISPLACEMENT STRUCTURE A SURVEY
MATRICES WITH DISPLACEMENT STRUCTURE A SURVEY PLAMEN KOEV Abstract In the following survey we look at structured matrices with what is referred to as low displacement rank Matrices like Cauchy Vandermonde
More informationLinear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
More informationMATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar
More informationMath 115A HW4 Solutions University of California, Los Angeles. 5 2i 6 + 4i. (5 2i)7i (6 + 4i)( 3 + i) = 35i + 14 ( 22 6i) = 36 + 41i.
Math 5A HW4 Solutions September 5, 202 University of California, Los Angeles Problem 4..3b Calculate the determinant, 5 2i 6 + 4i 3 + i 7i Solution: The textbook s instructions give us, (5 2i)7i (6 + 4i)(
More informationMath 115 Spring 2011 Written Homework 5 Solutions
. Evaluate each series. a) 4 7 0... 55 Math 5 Spring 0 Written Homework 5 Solutions Solution: We note that the associated sequence, 4, 7, 0,..., 55 appears to be an arithmetic sequence. If the sequence
More informationCONTINUED FRACTIONS AND PELL S EQUATION. Contents 1. Continued Fractions 1 2. Solution to Pell s Equation 9 References 12
CONTINUED FRACTIONS AND PELL S EQUATION SEUNG HYUN YANG Abstract. In this REU paper, I will use some important characteristics of continued fractions to give the complete set of solutions to Pell s equation.
More informationTHE SIGN OF A PERMUTATION
THE SIGN OF A PERMUTATION KEITH CONRAD 1. Introduction Throughout this discussion, n 2. Any cycle in S n is a product of transpositions: the identity (1) is (12)(12), and a k-cycle with k 2 can be written
More informationMatrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.
Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationMATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
More informationA characterization of trace zero symmetric nonnegative 5x5 matrices
A characterization of trace zero symmetric nonnegative 5x5 matrices Oren Spector June 1, 009 Abstract The problem of determining necessary and sufficient conditions for a set of real numbers to be the
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More informationv w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.
3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with
More informationQuotient Rings and Field Extensions
Chapter 5 Quotient Rings and Field Extensions In this chapter we describe a method for producing field extension of a given field. If F is a field, then a field extension is a field K that contains F.
More informationIntroduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
More informationMinimally Infeasible Set Partitioning Problems with Balanced Constraints
Minimally Infeasible Set Partitioning Problems with alanced Constraints Michele Conforti, Marco Di Summa, Giacomo Zambelli January, 2005 Revised February, 2006 Abstract We study properties of systems of
More information1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
More information4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
More informationAu = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
More information5.3 The Cross Product in R 3
53 The Cross Product in R 3 Definition 531 Let u = [u 1, u 2, u 3 ] and v = [v 1, v 2, v 3 ] Then the vector given by [u 2 v 3 u 3 v 2, u 3 v 1 u 1 v 3, u 1 v 2 u 2 v 1 ] is called the cross product (or
More informationLecture L3 - Vectors, Matrices and Coordinate Transformations
S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between
More informationMatrix Differentiation
1 Introduction Matrix Differentiation ( and some other stuff ) Randal J. Barnes Department of Civil Engineering, University of Minnesota Minneapolis, Minnesota, USA Throughout this presentation I have
More informationUniversity of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
More informationMATH10040 Chapter 2: Prime and relatively prime numbers
MATH10040 Chapter 2: Prime and relatively prime numbers Recall the basic definition: 1. Prime numbers Definition 1.1. Recall that a positive integer is said to be prime if it has precisely two positive
More informationF Matrix Calculus F 1
F Matrix Calculus F 1 Appendix F: MATRIX CALCULUS TABLE OF CONTENTS Page F1 Introduction F 3 F2 The Derivatives of Vector Functions F 3 F21 Derivative of Vector with Respect to Vector F 3 F22 Derivative
More informationMath 181 Handout 16. Rich Schwartz. March 9, 2010
Math 8 Handout 6 Rich Schwartz March 9, 200 The purpose of this handout is to describe continued fractions and their connection to hyperbolic geometry. The Gauss Map Given any x (0, ) we define γ(x) =
More informationChapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
More informationMath 319 Problem Set #3 Solution 21 February 2002
Math 319 Problem Set #3 Solution 21 February 2002 1. ( 2.1, problem 15) Find integers a 1, a 2, a 3, a 4, a 5 such that every integer x satisfies at least one of the congruences x a 1 (mod 2), x a 2 (mod
More information4. How many integers between 2004 and 4002 are perfect squares?
5 is 0% of what number? What is the value of + 3 4 + 99 00? (alternating signs) 3 A frog is at the bottom of a well 0 feet deep It climbs up 3 feet every day, but slides back feet each night If it started
More informationSF2940: Probability theory Lecture 8: Multivariate Normal Distribution
SF2940: Probability theory Lecture 8: Multivariate Normal Distribution Timo Koski 24.09.2015 Timo Koski Matematisk statistik 24.09.2015 1 / 1 Learning outcomes Random vectors, mean vector, covariance matrix,
More informationOverview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
More informationSOME PROPERTIES OF SYMBOL ALGEBRAS OF DEGREE THREE
SOME PROPERTIES OF SYMBOL ALGEBRAS OF DEGREE THREE CRISTINA FLAUT and DIANA SAVIN Communicated by the former editorial board In this paper, we study some properties of the matrix representations of the
More informationCITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION
No: CITY UNIVERSITY LONDON BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION ENGINEERING MATHEMATICS 2 (resit) EX2005 Date: August
More informationThe Determinant: a Means to Calculate Volume
The Determinant: a Means to Calculate Volume Bo Peng August 20, 2007 Abstract This paper gives a definition of the determinant and lists many of its well-known properties Volumes of parallelepipeds are
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
More informationArkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan
Arkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan 3 Binary Operations We are used to addition and multiplication of real numbers. These operations combine two real numbers
More informationWHEN DOES A CROSS PRODUCT ON R n EXIST?
WHEN DOES A CROSS PRODUCT ON R n EXIST? PETER F. MCLOUGHLIN It is probably safe to say that just about everyone reading this article is familiar with the cross product and the dot product. However, what
More informationSYSTEMS OF PYTHAGOREAN TRIPLES. Acknowledgements. I would like to thank Professor Laura Schueller for advising and guiding me
SYSTEMS OF PYTHAGOREAN TRIPLES CHRISTOPHER TOBIN-CAMPBELL Abstract. This paper explores systems of Pythagorean triples. It describes the generating formulas for primitive Pythagorean triples, determines
More information8 Square matrices continued: Determinants
8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You
More informationKevin James. MTHSC 412 Section 2.4 Prime Factors and Greatest Comm
MTHSC 412 Section 2.4 Prime Factors and Greatest Common Divisor Greatest Common Divisor Definition Suppose that a, b Z. Then we say that d Z is a greatest common divisor (gcd) of a and b if the following
More informationFinite dimensional C -algebras
Finite dimensional C -algebras S. Sundar September 14, 2012 Throughout H, K stand for finite dimensional Hilbert spaces. 1 Spectral theorem for self-adjoint opertors Let A B(H) and let {ξ 1, ξ 2,, ξ n
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationα = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationEquations, Inequalities & Partial Fractions
Contents Equations, Inequalities & Partial Fractions.1 Solving Linear Equations 2.2 Solving Quadratic Equations 1. Solving Polynomial Equations 1.4 Solving Simultaneous Linear Equations 42.5 Solving Inequalities
More informationarxiv:math/0202219v1 [math.co] 21 Feb 2002
RESTRICTED PERMUTATIONS BY PATTERNS OF TYPE (2, 1) arxiv:math/0202219v1 [math.co] 21 Feb 2002 TOUFIK MANSOUR LaBRI (UMR 5800), Université Bordeaux 1, 351 cours de la Libération, 33405 Talence Cedex, France
More information160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
More informationThe Notebook Series. The solution of cubic and quartic equations. R.S. Johnson. Professor of Applied Mathematics
The Notebook Series The solution of cubic and quartic equations by R.S. Johnson Professor of Applied Mathematics School of Mathematics & Statistics University of Newcastle upon Tyne R.S.Johnson 006 CONTENTS
More informationLecture 2 Matrix Operations
Lecture 2 Matrix Operations transpose, sum & difference, scalar multiplication matrix multiplication, matrix-vector product matrix inverse 2 1 Matrix transpose transpose of m n matrix A, denoted A T or
More informationNOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
More informationIntroduction to Algebraic Geometry. Bézout s Theorem and Inflection Points
Introduction to Algebraic Geometry Bézout s Theorem and Inflection Points 1. The resultant. Let K be a field. Then the polynomial ring K[x] is a unique factorisation domain (UFD). Another example of a
More informationName: Section Registered In:
Name: Section Registered In: Math 125 Exam 3 Version 1 April 24, 2006 60 total points possible 1. (5pts) Use Cramer s Rule to solve 3x + 4y = 30 x 2y = 8. Be sure to show enough detail that shows you are
More information5. Orthogonal matrices
L Vandenberghe EE133A (Spring 2016) 5 Orthogonal matrices matrices with orthonormal columns orthogonal matrices tall matrices with orthonormal columns complex matrices with orthonormal columns 5-1 Orthonormal
More informationGENERATING SETS KEITH CONRAD
GENERATING SETS KEITH CONRAD 1 Introduction In R n, every vector can be written as a unique linear combination of the standard basis e 1,, e n A notion weaker than a basis is a spanning set: a set of vectors
More informationRecall the basic property of the transpose (for any A): v A t Aw = v w, v, w R n.
ORTHOGONAL MATRICES Informally, an orthogonal n n matrix is the n-dimensional analogue of the rotation matrices R θ in R 2. When does a linear transformation of R 3 (or R n ) deserve to be called a rotation?
More informationInner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
More informationDecember 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
More informationTHREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
More informationMATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column
More informationThe Method of Partial Fractions Math 121 Calculus II Spring 2015
Rational functions. as The Method of Partial Fractions Math 11 Calculus II Spring 015 Recall that a rational function is a quotient of two polynomials such f(x) g(x) = 3x5 + x 3 + 16x x 60. The method
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More informationa 1 x + a 0 =0. (3) ax 2 + bx + c =0. (4)
ROOTS OF POLYNOMIAL EQUATIONS In this unit we discuss polynomial equations. A polynomial in x of degree n, where n 0 is an integer, is an expression of the form P n (x) =a n x n + a n 1 x n 1 + + a 1 x
More informationSUM OF TWO SQUARES JAHNAVI BHASKAR
SUM OF TWO SQUARES JAHNAVI BHASKAR Abstract. I will investigate which numbers can be written as the sum of two squares and in how many ways, providing enough basic number theory so even the unacquainted
More informationMatrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
More information1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
More informationLecture 3: Finding integer solutions to systems of linear equations
Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture
More informationSection 4.4 Inner Product Spaces
Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer
More informationThe last three chapters introduced three major proof techniques: direct,
CHAPTER 7 Proving Non-Conditional Statements The last three chapters introduced three major proof techniques: direct, contrapositive and contradiction. These three techniques are used to prove statements
More informationMath 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
More informationNotes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
More informationDIFFERENTIATION AND INTEGRATION BY USING MATRIX INVERSION
Journal of Applied Mathematics and Computational Mechanics 2014, 13(2), 63-71 DIFFERENTIATION AND INTEGRATION BY USING MATRIX INVERSION Dagmara Matlak, Jarosław Matlak, Damian Słota, Roman Wituła Institute
More informationSection 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
More informationLinear Algebra: Determinants, Inverses, Rank
D Linear Algebra: Determinants, Inverses, Rank D 1 Appendix D: LINEAR ALGEBRA: DETERMINANTS, INVERSES, RANK TABLE OF CONTENTS Page D.1. Introduction D 3 D.2. Determinants D 3 D.2.1. Some Properties of
More informationMAT188H1S Lec0101 Burbulla
Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u
More informationLecture 5 Principal Minors and the Hessian
Lecture 5 Principal Minors and the Hessian Eivind Eriksen BI Norwegian School of Management Department of Economics October 01, 2010 Eivind Eriksen (BI Dept of Economics) Lecture 5 Principal Minors and
More informationarxiv:1404.6042v1 [math.dg] 24 Apr 2014
Angle Bisectors of a Triangle in Lorentzian Plane arxiv:1404.604v1 [math.dg] 4 Apr 014 Joseph Cho August 5, 013 Abstract In Lorentzian geometry, limited definition of angles restricts the use of angle
More informationVariance Reduction. Pricing American Options. Monte Carlo Option Pricing. Delta and Common Random Numbers
Variance Reduction The statistical efficiency of Monte Carlo simulation can be measured by the variance of its output If this variance can be lowered without changing the expected value, fewer replications
More informationNotes on Linear Algebra. Peter J. Cameron
Notes on Linear Algebra Peter J. Cameron ii Preface Linear algebra has two aspects. Abstractly, it is the study of vector spaces over fields, and their linear maps and bilinear forms. Concretely, it is
More informationSF2940: Probability theory Lecture 8: Multivariate Normal Distribution
SF2940: Probability theory Lecture 8: Multivariate Normal Distribution Timo Koski 24.09.2014 Timo Koski () Mathematisk statistik 24.09.2014 1 / 75 Learning outcomes Random vectors, mean vector, covariance
More informationLinear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices
MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two
More informationIntroduction to Matrices for Engineers
Introduction to Matrices for Engineers C.T.J. Dodson, School of Mathematics, Manchester Universit 1 What is a Matrix? A matrix is a rectangular arra of elements, usuall numbers, e.g. 1 0-8 4 0-1 1 0 11
More information