MODIFIED INCOMPLETE CHOLESKY FACTORIZATION FOR SOLVING ELECTROMAGNETIC SCATTERING PROBLEMS
|
|
|
- Matthew McKenzie
- 9 years ago
- Views:
Transcription
1 Progress In Electromagnetics Research B, Vol. 13, 41 58, 2009 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION FOR SOLVING ELECTROMAGNETIC SCATTERING PROBLEMS T.-Z. Huang, Y. Zhang, and L. Li School of Applied Mathematics/Institute of Computational Science University of Electronic Science and Technology of China Chengdu , Sichuan, China W. Shao and S. Lai School of Physical Electronics University of Electronic Science and Technology of China Chengdu , Sichuan, China Abstract In this paper, we study a class of modified incomplete Cholesky factorization preconditioners LL T with two control parameters including dropping rules. Before computing preconditioners, the modified incomplete Cholesky factorization algorithm allows to decide the sparsity of incomplete factorization preconditioners by two fillin control parameters: (1) p, the number of the largest number p of nonzero entries in each row; (2) dropping tolerance. With RCM reordering scheme as a crucial operation for incomplete factorization preconditioners, our numerical results show that both the number of PCOCG and PCG iterations and the total computing time are reduced evidently for appropriate fill-in control parameters. Numerical tests on harmonic analysis for 2D and 3D scattering problems show the efficiency of our method. 1. INTRODUCTION The coefficient matrix of the linear equations which stem from the finite-element analysis of high-frequency electromagnetic field simulations such as scattering [1 9] is generally symmetric and indefinite. At present, the incomplete Cholesky factorization [10 13] (IC) preconditioners applied with preconditioned Conjugate Gradient (PCG) method and preconditioned Conjugate Orthogonal Conjugate
2 42 Huang et al. Gradient (PCOCG) method are rather popular [14 16]. Here incomplete Cholesky factorization is studied in the case of finite element (FEM) matrices arising from the discretization of the following electromagnetic scattering problem: ( ) ( ) 1 1 E sc k0ε 2 r E sc = E inc + k0e 2 inc, (1) µr µr with some absorbing boundary conditions, where E sc is the scattering field, E inc is the incident field and µ r and ε r are relative permeability and permittivity, respectively. The solution of Eq. (1) will result in a linear system Ax = b, (2) where A = (a ij ) n n C n n is sparse complex symmetric (usually indefinite), x, b C n. In order to solve (2) effectively, incomplete LU factorization preconditioners are often associated with some preconditioned Krylov subspace methods such as BICGSTAB, QMR, TFQMR, CG, COCG [17 19]. To make full use of symmetry of the systems, incomplete Cholesky factorization is normally utilized with some preconditioned Krylov subspace methods such as PCG and PCOCG. Incomplete Cholesky factorization was designed for solving symmetric positive definite systems. The performance of the incomplete Cholesky factorization often relies on drop tolerances [13, 17] to reduce fill-ins. The properties of the incomplete Cholesky factorization depend, in part, on the sparsity pattern S of the incomplete Cholesky factor L =(l ij ) n n, where L is a lower triangular matrix such that [10] A = LL T + R, l ij = 0if (i, j) / S. The aim of the presented numerical tests is to analyze the performance of the studied incomplete Cholesky factorization algorithms. Our consideration is to focus on the performance of the proposed modified incomplete Cholesky factorization preconditioners with tuning of sparsity with PCG and PCOCG as accelerators. And we intend to find their impacts on these scattering problems discretized by FEM. Many research papers about incomplete Cholesky factorization can be found, such as Lin and More [10], Fang and Leary [11], Margenov and Popov [12], the fixed fill factorization of Meijerink and Vorst [20], the ILUT factorization of Saad [13, 17]. For additional information on incomplete Cholesky factorizations, please refer to Saad [17]. Reordering methods are very important for incomplete
3 Progress In Electromagnetics Research B, Vol. 13, factorization; see more in [21 26]. In [10], a new incomplete Cholesky factorization algorithm is proposed which is designed to limit the memory requirement by specifying the amount of additional memory. In contrast with drop tolerance strategies, the new approach in [10] is more stable in terms of number of iterations and memory requirements. In this paper, we intend to apply this approach as preconditioners for solving scattering problems and get more effective incomplete Cholesky factorization algorithm based on the work of Lin and More in [10]. Additionally, the reordering in the matrix plays an important role in the application of preconditioning technologies because the ordering of the matrix affects the fill in the matrix and thus the incomplete Cholesky factorization [21]. In this paper, both the AMD and RCM orderings [17, 21] are applied to reorder our linear system. The rest of the paper is organized as follows: In section 2 we survey some relative preconditioning algorithms and Krylov subspace methods. The modified incomplete factorization algorithm is presented in section 3 with detailed description of its implementation. In Section 4, a set of numerical experiments are presented and short concluding remarks are given in Section PRECONDITIONERS AND ITERATIVE METHODS Our implementation of the incomplete Cholesky factorization is based on the jki version of the Cholesky factorization shown below in Algorithm 1 [10]. Note that diagonal elements are updated as the factorization proceeds. Obviously, Algorithm 1 is based on the columnoriented Cholesky factorization for sparse matrices. Algorithm 1: Column-oriented Cholesky Factorization [10, Algorithm 2.1] 1. for j =1:n 2. a jj = a jj 3. for k =1:j 1 4. for i = j +1:n 5. a ij = a ij a ik a jk 6. endfor 7. endfor 8. for i = j +1:n 9. a ij = a ij /a jj 10. a ii = a ii a 2 ij 11. endfor 12. endfor
4 44 Huang et al. For a symmetric coefficient matrix A, we only need to store the lower or the upper triangular parts of A. In Algorithm 1, only the lower triangular part of A including the diagonal entries is needed. And the access to A is column by column. Therefore, in order to compute the IC-type preconditioner by Algorithm 1, we only need to store the lower triangular part L as the incomplete Cholesky factor. In order to show the detailed computation process of L, we transform Algorithm 1 into a comprehensible version with explicit computation of L: Algorithm 2: Column-oriented Cholesky Factorization with Explicit Expression of L 1. for j =1:n 2. l jj = a jj 3. for k =1:j 1 4. for i = j +1:n 5. l ij = l ij l ik l jk 6. endfor 7. endfor 8. for i = j +1:n 9. l ij = l ij /l jj 10. a ii = a ii lij endfor 12. endfor In [10], the following Algorithm 3 has been discussed in details. Algorithm 3. Column-oriented Cholesky Factorization [10, Algorithm 2.2] 1. for j =1:n 2. a jj = a jj 3. L col len = size(i >j: a ij 0) 4. for k =1:j 1 and a jk 0 5. for i = j +1:n and a ik 0 6. a ij = a ij a ik a jk 7. endfor 8. endfor 9. for i = j +1:n and a ij a ij = a ij /a jj 11. a ii = a ii a 2 ij 12. endfor 13. Retain the largest L col len + p elements in a j+1:n,j 14. endfor Notice the symbol a j+1:n,j means these entries of the j-th column from row j +1 to row n of coefficient matrix A. For iterative solution of
5 Progress In Electromagnetics Research B, Vol. 13, the symmetric linear system (2), we choose the precondtioned COCG and precondtioned CG methods (See more in [14 16]). 3. MODIFIED INCOMPLETE CHOLESKY FACTORIZATION ALGORITHM WITH ITS IMPLEMENTATION In the light of ILUT algorithm in [17, p.287] and Algorithm 3 in [10, p.29], we present the following column-oriented MIC(p, τ) algorithm for obtaining the incomplete Cholesky factor L. Algorithm 4. Modified Incomplete Cholesky factorization (MIC(p, τ)) 1. for j =1:n 2. l jj = a jj 3. w = a j+1:n,j 4. for k =1:j 1 5. for i = j +1:n and when l jk 0 6. w i = w i l ik l jk 7. endfor 8. endfor 9. for i = j +1:n 10. w ij = w ij /l jj 11. endfor 12. τ j = τ w 13. for i = j +1:n 14. w i = 0when w i <τ j 15. endfor 16. An integer array I =(i k ) k=1,,p contains indices of the first largest p entries of w i,i= j +1:n. 17. for k =1:p 18. l ik j = w ik j 19. endfor 20. for i = j +1:n 21. a ii = a ii lij endfor 23. endfor In the case that the lower triangular matrix L keeps the same nonzero pattern as that of the lower triangular part of A, Algorithm 1 leads to IC(0) algorithm (i.e., incomplete Cholesky factorization preconditioners with the same nonzero pattern as that of coefficient matrix A).
6 46Huang et al. In order to implement Algorithm 4, we store the upper triangular part of the coefficient matrix A in compressed sparse row (CSR) format. However, for convenience, Algorithm 4 needs to access lower triangular part of A in compressed sparse column (CSC) format. Observed from the data structures of CSR and CSC, the upper triangular part of the coefficient matrix A stored in CSR is exactly the lower triangular part of A stored in CSC. So, we don t need to perform the transform operation from the input matrix (the upper triangular part of the coefficient matrix A) into the CSC format of the lower triangular part of the coefficient matrix A. For simplicity, variable L here denotes the CSC format of L. From Line 6 in Algorithm 4, in order to compute the linear combination vector w, we need to access the j-th row and k-th column of L. The access of k th column of L is convenient because L is just stored in CSC. The difficulty in Line 6 is how to access the j-th row of L which is stored in CSC format. In order to get high efficiency of accessing rows of L, we introduce a temporary CSR variable U to store the CSR format of L. In iterative methods, we need to solve the preconditioning system LL T x = y. Normally, L and L T are stored in CSR format, respectively. In fact, the transformation from the CSC format of L to the CSR foramt of L is unnecessary because variable U in CSR format is just the CSR format of L and variable L in CSC format is just the CSR format of L T. 4. NUMERICAL TESTS All numerical tests are performed on Linux operating system. All codes are programmed in C language and implemented on a PC, with 2 GB memory and a 2.66 GHz Intel(R) Core(TM)2 Duo CPU. In order to operate the complex type elements in computation, we declare double complex type variables which are supported directly by gcc compiler. The maximal iteration number is The iteration stops when r (k) / r (0) < Since ordering is crucial to a good factorized preconditioner, the reverse Cuthill-McKee (RCM) reordering and Approximate Minimum Degree (AMD) reordering are applied before computing L. Denote the number of nonzero elements of a matrix as nnz (matrix name), the iteration number as its, the incomplete factorization CPU time as P-t and iteration CPU time as I-t, total computation time (preconditioning time plus iteration time) as T-t. All consuming time is measured in seconds. Denote sparse ratio of a preconditioner i.e., nnz(l+lt ) n nnz(a) as sp-r.
7 Progress In Electromagnetics Research B, Vol. 13, Problem 1 (Harmonic Analysis for Plane Wave Scattering from a Metallic Plate): In this problem, we use edge-based FEM to calculate the RCS of a PEC plate (1λ o 1λ o ) where λ o stands for the free space wavelength of the incident plane wave. Applying PEC boundary condition, we need to solve a system of linear equations of size 5381 with a complex coefficient matrix containing nonzero elements in the upper triangular. Problem 2 (Harmonic Analysis for Scattering of a Dielectric Sphere): In this problem, perfectly matched layers (PML) are used to truncate the finite element analysis domain in order to determine the radar cross section (RCS) from the scattering of a dielectric sphere. The relative permittivity of the dielectric sphere is ε r = To calculate the bistatic RCS of the dielectric sphere, the incident plane wave is taken as x-polarized with incident angles φ =0 and θ =0, which leads to a system of linear equations of size with a complex coefficient matrix containing nonzero elements in the upper triangular. (a) Original (b) RCM Figure 1. Nonzero pattern of the coefficient matrix from Problem 1 with Original ordering and RCM reordering. For Problems 1 and 2, without RCM reordering, PCOCG and PCG methods do not converge within 5000 iterations. In order to evaluate the performance of the proposed algorithm, the IC(0) preconditioner (i.e. L + L T has the same nonzero pattern with that of A) and the diagonal preconditioner are exploited (see Table 1 for details). Numerical results of the solution of Problem 1 and Problem 2 with PCOCG and PCG methods associated with RCM reordering are presented in Tables 2-7. For Problem 1, AMD reordering nearly failed in all cases with the same parameters of RCM reordering, except in two cases. For Problem 2, AMD reordering failed in all cases. So the results with AMD reordering are all ignored in this section.
8 48 Huang et al. Table 1. Problem 1. PCOCG with IC(0) and diagonal preconditioners for Preconditioner IC(0) Diagonal preconditioner ordering NO RCM NO RCM Its By comparing Table 1 with Table 2, it is obvious that our MIC (p, τ) preconditioner is much more efficient than IC(0) and diagonal preconditioners. Observed from Tables 2 and 3, the number of iterations and total computation time of the two kinds of iterative methods of PCOCG and PCG are almost the same in all cases with the same parameters in MIC(p, τ). From Tables 2 4 and Fig. 1, it is noticed that the effect of dropping tolerance τ in such a wide range for Problem 1 is not prominent. Observed from Fig. 1, there is a jump of computation time with parameters p = 30and τ =10 3. From a general view, it is a specific case. However, from Tables 5-7 and Fig. 2 of the test of Problem 2, it is possible for us to draw the conclusion that the effect of dropping tolerance τ is obvious in the solution of larger scale problems. And the reasonable range for τ should arrange from 10 4 to In addition, the parameter τ has minor effect on memory requirement of our MIC preconditioner. What affects remarkably is the parameter p which also decides both fill-ins and efficiency of MIC preconditioner. The larger p is, (a) Original (b) RCM Figure 2. Nonzero pattern of the coefficient matrix from Problem 2 with Original ordering and RCM reordering.
9 Progress In Electromagnetics Research B, Vol. 13, Table 2. Problem 1. PCOCG with RCM reordering and Algorithm 6 for τ p nnz(l) sp-r its P-t I-t T-t
10 50 Huang et al. Table 3. PCG with RCM reordering and Algorithm 6 for Problem 1. τ p nnz(l) sp-r its P-t I-t T-t
11 Progress In Electromagnetics Research B, Vol. 13, Table 4. PCOCG with RCM reordering and Algorithm 6(τ = 0) for Problem 1. p nnz(l) sp-r its P-t I-t T-t the less the iteration number becomes while the more the fill-ins are required. However, the total computation time is not necessarily decreasing with the growth of p, which implies that it is crucial to select an appropriate parameter p. Generally, parameter p can be evaluated by setting the number of nonzero entries of incomplete Cholesky preconditioners, i.e., p = nnz(l) n where L is the incomplete Cholesky preconditioner and n is the dimension of coefficient matrix A. However, the number of nonzero entries of incomplete Cholesky preconditioners L is determined by the coefficient matrix A of linear system. For small-scale linear system such as Problem 1, proper set of the number of nonzero entries of L is about 2 times (or more) of that nnz(l) 5.5 x p=30 p=40 p=50 p=60 p=70 p=80 p=90 p= log τ 10 (a) fill-in Total computation time (s) p=30 p=40 p=50 p=60 p=70 p=80 p=90 p= log τ (b) Computation time with PCOCG Figure 3. Comparisons using fill-in and total computation time for Problem 1.
12 52 Huang et al. Table 5. Problem 2. PCOCG with RCM reordering and Algorithm 6 for τ p nnz(l) sp-r its P-t I-t T-t
13 Progress In Electromagnetics Research B, Vol. 13, Table 6. PCG with RCM reordering and Algorithm 6 for Problem 2. τ p nnz(l) sp-r its P-t I-t T-t
14 54 Huang et al. Table 7. PCOCG with RCM reordering and Algorithm 6(τ =0) for Problem 2. p nnz(l) sp-r its P-t I-t T-t of A. For middle-scale linear system such as Problem 2, the select of the number of nonzero entries of L could be 5 (or more) times of that of A. In order to compare the performance of Algorithm 4 (MIC(p,τ) with that of Algorithm 3, numerical experiments with Algorithm 3 are also performed. Note that parameter p in Algorithms 3 and 4 has different meanings. Observed from Tables 8 and 9, Algorithm 3 needs more memory than Algorithm 4 under the requirement of the same total computation time. Take Problem 2 for example. The minimum computation time with Algorithm 3 is (s) and the fill-ins of L is Nevertheless, using Algorithm 4, it consumes nnz(l) 3.4 x p=180 p=190 p=200 p=210 p=220 p=230 p=240 p= log τ 10 (a) Fill-in Total computation time (s) p=180 p=190 p=200 p=210 p=220 p=230 p=240 p= log τ (b) Computation time with PCOCG Figure 4. Comparisons using fill-in and total computation time for Problem 2.
15 Progress In Electromagnetics Research B, Vol. 13, Table 8. Problem 1. PCOCG with RCM reordering and Algorithm 3 for p nnz(l) sp-r its P-t I-t T-t Table 9. Problem 2. PCOCG with RCM reordering and Algorithm 3 for p nnz(l) sp-r its P-t I-t T-t (s) with parameters p = 230and τ =10 5 and the fill-ins of L is Additionally, as illustrated in Table 10 for Problem 1, Algorithm 4 is dramatically superior to Algorithm 3 in both aspects of total computation time and memory requirement.
16 56Huang et al. Table 10. Comparison results with respect to the minimum of total computation time and the corresponding memory between Algorithms 3 and 6 for Problem 1. min(t-t) nnz(l) Reduction Reduction Alg. 3 Alg. 6 Alg. 3 Alg. 6 ratio(%) ratio(%) CONCLUSIONS A column-oriented modified incomplete Cholesky factorization MIC (p, τ) with two controlling parameters for solution of systems of linear equations with sparse complex symmetric coefficient matrices resulted from finite-element analysis of the electromagnetic scattering problem (1) is presented in this paper. Proper choices of the controlling parameters in Algorithm 6 can evidently reduce the total computation time and memory requirements compared with Algorithm 3. It is worthwhile to emphasize that the involved parameter p, which prescribes the maximal fill-ins in each row of preconditioners, makes Algorithm 6 evidently superior to Algorithm 3 in the number of fill-ins, and helps to reduce total computation time of Algorithm 6. As shown in the numerical experiments, RCM ordering is obviously superior to AMD ordering. Moreover, RCM ordering is significant to our modified incomplete Cholesky factorization. Numerical experiments show that further developments of more proper incomplete factorization algorithms and reordering schemes for electromagnetic scattering problems are deserved to be taken into consideration in the future. ACKNOWLEDGMENT This research is supported by 973 Programs (2008CB317110), NSFC ( ), the Scientific and Technological Key Project of the Chinese Education Ministry (107098), the Specialized Research Fund for the Doctoral Program of Higher Education ( ), Sichuan Province Project for Applied Basic Research (2008JY0052) and the Project for Academic Leader and Group of UESTC.
17 Progress In Electromagnetics Research B, Vol. 13, REFERENCES 1. Jin, J. M., The Finite Element Method in Electromagnetics, Wiley, New York, Volakis, J. L., A. Chatterjee, and L. C. Kempel, Finite Element Method for Electromagnetics: Antennas, Microwave Circuits and Scattering Applications, IEEE Press, New York, Ahmed, S. and Q. A. Naqvi, Electromagnetic scattering of two or more incident plane waves by a perfect, electromagnetic conductor cylinder coated with a metamaterial, Progress In Electromagnetics Research B, Vol. 10, 75 90, Fan, Z. H., D. Z. Ding, and R. S. Chen, The efficient analysis of electromagnetic scattering from composite structures using hybrid CFIE-IEFIE, Progress In Electromagnetics Research B, Vol. 10, , Botha, M. M. and D. B. Davidson, Rigorous auxiliary variablebased implementation of a second-order ABC for the vector FEM, IEEE Trans. Antennas Propagat., Vol. 54, , Harrington, R. F., Field Computation by Moment Method, 2nd edition, IEEE Press, New York, Choi, S. H., D. W. Seo, and N. H. Myung, Scattering analysis of open-ended cavity with inner object, J. of Electromagn. Waves and Appl., Vol. 21, No. 12, , Ruppin, R., Scattering of electromagnetic radiation by a perfect electromagnetic conductor sphere, J. of Electromagn. Waves and Appl., Vol. 20, No. 12, , Ho, M., Scattering of electromagnetic waves from, vibrating perfect surfaces: Simulation using relativistic boundary conditions, J. of Electromagn. Waves and Appl., Vol. 20, No. 4, , Lin, C.-J. and J. J. More, Incomplete cholesky factorizations with limited memory, SIAM J. Sci. Comput., Vol. 21, 24 45, Fang, H.-R. and P. O. Dianne, Leary, modified Cholesky algorithms: A catalog with new approaches, Mathematical Programming, July Margenov, S. and P. Popov, MIC(0) DD preconditioning of FEM elasticity problem on non-structured meshes, Proceedings of ALGORITMY 2000 Conference on Scientific Computing, , Saad, Y., ILUT: A dual threshold incomplete LU factorization, Numer. Linear Algebra Appl., Vol. 4, , 1994.
18 58 Huang et al. 14. Freund, R. and N. Nachtigal, A quasi-minimal residual method for non-hermitian linear systems, Numer. Math., Vol. 60, , Van der Vorst, H. A. and J. B. M. Melissen, A Petrov-Galerkin type method for solving Ax = b, where A is symmetric complex, IEEE Trans. Mag., Vol. 26, No. 2, , Barrett, R., M. Berry, T. F. Chan, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, C. Romine, and H. van der Vorst, Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods, 2nd edition, SIAM, Philadelphia, PA, Saad, Y., Iterative Methods for Sparse Linear Systems, 2nd edition, SIAM, Philadelphia, PA, Saad, Y., Sparskit: A basic tool kit for sparse matrix computations, Report RIACS-90-20, Research Institute for Advanced Computer Science, NASA Ames Research Center, Moffett Field, CA, Benzi, M., Preconditioning techniques for large linear systems: A survey, J. Comp. Physics, Vol. 182, , Meijerink, J. A. and H. A. van der Vorst, An iterative solution method for linear equations systems of which the coefficient matrix is a symmetric M-matrix, Math. Comp., Vol. 31, , Lee, I., P. Raghavan, and E. G. Ng, Effective preconditioning through ordering interleaved with incomplete factorization, Siam J. Matrix Anal. Appl., Vol. 27, , Ng, E. G. and P. Raghavan, Performance of greedy ordering heuristics for sparse Cholesky factorization, Siam J. Matrix Anal. Appl., Vol. 20, No. 2, , Benzi, M., D. B. Szyld, and A. van Duin, Orderings for incomplete factorization preconditioning of nonsymmetric problems, SIAM J. Sci. Comput., Vol. 20, , Benzi, M., W. Joubert, and G. Mateescu, Numerical experiments with parallel orderings for ILU preconditioners, Electronic Transactions on Numerical Analysis, Vol. 8, , Chan, T. C. and H. A. van der Vorst, Approximate and incomplete factorizations, Preprint 871, Department of Mathematics, University of Utrecht, The Netherlands, Zhang, Y., T.-Z. Huang, and X.-P. Liu, Modified iterative methods for nonnegative matrices and M-matrices linear systems, Computers and Mathematics with Applications, Vol. 50, , 2005.
6. Cholesky factorization
6. Cholesky factorization EE103 (Fall 2011-12) triangular matrices forward and backward substitution the Cholesky factorization solving Ax = b with A positive definite inverse of a positive definite matrix
Yousef Saad University of Minnesota Computer Science and Engineering. CRM Montreal - April 30, 2008
A tutorial on: Iterative methods for Sparse Matrix Problems Yousef Saad University of Minnesota Computer Science and Engineering CRM Montreal - April 30, 2008 Outline Part 1 Sparse matrices and sparsity
GPR Polarization Simulation with 3D HO FDTD
Progress In Electromagnetics Research Symposium Proceedings, Xi an, China, March 6, 00 999 GPR Polarization Simulation with 3D HO FDTD Jing Li, Zhao-Fa Zeng,, Ling Huang, and Fengshan Liu College of Geoexploration
Solution of Linear Systems
Chapter 3 Solution of Linear Systems In this chapter we study algorithms for possibly the most commonly occurring problem in scientific computing, the solution of linear systems of equations. We start
Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems
Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems Aleksandar Donev Courant Institute, NYU 1 [email protected] 1 Course G63.2010.001 / G22.2420-001,
A matrix-free preconditioner for sparse symmetric positive definite systems and least square problems
A matrix-free preconditioner for sparse symmetric positive definite systems and least square problems Stefania Bellavia Dipartimento di Ingegneria Industriale Università degli Studi di Firenze Joint work
High-fidelity electromagnetic modeling of large multi-scale naval structures
High-fidelity electromagnetic modeling of large multi-scale naval structures F. Vipiana, M. A. Francavilla, S. Arianos, and G. Vecchi (LACE), and Politecnico di Torino 1 Outline ISMB and Antenna/EMC Lab
SOLVING LINEAR SYSTEMS
SOLVING LINEAR SYSTEMS Linear systems Ax = b occur widely in applied mathematics They occur as direct formulations of real world problems; but more often, they occur as a part of the numerical analysis
AN INTERFACE STRIP PRECONDITIONER FOR DOMAIN DECOMPOSITION METHODS
AN INTERFACE STRIP PRECONDITIONER FOR DOMAIN DECOMPOSITION METHODS by M. Storti, L. Dalcín, R. Paz Centro Internacional de Métodos Numéricos en Ingeniería - CIMEC INTEC, (CONICET-UNL), Santa Fe, Argentina
General Framework for an Iterative Solution of Ax b. Jacobi s Method
2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,
TESLA Report 2003-03
TESLA Report 23-3 A multigrid based 3D space-charge routine in the tracking code GPT Gisela Pöplau, Ursula van Rienen, Marieke de Loos and Bas van der Geer Institute of General Electrical Engineering,
Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms
Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms Björn Rocker Hamburg, June 17th 2010 Engineering Mathematics and Computing Lab (EMCL) KIT University of the State
7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
Notes on Cholesky Factorization
Notes on Cholesky Factorization Robert A. van de Geijn Department of Computer Science Institute for Computational Engineering and Sciences The University of Texas at Austin Austin, TX 78712 [email protected]
Scalable Distributed Schur Complement Solvers for Internal and External Flow Computations on Many-Core Architectures
Scalable Distributed Schur Complement Solvers for Internal and External Flow Computations on Many-Core Architectures Dr.-Ing. Achim Basermann, Dr. Hans-Peter Kersken, Melven Zöllner** German Aerospace
The Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix
7. LU factorization EE103 (Fall 2011-12) factor-solve method LU factorization solving Ax = b with A nonsingular the inverse of a nonsingular matrix LU factorization algorithm effect of rounding error sparse
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
Lecture 1: Systems of Linear Equations
MTH Elementary Matrix Algebra Professor Chao Huang Department of Mathematics and Statistics Wright State University Lecture 1 Systems of Linear Equations ² Systems of two linear equations with two variables
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
1 Determinants and the Solvability of Linear Systems
1 Determinants and the Solvability of Linear Systems In the last section we learned how to use Gaussian elimination to solve linear systems of n equations in n unknowns The section completely side-stepped
Optimal Scheduling for Dependent Details Processing Using MS Excel Solver
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 8, No 2 Sofia 2008 Optimal Scheduling for Dependent Details Processing Using MS Excel Solver Daniela Borissova Institute of
Optimization of a Broadband Discone Antenna Design and Platform Installed Radiation Patterns Using a GPU-Accelerated Savant/WIPL-D Hybrid Approach
Optimization of a Broadband Discone Antenna Design and Platform Installed Radiation Patterns Using a GPU-Accelerated Savant/WIPL-D Hybrid Approach Tod Courtney 1, Matthew C. Miller 1, John E. Stone 2,
MAT188H1S Lec0101 Burbulla
Winter 206 Linear Transformations A linear transformation T : R m R n is a function that takes vectors in R m to vectors in R n such that and T (u + v) T (u) + T (v) T (k v) k T (v), for all vectors u
NOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
Multigrid preconditioning for nonlinear (degenerate) parabolic equations with application to monument degradation
Multigrid preconditioning for nonlinear (degenerate) parabolic equations with application to monument degradation M. Donatelli 1 M. Semplice S. Serra-Capizzano 1 1 Department of Science and High Technology
MINIMUM USAGE OF FERRITE TILES IN ANECHOIC CHAMBERS
Progress In Electromagnetics Research B, Vol. 19, 367 383, 2010 MINIMUM USAGE OF FERRITE TILES IN ANECHOIC CHAMBERS S. M. J. Razavi, M. Khalaj-Amirhosseini, and A. Cheldavi College of Electrical Engineering
A PRACTICAL MINIATURIZED U-SLOT PATCH ANTENNA WITH ENHANCED BANDWIDTH
Progress In Electromagnetics Research B, Vol. 3, 47 62, 2008 A PRACTICAL MINIATURIZED U-SLOT PATCH ANTENNA WITH ENHANCED BANDWIDTH G. F. Khodaei, J. Nourinia, and C. Ghobadi Electrical Engineering Department
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms
by the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that
Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
Derivative Free Optimization
Department of Mathematics Derivative Free Optimization M.J.D. Powell LiTH-MAT-R--2014/02--SE Department of Mathematics Linköping University S-581 83 Linköping, Sweden. Three lectures 1 on Derivative Free
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
THE Walsh Hadamard transform (WHT) and discrete
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I: REGULAR PAPERS, VOL. 54, NO. 12, DECEMBER 2007 2741 Fast Block Center Weighted Hadamard Transform Moon Ho Lee, Senior Member, IEEE, Xiao-Dong Zhang Abstract
State of Stress at Point
State of Stress at Point Einstein Notation The basic idea of Einstein notation is that a covector and a vector can form a scalar: This is typically written as an explicit sum: According to this convention,
P164 Tomographic Velocity Model Building Using Iterative Eigendecomposition
P164 Tomographic Velocity Model Building Using Iterative Eigendecomposition K. Osypov* (WesternGeco), D. Nichols (WesternGeco), M. Woodward (WesternGeco) & C.E. Yarman (WesternGeco) SUMMARY Tomographic
An Additive Neumann-Neumann Method for Mortar Finite Element for 4th Order Problems
An Additive eumann-eumann Method for Mortar Finite Element for 4th Order Problems Leszek Marcinkowski Department of Mathematics, University of Warsaw, Banacha 2, 02-097 Warszawa, Poland, [email protected]
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
Fast Iterative Solvers for Integral Equation Based Techniques in Electromagnetics
Fast Iterative Solvers for Integral Equation Based Techniques in Electromagnetics Mario Echeverri, PhD. Student (2 nd year, presently doing a research period abroad) ID:30360 Tutor: Prof. Francesca Vipiana,
Preconditioning Sparse Matrices for Computing Eigenvalues and Solving Linear Systems of Equations. Tzu-Yi Chen
Preconditioning Sparse Matrices for Computing Eigenvalues and Solving Linear Systems of Equations by Tzu-Yi Chen B.S. (Massachusetts Institute of Technology) 1995 B.S. (Massachusetts Institute of Technology)
Direct Methods for Solving Linear Systems. Matrix Factorization
Direct Methods for Solving Linear Systems Matrix Factorization Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011
Matrix Multiplication
Matrix Multiplication CPS343 Parallel and High Performance Computing Spring 2016 CPS343 (Parallel and HPC) Matrix Multiplication Spring 2016 1 / 32 Outline 1 Matrix operations Importance Dense and sparse
Systems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
HSL and its out-of-core solver
HSL and its out-of-core solver Jennifer A. Scott [email protected] Prague November 2006 p. 1/37 Sparse systems Problem: we wish to solve where A is Ax = b LARGE Informal definition: A is sparse if many
Elasticity Theory Basics
G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold
A Direct Numerical Method for Observability Analysis
IEEE TRANSACTIONS ON POWER SYSTEMS, VOL 15, NO 2, MAY 2000 625 A Direct Numerical Method for Observability Analysis Bei Gou and Ali Abur, Senior Member, IEEE Abstract This paper presents an algebraic method
On the representability of the bi-uniform matroid
On the representability of the bi-uniform matroid Simeon Ball, Carles Padró, Zsuzsa Weiner and Chaoping Xing August 3, 2012 Abstract Every bi-uniform matroid is representable over all sufficiently large
Computational Optical Imaging - Optique Numerique. -- Deconvolution --
Computational Optical Imaging - Optique Numerique -- Deconvolution -- Winter 2014 Ivo Ihrke Deconvolution Ivo Ihrke Outline Deconvolution Theory example 1D deconvolution Fourier method Algebraic method
Lecture L3 - Vectors, Matrices and Coordinate Transformations
S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between
1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
National Laboratory of Antennas and Microwave Technology Xidian University Xi an, Shaanxi 710071, China
Progress In Electromagnetics Research, PIER 76, 237 242, 2007 A BROADBAND CPW-FED T-SHAPE SLOT ANTENNA J.-J. Jiao, G. Zhao, F.-S. Zhang, H.-W. Yuan, and Y.-C. Jiao National Laboratory of Antennas and Microwave
Design and Electromagnetic Modeling of E-Plane Sectoral Horn Antenna For Ultra Wide Band Applications On WR-137 & WR- 62 Waveguides
International Journal of Engineering Science Invention ISSN (Online): 2319 6734, ISSN (Print): 2319 6726 Volume 3 Issue 7ǁ July 2014 ǁ PP.11-17 Design and Electromagnetic Modeling of E-Plane Sectoral Horn
AMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: SVD revisited; Software for Linear Algebra Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 9 Outline 1 Computing
By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal
1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
ECE 842 Report Implementation of Elliptic Curve Cryptography
ECE 842 Report Implementation of Elliptic Curve Cryptography Wei-Yang Lin December 15, 2004 Abstract The aim of this report is to illustrate the issues in implementing a practical elliptic curve cryptographic
A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form
Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...
Calculation of Eigenfields for the European XFEL Cavities
Calculation of Eigenfields for the European XFEL Cavities Wolfgang Ackermann, Erion Gjonaj, Wolfgang F. O. Müller, Thomas Weiland Institut Theorie Elektromagnetischer Felder, TU Darmstadt Status Meeting
A note on fast approximate minimum degree orderings for symmetric matrices with some dense rows
NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. (2009) Published online in Wiley InterScience (www.interscience.wiley.com)..647 A note on fast approximate minimum degree orderings
Dynamic Eigenvalues for Scalar Linear Time-Varying Systems
Dynamic Eigenvalues for Scalar Linear Time-Varying Systems P. van der Kloet and F.L. Neerhoff Department of Electrical Engineering Delft University of Technology Mekelweg 4 2628 CD Delft The Netherlands
SIW 2D PLANAR ARRAY WITH FOUR CROSS SLOTS RADIATOR AND TUNING VIAS
Progress In Electromagnetics Research C, Vol. 40, 83 92, 2013 SIW 2D PLANAR ARRAY WITH FOUR CROSS SLOTS RADIATOR AND TUNING VIAS P. Sanchez-Olivares, J. L. Masa-Campos *, J. A. Ruiz-Cruz, and B. Taha-Ahmed
Section 5.3. Section 5.3. u m ] l jj. = l jj u j + + l mj u m. v j = [ u 1 u j. l mj
Section 5. l j v j = [ u u j u m ] l jj = l jj u j + + l mj u m. l mj Section 5. 5.. Not orthogonal, the column vectors fail to be perpendicular to each other. 5..2 his matrix is orthogonal. Check that
Model Order Reduction of Parameterized Interconnect Networks via a Two-Directional
Model Order Reduction of Parameterized Interconnect Networks via a Two-Directional Arnoldi Process Yung-Ta Li joint work with Zhaojun Bai, Yangfeng Su, Xuan Zeng RANMEP, Jan 5, 2008 1 Problem statement
P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE
1 P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE JEAN-MARC GRATIEN, JEAN-FRANÇOIS MAGRAS, PHILIPPE QUANDALLE, OLIVIER RICOIS 1&4, av. Bois-Préau. 92852 Rueil Malmaison Cedex. France
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
Matrix Differentiation
1 Introduction Matrix Differentiation ( and some other stuff ) Randal J. Barnes Department of Civil Engineering, University of Minnesota Minneapolis, Minnesota, USA Throughout this presentation I have
Inner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar
Math 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster
Acta Technica Jaurinensis Vol. 3. No. 1. 010 A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster G. Molnárka, N. Varjasi Széchenyi István University Győr, Hungary, H-906
Distributed block independent set algorithms and parallel multilevel ILU preconditioners
J Parallel Distrib Comput 65 (2005) 331 346 wwwelseviercom/locate/jpdc Distributed block independent set algorithms and parallel multilevel ILU preconditioners Chi Shen, Jun Zhang, Kai Wang Department
Representing Reversible Cellular Automata with Reversible Block Cellular Automata
Discrete Mathematics and Theoretical Computer Science Proceedings AA (DM-CCG), 2001, 145 154 Representing Reversible Cellular Automata with Reversible Block Cellular Automata Jérôme Durand-Lose Laboratoire
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
Linear Algebra: Vectors
A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector
Solution to Homework 2
Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if
Lecture 3: Finding integer solutions to systems of linear equations
Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture
Examination paper for TMA4205 Numerical Linear Algebra
Department of Mathematical Sciences Examination paper for TMA4205 Numerical Linear Algebra Academic contact during examination: Markus Grasmair Phone: 97580435 Examination date: December 16, 2015 Examination
The Basics of FEA Procedure
CHAPTER 2 The Basics of FEA Procedure 2.1 Introduction This chapter discusses the spring element, especially for the purpose of introducing various concepts involved in use of the FEA technique. A spring
MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =
MAT 200, Midterm Exam Solution. (0 points total) a. (5 points) Compute the determinant of the matrix 2 2 0 A = 0 3 0 3 0 Answer: det A = 3. The most efficient way is to develop the determinant along the
a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data Jun Wang Department of Mechanical and Automation Engineering The Chinese University of Hong Kong Shatin, New Territories,
MATH 551 - APPLIED MATRIX THEORY
MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
