Applied Mathematics
Vol.07 No.11(2016), Article ID:68223,16 pages
10.4236/am.2016.711108
A Class of Generalized Approximate Inverse Solvers for Unsymmetric Linear Systems of Irregular Structure Based on Adaptive Algorithmic Modelling for Solving Complex Computational Problems in Three Space Dimensions
Anastasia-Dimitra Lipitakis
Department of Informatics and Telematics, Harokopio University, Athens, Greece
Copyright © 2016 by author and Scientific Research Publishing Inc.
This work is licensed under the Creative Commons Attribution International License (CC BY).
Received 18 May 2016; accepted 9 July 2016; published 12 July 2016
ABSTRACT
A class of general inverse matrix techniques based on adaptive algorithmic modelling methodologies is derived yielding iterative methods for solving unsymmetric linear systems of irregular structure arising in complex computational problems in three space dimensions. The proposed class of approximate inverse is chosen as the basis to yield systems on which classic and preconditioned iterative methods are explicitly applied. Optimized versions of the proposed approximate inverse are presented using special storage (k-sweep) techniques leading to economical forms of the approximate inverses. Application of the adaptive algorithmic methodologies on a characteristic nonlinear boundary value problem is discussed and numerical results are given.
Keywords:
Adaptive Algorithms, Algorithmic Modelling, Approximate Inverse, Incomplete LU Factorization, Approximate Decomposition, Unsymmetric Linear Systems, Preconditioned Iterative Methods, Systems of Irregular Structure
1. Introduction
In recent years, extensive research work has been focused on the computation of exact and approximate inverse matrices for solving efficiently complex computational problems particularly on parallel computer systems [1] - [20] . In this article, a new class of Sparse Approximate Inverses matrices based on adaptive algorithmic modelling methods is presented. These adaptive algorithmic solution methods can be used for solving large sparse linear finite difference (FD) and finite element (FE) systems of irregular structures derived mainly from the discretization of parabolic and elliptic PDE’s in both two and three space dimensions [21] - [31] .
Let us consider a class of boundary value problems defined by the equation
(1.1)
subject to the general boundary conditions
(1.1a)
where D is a closed bounded domain in RN, with N ≤ 3 and ∂D the boundary of D, a predetermined singular perturbation parameter,; and the coefficients are continuous and differentiable functions in D,; f are sufficiently smooth functions on D and is the direction of the outward normal derivative.
The discrete analogue of Equation (1.1) leads to the solution of the general linear system
, (1.2)
where the coefficient matrix A is a large sparse real (n ´ n) matrix of irregular structure. The structure of A is shown in the following Figure 1.
For solving the system (1.2), there is a choice between direct and iterative, assuming that there are no barriers due to memory requirements for the former or excessive runtimes (e.g. time dependent problems) for the latter. Note that for generality purposes the coefficient matrix is assumed to be unsymmetric (case occurring in the discretization of flow equations that arise in certain Hydrology studies [32] and of irregular non-zero structure (case resulting from the triangulation of irregular or regular domains into irregular elements in pipeline networks [33] . Algorithmic solution methods for the linear systems (1.2) applicable to both two and three space dimensions can be applied [22] , where the unsymmetric coefficient matrix, in which all the off-centre terms are grouped into regular bands, can be factorized exactly to yield direct algorithmic procedures for the FD or FE solution.
It should be noted that in the case of very large sparse linear and nonlinear systems with coefficients of irregular structure, the memory requirements and the corresponding computational work are prohibitively high and the use of exact inverse solvers is usually not recommended. In such cases, preconditioned iterative techniques for solving numerically the FD or FE linear systems (1.2) can be used by deriving semi-direct solution methods following the principle [27] [34] that implicit procedures based on approximately decomposing discrete operators into easily invertible factors facilitating the solution of (1.2). Sparse factorization procedures yield efficient
Figure 1. The structure of coefficient matrix A.
procedures for the FE or FD solution by manipulating the problem of the fill-in terms, which occur during the factorization [22] [35] . Note that simple compact storage schemes for the considered data can be used and preconditioned algorithmic solvers do not require any searching operations. An important feature of the proposed adaptive algorithmic methods is the provision of a class of iterative methods for solving large sparse unsymmetric systems of irregular non-zero structure, with additional computational facilities, i.e. the choice of fill-in parameters, rejection parameters, entropy-adaptivity-uncertainty (EAU) parameters [36] , by which the best method for a given problem can be selected. The proposed methods have a universal scope of application for numerically solving of elliptic and parabolic boundary value problems by either FD or FE discretization methods in both two and three space dimensions with the only restriction being that the coefficient matrix should be diagonally dominant.
2. Approximate LU Decomposition and Approximate Inverse Methods
The approximate factorization techniques and approximate inverse methodologies have been widely used for solving a large class of linear and nonlinear systems resulting in complex computational problems [12] [29] [31] [37] - [49] . The LU factorization of a given matrix is characterized as a high level algebraic description of Gaussian elimination and by expressing the outcomes of matrix algorithms in the language of matrix factorizations facilitates generalization and certain connections between algorithms that may appear different at scalar level [47] . The solution of linear system can be computed by a two-step triangular solving process, i.e.
(2.1)
For solving the symmetric problem, a variant of the LU factorization in which A is decomposed into three-matrix product, i.e., where D is diagonal and L, M are lower unit lower triangular. In this case the solution can be obtained in flops by solving (forward elimination), and (by substitution). Note that if then L = M and the computational work for the factorization is half of that required by Gaussian elimination. The factorization takes the form and can be used for solving symmetric problems, as well as the four-matrix product, i.e., where is the transpose of T (Varga, 1962). In the case of symmetric positive definite systems the factorization exists and is computationally stable. The factorization is known as Cholesky factorization and by solving the triangular system and, then. Note that the Gaxpy Cholesky version requires flops, where Gaxpy is a BLAS level-2 routine defined algebraically as requiring operations.
In the general case, such as the case of three space dimensions and the finite element discretization, the coefficient matrix has an irregular structure of the nonzero elements, where the non-diagonal elements can be grouped in regular bands of width and (width parameters) in distances m and p (semi-bandwidths) respectively.
The linear system (1.2) can be solved by direct (explicitly) or iterative (implicitly) methods depending on the availability of memory requirements. Several factorization/decomposition techniques can be used for facilitating the numerical solution of linear system (1.2), i.e. two, three, four term factorization schemes of the coefficient matrix A. Following an explicit solution of system (1.2) this system can equivalently written as
, (2.2)
where M is the inverse matrix of A, i.e.. Since the computation of the exact inverse is a difficult computational problem particularly in the case of complex 3D problems, the approximate inverse matrix approach can be alternatively used.
Let us consider an approximate factorization of the coefficient matrix A,
(2.3)
where and, are lower and upper sparse triangular matrices of irregular structures of semi-bandwidths m and p retaining and fill-in terms respectively. The decomposition factors and are banded matrices with l1 and l2 the numbers of diagonals retained in semi-bandwidths m and p respectively (Figure 2 and Figure 3), of the following form.
The computation of the elements of the sparse decomposition factors has been presented in [23] [24] [27] .
Figure 2. The structure of lower decomposition factor.
Figure 3. The structure of upper decomposition factor.
The relationships of the elements of V matrix and the corresponding conventional (for and) is shown in the following Figure 4.
An analogue scheme can be obtained for matrix W, while the relationships of the elements of H matrix and the corresponding conventional (for and) is shown in the following Figure 5 (the same holds for the matrix F).
The (near) optimum values of fill-in parameters are mainly depended on the nature of the problem and structure of the coefficient matrix A [22] [47] - [49] .
3. Generalized Approximate Inverse Solvers for Unsymmetric Linear Systems of Irregular Structure
3.1. Introduction
An exact inverse algorithm based on adaptive algorithmic methodologies for solving linear unsymmetric systems of irregular structure arising in FD/FE discretization of boundary-value problems in three space dimensions has been recently presented [36] . This algorithm computes the elements of an exact inverse of a given unsymmetric matrix of irregular structure using an exact LU factorization [22] . The computational work of the EBAIM-1 algorithm is multiplications, while the memory requirements are (n ´ n) words. In the case of very large systems the memory requirements could be prohibitively high and the usage of approximate inverse iterative techniques is desirable.
It should be also noted that in the case that only and fill-in terms are retained in semi-bandwidths m and p respectively, then a class of approximate inverse matrix algorithms for solving large sparse unsymmetric linear systems of irregular structure [50] arising in the FD/FE discretization of elliptic and
Figure 4. The relationship of matrix V in its banded stored form and the corres- ponding one in Figure 1.
Figure 5. The relationship of matrix H in its stored form and the corres- ponding one in Figure 3.
parabolic boundary values, can be obtained. Such algorithms are described in the next sections.
A class of optimized approximate inverse variants can be obtained by considering a (near) optimized choice of the approximate inverse M depends on the selection of related parameters, i.e. fill-in parameters r1, r2, retention parameters δl1, δl2 and entropy-adaptivity-uncertainty (EAU) parameters [36] [51] . Note that the selection of retention parameter values as multiples of the corresponding semi-bandwidths of the original matrix leads to improved numerical results [22] . Then, the following sub-classes of approximate inverses, depending on the accuracy, storage and computational work requirements, can be derived as indicated in the following Figure 6.
where of sub-class I is a banded form of the exact inverse retaining elements along each row and column respectively, while its elements are equal to the corresponding elements of the exact inverse. The term of sub-class II is a banded form of M, retaining only elements along each row and column during the computational procedure of the approximate inverse and under certain hypotheses can be considered as a good approximation of the original inverse, while the entries of the approximate inverse in sub-class III have been retained after computing M* and are less accurate than the corresponding entries of. Finally, in sub-class IV the elements of the approximate inverse can be computed.
3.2. Approximate Inverse Algorithmic Methodologies
Algorithmic solution methods for the linear systems (1.2) applicable to both two and three space dimension can be applied [23] , where the unsymmetric coefficient matrix, in which all the off-centre terms are grouped into regular bands, can be factorized exactly to yield direct algorithmic procedures for the FD or FE solution. Alternatively, preconditioned iterative techniques for solving numerically the FD or FE linear systems (1.2) can be used by deriving semi-direct solution methods following the principle [28] that implicit procedures based on approximately decomposing discrete operators into easily invertible factors facilitating the solution of (1.2). Sparse factorization procedures yield efficient procedures for the FE or FD solution by manipulating the problem of the fill-in terms, which occur during the factorization. Note that simple compact storage schemes for the considered data can be used and preconditioned algorithmic solvers do not require any searching operations. An important feature of the proposed adaptive algorithmic methods is the provision of both direct and iterative methods for solving large sparse unsymmetric systems of irregular non-zero structure, with additional computational facilities, i.e. the choice of fill-in parameters, rejection parameters, entropy-adaptivity-uncertainty (EAU) parameters [36] [51] , by which the best method for a given problem can be selected. The proposed methods have a universal
Figure 6. Subclasses of approximate inverses.
scope of application for numerically solving of elliptic and parabolic boundary value problems by either FD or FE discretization methods in both two and three space dimensions with the only restriction being that the coefficient matrix be diagonally dominant.
Let us assume that, a non-singular matrix, is an approximate inverse of A, i.e. ,. Note that if and non-zero elements have been retained in the corresponding decomposition factors, then, where M is the exact inverse of A. The elements of M can be determined by solving recursively the systems
, (3.1)
having main disadvantages, i.e. high storage requirements and computational work involved particularly in the case of solving very large unsymmetric linear systems. A class of approximate inverses can be obtained by retaining only and diagonals in the lower and upper triangular parts of inverse respectively, the remaining elements being just not computed at all. Optimized forms of this algorithm are particularly effective for solving banded sparse FE systems of very large order, i.e. or in the case of narrow-banded sparse FE systems of very large order, i.e..
Let us consider now the approximate inverse of A with the form
(3.2)
where are the fill-in parameters, i.e. the number of outermost off-diagonal entries retained in semi-band- widths m and p respectively.
Then, by post-multiplying Equation (3.2) by and pre-multiplying the same equation by, we obtain
(3.3)
where. Note that in the 2D symmetric case, i.e., , where r is the fill-in parameter, i.e. the number of outer-most off diagonal entries retained in semi-bandwidth of the tridiagonal factor Lr., by considering the equations in the analytical form for i-row with and the j-column with respectively we can obtain
(3.4)
where is the Kronecker delta [21] [27] [52] .
The elements of the approximate inverse for can be determined successively as (i.e. elements of the n-th row of the inverse) and for we obtain (i.e. the n-th column of the inverse). Proceeding in a similar manner we can explicitly determine for and respectively the remaining elements. In the following Figure 7 the form of an (8 × 8) approximate inverse matrix is indicatively demonstrated.
where is a (8 × 8) banded approximate inverse matrix with retention parameters and. Note that for simplicity reasons the case is considered.
3.3. Optimized Approximate Inverse Matrices and Storage Techniques
Let us consider the exact inverse M of the original coefficient matrix A in equation (2.1). Note that the computation of the inverse is indicated in the following characteristic diagram (Figure 8).
It should be noted that the diagonal elements (in bold) are firstly computed (starting from the last element of the inverse, i.e.) and then computing upwards/column-wise and from right to left/row-wise. Note also that
Figure 7. The structure of an (8 × 8) banded approximate inverse.
Figure 8. Computing the elements of the approximate inverse M* of sub-class IV following the KS technique (K = 2)*. (*) Note that this computational technique will be referred as double-sweep (DS) technique.
the last diagonal element is computed and then all the elements of n-row (from right to left) and n-column (upwards). Then, only the diagonal element is computed, and next the diagonal element is computed and all the elements of (n-2)-row (from right to left) and the elements of (n-2)-column (upwards). Continuing in this way the rest elements of this approximate inverse are computed. It should be noted that the computational work of the resulting inverse M* of sub-class IV, is almost the half of that required by the approximate inverse of sub-class III. Note diagrammatically, as it is shown in Figure 8, only the underlined elements of the approximate inverse M* are computed.
By generalized this storage saving computational technique, we consider the above DS technique can be replaced by k-sweep (KS) technique, i.e. after the computation of the last diagonal element, all the elements of n-row (from right to left) and n-column (upwards). Then, only the diagonal elements are computed, and next the diagonal element is computed and all the elements of (n-k-2)-row (from right to left) and the elements of (n-k-2)-column (upwards). Continuing in this way the rest elements of this approximate inverse are computed. It should be noted that the computational work of the resulting inverse M* of this sub-class by using the KS-storage technique is considerably smaller than that required by the approximate inverse resulting from the application of the DS storage technique. In the case of k = 2 the KS-storage technique reduces to the example shown in Figure 4.
An optimized explicit banded approximate inverse by minimizing the memory requirements of EBAIM-1 algorithm
In order to minimize the memory requirements of EBAIM-1 algorithm, which in particular in the case of very large matrices of irregular structure can be prohibitively high, we consider the inverse M of equation (5.4) revolving its elements by 180˚ about the anti-diagonal removing the diagonal and the (δl-1) super diagonals in the first δl columns, while the rest δl sub-diagonals in the rest δl columns, then results the following form of the inverse (Figure 9).
4. The Optimized Approximate Inverse Algorithm
The application of this storage scheme on the approximate inverse leads to the following optimized approximate inverse algorithm. Note that the computation of the approximate inverse algorithm pre-assumes the approximate factorization of the coefficient matrix A, i.e., where and are the lower and upper
Figure 9. Transformed forms of optimized approximate inverse M* (in banded storage).
sparse triangular decomposition factors [21] [23] - [25] [49] .
Algorithm OBAIM-1 (a, b, c, n, F, H, g, Γ, Ζ, ω, β, r1, r2, m, p, l1, l2, δl, M*)
Purpose: This algorithm computes the elements of the approximate inverse of a given real (n ´ n) matrix of irregular structure
Input: diagonal elements a of matrix A; superdiagonal elements b, subdiagonal elements c, n order of A; submatrices F, H, of upper triadiagonal decomposition factor U, superdiagonal elements g of L; submatrices Γ, Z, of lower tridiagonal matrix L; diagonal elements ω of L; subdiagonal elements β of L; fill-in parameters r1, r2; semi-bandwidths m, p; l1 and l2 numbers of diagonals retained in semi-bandwidths m and p respectively, and are the numbers of diagonal retained in approximate inverse M*/for simplicity reasons is chosen/
Output: elements of the approximate inverse
Computational Procedure:
step 1: let;;;;;; ;;;
step 2: for
step 3: for
step 4: if then
step 5: if then
step 6: if then
step 7:
step 8: else
step 9:
step 10:
step 11: else
step 12:
Step 13: else
if and then
step 14: if then
step 15:
step 16:
step 17:
step 18:
step 19: else
if and then
step 20: if then
step 21:
step 22:
step 23:
step 24:
step 25:
step 26:
step 27: else
step 28:
step 29:
step 30:
step 31:
step 32:
step 33: else
if then
step 34: if then
step 35:
step 36:
step 37:
step 38:
step 39:
step 40:
step 41: else
step 42:
step 43:
step 44:
step 45:
step 46:
step 47:
step 48:
step 49:
step 50:
step 51: else
step 52:
step 53:
step 54:
step 55:
step 56:
step 57:
step 58:
step 59:
step 60:
step 61: for
step 62:
step 63: form the approximate inverse matrix
The subroutine mw (n, δl, s, q, x, y) performs the transformation in the indexes of the explicit approximate inverse matrix from its banded form to the optimized form. This routine has the following form:
Subroutine mw (n, δl, s, q, x, y)
If
Then
else
The computational work of the optimized OBAIM-1 algorithm is multiplica-
tions, while the memory requirements have been reduced down to words. It should be also noted that a class of approximate inverse matrix can be considered containing several sub-classes of approximate inverses according to memory requirements, computational work, accuracy, as indicated in the diagrammatic schemes (Figure 3 and Figure 7).
5. Explicit Adaptive Iterative Methods
A class of Adaptive Iterative Schemes for solving large sparse linear systems includes the following adaptive preconditioned iterative methods:
, (Explicit preconditioned simultaneous displacement) (5.1)
, (Explicit Preconditioned first order Richardson) (5.2)
(Explicit Preconditioned second order Richardson) (5.3)
(Explicit Preconditioned Chebyshev) (5.4)
where, α and β are predetermined acceleration parameters, and are sequences of preconditioned acceleration parameters and, i ≥ 0.
5.1. The Explicit Preconditioned Iterative Method
During the last decades extensive research work has been focused in the preconditioned approach and preconditioned iterative methods for solving large linear and nonlinear problems in sequential and parallel environments [5] [8] [11] [12] [14] [26] [29] [30] [38] [39] [41] [47] [53] - [58] . A predominant role in the usage of the preconditioned iterative schemes possess the explicit preconditioned Conjugate Gradient (EPCG) method and its variants using the sparse approximate inverse M* due to its superior convergence rate for solving very large complex computational problems [47] . A characteristic explicit solver of this sub-class is the Explicit Preconditioned Generalized Conjugate Gradient (EPGCG) method [58] . This basic EPCG method can be expressed in the following compact form:
Algorithm EPCG-1 (A, n, s, u0, u, r)
Purpose: This algorithm computes the solution vector of the linear system by using the explicit preconditioned generalized Conjugate Gradient method.
Input: A given matrix, n order of A, s known rhs vector, u0 initial guess
Output: solution vector u, residual r
Computational Procedure:
Step 1: let be an arbitrary initial approximation to the solution u
Step 2: set, form and set
Step 3: for (until convergence)
compute
//compute scalar quantities as follows://
Step 4: form and set (only for)
Step 5: evaluate and compute
Step 6: compute and form
Step 7: compute and evaluate
Step 8: compute
Step 9: if there is no convergence go to step 3,
Step 10: else
print the approximate solution and corresponding residual.
Note that a good approximant M* leads obviously to an improved EPCG method. The effectiveness of the explicit preconditioned iterative methods for solving certain classes of elliptic boundary value problems on regular domains is related to the fact that the exact inverse of A (although is full) exhibits a similar fuzzy structure around the principal diagonal and m-diagonals [22] .
5.2. The Symmetric Case
In the case of symmetric coefficient matrix by using the four-matrix decomposition [27] [41] the corresponding inverse subclasses can be enlarged as follows in Figure 10.
where 1) the elements of exact inverse of subclass I are obtained after the exact decomposition of M+, with excessive memory and computational requirements, 2) the elements of the inverse of subclass II have been computed after the application of the exact inverse algorithm, while only and diagonals have been retained, 3) the elements of inverse MS2 of subclass III have been computed from the approximate inverse, while the exact decomposition has been applied, 4) the elements of the inverse of subclass IV have been computed from the approximate factorization and the banded approximate inverse algorithm has been used for computing the elements of the inverse, 5) the elements of inverse of subclass V have been retained only on the diagonal elements of the inverse, i.e., that is, leading to a fast algorithm for computing of approximate inverse.
Note that the largest elements of inverse matrix are mainly gathered around the main diagonal in distances and in a recurring wave like pattern (Lipitakis, 1984), where m and p are respectively the semi-bandwidths of the coefficient matrix A, and and. Based on this observation the selection of retention parameters, is recommended to be multiples of values of m and p, leading to preconditioners with better performance [22] .
An indication of the sparsity and memory requirements of optimized versions of approximate inverses is given in the following Table 1.
It should be noted that in the case that δl2 = 0 the approximate inverse algorithm is reduced to an algorithm for solving FE systems in two space dimensions of semi-bandwidth m, while if then the algorithm reduces to one for solving FD linear systems in three space dimensions of semi-bandwidths m and p [23] . In the case of δl1 = 1 and δl2 = 0, then the approximate inverse reduces to one for solving linear FD systems in two space dimensions of semi-bandwidth m [25] , while if then the approximate inverse reduces to the one for solving tridiagonal linear systems (Thomas algorithm) [59] .
Figure 10. Subclasses of inverses for the symmetric case.
Table 1. Memory and sparsity requirements of the approximate inverse matrix (n = 8000, m = 21, p = 401), where δl denotes here the retention parameters.
6. Numerical Experiments
In this section a nonlinear case study by using approximate inverse preconditioned methods are presented.
The nonlinear case
Let us consider the nonlinear elliptic PDE
(6.1)
where subject to the Dirichlet boundary conditions
(6.1a)
where ∂R is the exterior boundary of the domain R.
Equation (6.1) arises in magnetohydrodynamics (diffusion-reaction, vortex problems, electric space charge considerations) with its existence and uniqueness assured by the classical theory [32] [33] . The solution of Equation (6.1) can be obtained by the linearized Picard and quasi-linearized Newton iterative schemes as outer iterative schemes of the form:
and (6.2)
(6.3)
where δ denotes here the usual central difference operator.
The resulting large sparse nonlinear system is of the form
, (6.4)
where Ω is a block tridiagonal matrix [23] .
Then, composite iterative schemes can be used, where Picard/Newton iterations are the outer iteration, while the inner iteration can be carried out either directly by an exact algorithm or by an approximate algorithm in conjunction with an explicit iterative method (6.3). The latter method can be written as
, (6.5)
where the superscript l denotes the outer iteration index, the subscript I denotes the inner iteration and.
The outer iteration was terminated when the following criterion was satisfied
, (6.6)
while the termination criterion of the inner iteration was
(6.7)
where ε1 was taken initially as and then was decreased at each iterative step by 1/10 to 10−6, where it remained constant during the next iterative steps. Numerical experiments were carried out for nonlinear problem (6.4) with and the initial guesses when u was on the boundary 0.0, 5.0, 10.0 were chosen as 0.0, 4.0, 6.0 respectively. The performance of the composite schemes Newton-Explicit Preconditioned Simultaneous Displacement (EPSD) and Picard/Newton-EPCG are given in the following Table 2 and Table 3.
7. Conclusions
A class of exact and approximate inverse adaptive algorithmic procedures has been presented for solving numerically initial/boundary value problems. Several subclasses of optimized variants of these algorithms have been also proposed for solving economically highly nonlinear systems of irregular structure. It should be stated that the proposed explicit preconditioned iterative methods and their related variants can be efficiently used for solving large sparse nonlinear systems of irregular structure of complex computational problems and for the numerical solution of highly nonlinear initial/ boundary value problems in two and three space dimensions.
Table 2. The performance of composite iterative schemes Picard/Newton for the nonlinear system (6.4) using the EPSD method (r = 4), with n = 361, m = 20, for several values of acceleration parameter α and retention parameters δl.
Table 3. The performance of composite iterative schemes Picard/Newton for the nonlinear system (6.4) using the EPCG method, with n = 361, m = 20, for several values of retention parameters δl and fill-in parameters r.
Future research work includes the parallelization of the proposed class of exact and approximate inverse matrices of irregular structure. These adaptive exact and approximate inverse algorithmic techniques can be used for solving efficiently highly nonlinear large sparse systems arising in the numerical solution of complex computational problems in parallel computer environments.
Cite this paper
Anastasia-Dimitra Lipitakis, (2016) A Class of Generalized Approximate Inverse Solvers for Unsymmetric Linear Systems of Irregular Structure Based on Adaptive Algorithmic Modelling for Solving Complex Computational Problems in Three Space Dimensions. Applied Mathematics,07,1225-1240. doi: 10.4236/am.2016.711108
References
- 1. Guillaume, P., Huard, A. and Le Calvez, C. (2002) A Block Constant Approximate Inverse for Preconditioning Large Linear Systems. SIAM Journal on Matrix Analysis and Applications, 24, 822-851.
http://dx.doi.org/10.1137/S0895479802401515 - 2. He, B. and Teixeira, F.L. (2007) Differential Forms, Galerkin Duality, and Sparse Inverse Approximations in Finite Element Solutions of Maxwell Equations. IEEE Transactions on Antennas and Propagation, 55, 1359-1368.
http://dx.doi.org/10.1109/TAP.2007.895619 - 3. Helsing, J. (2006) Approximate Inverse Preconditioners for Some Large Dense Random Electrostatic Interaction Matrices. BIT Numerical Mathematics, 46, 307-323.
http://dx.doi.org/10.1007/s10543-006-0057-0 - 4. Cerdan, J., Faraj, T., Malla, N., Marin, J. and Mas, J. (2010) Block Approximate Inverse Preconditioners for Sparse Nonsymmetric Linear Systems. Electronic Transactions on Numerical Analysis, 37, 23-40.
- 5. Cui, X. and Havami, K. (2009) Generalized Approximate Inverse Preconditioners for Least Squares Problems. Japan Journal of Industrial and Applied Mathematics, 26, 1-14.
http://dx.doi.org/10.1007/BF03167543 - 6. Gravvanis, G.A., Filelis-Papadopoulos, C.K., Giannoutakis, K.M. and Lipitakis, E.A. (2013) A Note on Parallel Finite Difference Approximate Inverse Preconditioning on Multicore Systems Using POSIX Threads. International Journal of Computational Methods, 10, Article ID: 1350032.
http://dx.doi.org/10.1142/S0219876213500321 - 7. Giannoutakis, K., Gravvanis, G., Clayton, B., Patil, A., Enright, T. and Morrison, J. (2007).Matching High Performance Approximate Inverse Preconditioning to Architectural Platforms. The Journal of Supercomputing, 42, 145-163.
http://dx.doi.org/10.1007/s11227-007-0129-1 - 8. Gravvanis, G.A., Epitropou, V.N. and Giannoutakis, K.M. (2007) On the Performance of Parallel Approximate Inverse Preconditioning Using Java Multithreading Techniques. Applied Mathematics and Computation, 190, 255-270.
http://dx.doi.org/10.1016/j.amc.2007.01.024 - 9. Ngo, C., Yashtini, M. and Zhang, H.C. (2015) An Alternating Direction Approximate Newton Algorithm for Ill-Conditioned Inverse Problems with Application to Parallel MRI. Journal of the Operations Research Society of China, 3, 139-162.
http://dx.doi.org/10.1007/s40305-015-0078-y - 10. Wang, H., Li, E. and Li, G. (2013) A Parallel Reanalysis Method Based on Approximate Inverse Matrix for Complex Engineering Problems. Journal of Mechanical Design, 135, Article ID: 081001.
http://dx.doi.org/10.1115/1.4024368 - 11. Xu, S., Xue, W., Wang, K. and Lin, H. (2011) Generating Approximate Inverse Preconditioners for Sparse Matrices Using CUDA and GPGPU. Journal of Algorithms and Computational Technology, 5, 475-500.
http://dx.doi.org/10.1260/1748-3018.5.3.475 - 12. Alleon, G., Benzi, M. and Giraud, L. (1997) Sparse Approximate Inverse Preconditioning for Dense Linear Systems Arising in Computational Electromagnetics. Numerical Algorithms, 16.
http://dx.doi.org/10.1023/A:1019170609950 - 13. Dubois, P.F., Greenbaum, A. and Rodrigue, G.H. (1979) Approximating the Inverse of a Matrix for Use in Iterative Algorithms on Vector Processors. Computing, 22, 257-268.
http://dx.doi.org/10.1007/BF02243566 - 14. Grote, M.J. and Huckle, T. (1997) Parallel Preconditioning with Sparse Approximate Inverses. SIAM Journal on Scientific Computing, 18, 838-853.
http://dx.doi.org/10.1137/S1064827594276552 - 15. Van der Vorst, H.A. (1989) High Performance Preconditioning. SIAM Journal on Scientific Computing, 10, 1174-1185.
http://dx.doi.org/10.1137/0910071 - 16. Broker, O., Grote, J., Mayer, C. and Reusken, A. (2001) Robust Parallel Smoothing for Multigrid via Sparse Approximate Inverses. SIAM Journal on Scientific Computing, 23, 1396-1417.
http://dx.doi.org/10.1137/S1064827500380623 - 17. Chow, E. (2000) A Priori Sparsity Patterns for Parallel Sparse Approximate Inverse Preconditioners. SIAM Journal on Scientific Computing, 21, 1804-1822.
http://dx.doi.org/10.1137/S106482759833913X - 18. Chow, E. (2001) Parallel Implementation and Practical Use of Sparse Approximate Inverse Preconditioners with a Priori Sparsity Patterns. International Journal of High Performance Computing Applications, 15, 56-74.
http://dx.doi.org/10.1177/109434200101500106 - 19. Kallischko, A. (2008) Modified Sparse Approximate Inverses (MSPAI) for Parallel Preconditioning. Doctoral Dissertation, Technische Universitat Munchen Zentrum Mathematik, Germany.
- 20. Tanaka, T. and Nodera, T. (2002) Effectiveness of Approximate Inverse Preconditioning by Using the MR Algorithm on an Origin 2400. In: Topping, B.H.V., Bittnar, Z., Topping, B.H.V. and Bittnar, Z., Eds., Proceedings of the 3rd International Conference on Engineering Computational Technology, Paper 44, 115-116.
http://dx.doi.org/10.4203/ccp.76.44 - 21. Lipitakis, E.A. (1983) A Normalized Sparse Linear Equations Solver. Journal of Computational and Applied Mathematics, 9, 287-298.
http://dx.doi.org/10.1016/0377-0427(83)90021-3 - 22. Lipitakis, E.A. (1984) Generalized Extended to the Limit Sparse Factorization Techniques for Solving Unsymmetric Finite Element Systems. Computing, 32, 255-270.
http://dx.doi.org/10.1007/BF02243576 - 23. Lipitakis, E.A. (1989) Numerical Solution of 3D Boundary Value Problems by Generalized Approximate Inverse Matrix Techniques. International Journal of Computer Mathematics, 31, 69-89.
http://dx.doi.org/10.1080/00207168908803789 - 24. Lipitakis, E.A. and Evans, D.J. (1979) Normalized Factorization Procedures for Solving Self-Adjoint Elliptic PDE’s in 3D Space Dimensions. Mathematics and Computers in Simulation, 21, 189-196.
http://dx.doi.org/10.1016/0378-4754(79)90133-2 - 25. Lipitakis, E.A. and Evans, D.J. (1980) Solving Non-Linear Elliptic Difference Equations by Extendable Sparse Factorization Procedures. Computing, 24, 325-339.
http://dx.doi.org/10.1007/BF02237818 - 26. Lipitakis, E.A. and Evans, D.J. (1981) The Rate of Convergence of an Approximate Matrix Factorization Semi-Direct Method. Numerische Mathematik, 36, 237-251.
http://dx.doi.org/10.1007/BF01396653 - 27. Lipitakis, E.A. and Evans, D.J. (1984) Solving Linear FE Systems by Normalized Approximate Matrix Factorization Semi-Direct Methods. Computer Methods in Applied Mechanics and Engineering, 43, 1-19.
http://dx.doi.org/10.1016/0045-7825(84)90091-4 - 28. Lipitakis, E.A. and Evans, D.J. (1987) Explicit Semi-Direct Methods Based on Approximate Inverse Matrix Techniques for Solving BV Problems on Parallel Processors. Mathematics and Computers in Simulation, 29, 1-17.
http://dx.doi.org/10.1016/0378-4754(87)90062-0 - 29. Gravvanis, G.A. (2000) Fast Explicit Approximate Inverses for Solving Linear and Non-Linear Finite Difference Equations. International Journal of Differential Equations and Applications, 1, 451-473.
- 30. Gravvanis, G.A. and Giannoutakis, K.M. (2005) Normalized Explicit Preconditioned Methods for Solving 3D Boundary Value Problems on Uniprocessor and Distributed Systems. International Journal for Numerical Methods in Engineering, 65, 84-110.
http://dx.doi.org/10.1002/nme.1494 - 31. Gravvanis, G.A. and Gianoutakis, K.M. (2003) Normalized Explicit FE Approximate Inverses. International Journal of Differential Equations and Applications, 6, 253-267.
- 32. Pinter, G.F. and Gray, W.G. (1977) Finite Element Simulation in Surface Hydrology. Academic Press, New York.
- 33. Porsching, T.A. (1976) On the Origins and Numerical Solution of Some Sparse Non-Linear Systems. In: Bunch, J.R. and Rose, D.J., Eds., Sparse Matrix Computations, Academic Press, New York, 439.
- 34. Gravvanis, G.A. and Lipitakis, E.A. (1996) An Explicit Sparse Unsymmetric FE Solver. Communications in Numerical Methods in Engineering, 12, 21-29.
http://dx.doi.org/10.1002/(SICI)1099-0887(199601)12:1<21::AID-CNM948>3.0.CO;2-K - 35. Gravvanis, G.A. (2000) Generalized Approximate Inverse Preconditioning for Solving Non-Linear Elliptic BV Problems. International Journal of Applied Mathematics, 2, 1363-1378.
- 36. Lipitakis, A.D. and Lipitakis, E.A.E.C. (2012) On the E-Valuation of Certain E-Business Strategies on Firm Performance by Adaptive Algorithmic Modelling: An Alternative Strategic Managerial Approach. Computer Technology and Application, 3, 38-46.
- 37. Bollhofer, M. (2003) A Robust and Efficient ILU that Incorporates the Growth of the Inverse Triangular Factors. SIAM Journal on Scientific Computing, 25, 86-103.
http://dx.doi.org/10.1137/S1064827502403411 - 38. Cosgrove, J., Dias, J. and Griewank, A. (1992) Approximate Inverse Preconditioning for Sparse Linear Systems. International Journal of Computer Mathematics, 44, 91-110.
http://dx.doi.org/10.1080/00207169208804097 - 39. Evans, D.J. (1967) The Use of Preconditioning in Iterative Methods for Solving Linear Equations with Symmetric Positive Definite Matrices. IMA Journal of Applied Mathematics, 4, 295-314.
- 40. Il’lin, V.P. (1972) Iterative Incomplete Factorization Methods. World Scientific Press, Singapore.
- 41. Varga, R.S. (1962) Matrix Iterative Analysis. Prentice-Hall, Upper Saddle River.
- 42. Gravvanis, G.A. and Giannoutakis, K.M. (2003) Normalized FE Approximate Inverse Preconditioning for Solving Nonlinear BV Problems. In: Bathe, K.J., Ed., Computational Fluid and Solid Mechanics 2003, Proceedings of 2nd MIT Conference on Computational Fluid and Solid Mechanics, 2, 1958-1962.
- 43. Huckle, T. (1998) Efficient Computation of Sparse Approximate Inverses. Numerical Linear Algebra with Applications, 5, 57-71.
http://dx.doi.org/10.1002/(SICI)1099-1506(199801/02)5:1<57::AID-NLA129>3.0.CO;2-C - 44. Jia, Z. and Zhu, B. (2009) A Power Sparse Approximate Inverse Preconditioning Procedure for Large Sparse Linear Systems. Numerical Linear Algebra with Applications, 16, 259-299.
http://dx.doi.org/10.1002/nla.614 - 45. Kolotilina, L.Y. and Yeremin, A.Y. (1993) Factorized Sparse Approximate Inverse Preconditioning. SIAM Journal on Matrix Analysis and Applications, 14, 45-58.
http://dx.doi.org/10.1137/0614004 - 46. Tang, W.-P. and Wan, W.L. (2000) Sparse Approximate Inverse Smoother for Multigrid. SIAM Journal on Matrix Analysis and Applications, 21, 1236-1252.
http://dx.doi.org/10.1137/S0895479899339342 - 47. Golub, G.H. and Van Loan, C. (1989) Matrix Computations. John Hopkins University Press, Baltimore.
- 48. Evans, D.J. and Lipitakis, E.A. (1980) A Normalized Implicit Conjugate Gradient Method for the Solution of Large Sparse Systems of Linear Equations. Computer Methods in Applied Mechanics and Engineering, 23, 1-19.
http://dx.doi.org/10.1016/0045-7825(80)90075-4 - 49. Giannoutakis, C.M. (2008) Study of Advanced Computational Methods of High Performance: Preconditioned Methods and Techniques of Matrix Inversion. Doctoral Thesis, Department of Electrical and Computer Engineering, Democritus University of Thrace, Hellas.
- 50. Tewarson, R.P. (1966) On the Product Form of Inverses of Sparse Matrices. SIAM Review, 8, 336-342.
http://dx.doi.org/10.1137/1008066 - 51. Lipitakis, A.D. (2016) A Generalized Exact Inverse Solver Based on Adaptive Algorithmic Methodologies for Solving Complex Computational Problems in Three Space Dimensions. HU-DP-TR15-2016, Technical Report.
- 52. Axelsson, O. (1985) A Survey of Preconditioned Iterative Methods of Linear Systems of Algebraic Equations. BIT Numerical Mathematics, 25, 166-187.
http://dx.doi.org/10.1007/BF01934996 - 53. Saad, Y. and Van der Vorst, H.A. (2000) Iterative Solution of Linear Systems in the 20th Century. Journal of Computational and Applied Mathematics, 123, 1-33.
http://dx.doi.org/10.1016/S0377-0427(00)00412-X - 54. Chen, K. (2005) Matrix Preconditioning Techniques and Applications. Cambridge Monographs on Applied and Computational Mathematics. Cambridge University Press, Cambridge.
http://dx.doi.org/10.1017/CBO9780511543258 - 55. Axelsson, O., Carey, G. and Lindskog, G. (1985) On a Class of Preconditioned Iterative Methods for Parallel Computers. International Journal for Numerical Methods in Engineering, 27, 637-654.
http://dx.doi.org/10.1002/nme.1620270314 - 56. Benji, M. (2002) Preconditioning Techniques for Large Linear Systems: A Survey. Journal of Computational Physics, 182, 418-477.
http://dx.doi.org/10.1006/jcph.2002.7176 - 57. Evans, D.J. (1972) The Analysis and Applications of Sparse Matrix Algorithms in FE Applications. In: Whiteman, J.R., Ed., Proceedings of Brunel University Conference, Academic Press, London and New York, 427-447.
- 58. Gravvanis, G.A. (2002) Explicit Approximate Inverse Preconditioning Techniques. Archives of Computational Methods in Engineering, 9, 371-402.
http://dx.doi.org/10.1007/BF03041466 - 59. Thomas, L.H. (1949) Elliptic Problems in Linear Difference Equations over a Network. Watson Sc. Comp. Lab. Rep., Columbia University, New York.