The conjugate gradient method is an ingenious method for iterative solution of sparse linear equations. It is now a standard benchmark for parallel scientific computing. In the author's opinion, the apparent mystery of this method is largely due to the inadequate way in which it is presented in text
The solution of sparse linear equations by the conjugate gradient method
โ Scribed by A. Jennings; G. M. Malik
- Publisher
- John Wiley and Sons
- Year
- 1978
- Tongue
- English
- Weight
- 930 KB
- Volume
- 12
- Category
- Article
- ISSN
- 0029-5981
No coin nor oath required. For personal study only.
โฆ Synopsis
Abstract
The convergence properties of the conjugate gradient method are discussed in relation to relaxation methods and Chebyshev accelerated Jacobi iteration when applied to the solution of large sets of linear equations which have a sparse, symmetric and positive definite coefficient matrix. The conclusion is reached that its convergence rate is unlikely to be much worse than these methods, and may be considerably better. The conjugate gradient method may either be applied to the basic unscaled or scaled equations or alternatively to various transformed equations. Preconditioning, block elimination and partial elimination methods of transforming equations are considered, and some comparative tests given for six problems.
๐ SIMILAR VOLUMES
We introduce a new method for the solution of linear differential equations with constant coefficients. The solutions are obtained by the application of a divided differences functional to a kernel function in two variables. For homogeneous equations the kernel is the product of a polynomial, which