This chapter is devoted to several iterative methods for solving the large sparse linear systems
Iterative methods are suited for use with sparse matrices. If A is dense, your best course of action is probably to factor A and solve the equation by back substitution. The time spent factoring a dense A is roughly equivalent to the time spent solving the system iteratively; and once A is factored, the system can be back solved quickly for multiple values of b. Compare this dense matrix with a sparse matrix of larger size that fills the same amount of memory. The triangular factors of a sparse A usually have many more nonzero elements than A itself. Factoring may be impossible due to limited memory, and will be time-consuming as well; even the backsolving step may be slower than iterative solution. On the other hand, most iterative methods are memory-efficient and run quickly with sparse matrices.
Except when the matrix A in Eq.\eqref{EqPart4.1} has very special structure and fast direct methods apply, iterative methods are usually the method of choice for large sparse linear systems. All iterative methods can be divided into two categories: stationary methods and nonstationary iterative methods. Since the linear matrix/vector equation \eqref{EqPart4.1} is not suitable for stationary iterative schemes, it requires preconditioning operation that transforms the given problem into fixed point problem
Stationary Iterative Methods
Theorem 1: Given vector b and splitting A = S + (A − S) of matrix A, with A and S nonsingular, the iteration scheme (4) converges if and only if the spectral radius \[ \rho \left( {\bf S}^{-1} {\bf M} \right) < 1 . \] Here ρ(X) is the largest modulus of eigenvalues of matrix X.
Theorem 2: If H is a square matrix, then Hkx → 0 as k → ∞ for every vector x ∈ ℂn×1 if and only if each eigenvalue λ of H is less than one in modulus.
Then we have only to prove the theorem for an arbitrary block of the form above.
Every Jordan block can be decomposed into a sum of the diagonal matrix λiI and the nilpotent matrix C:
\[
{\bf J}_i = \lambda_i {\bf I} + {\bf C} \in \mathbb{R}^{m \times m} ,
\]
where m = mi depends on index i, which we drop for simplicity. Matrix C is independent of th eigenvalue and depends only on the dimension of eigenspace; it consists of one on the superdiagonal and zeroes elsewhere:
\[
{\bf C}_i =
\begin{bmatrix} 0& 1 & 0 & \cdots & 0 \\
0&0 & 1 & \cdots & 0 \\
\vdots & \vdots & \vdots & \ddots & \vdots \\
0&0&0& \cdots & 0 \end{bmatrix} \in \mathbb{R}^{m \times m} .
\]
It can be shown that power matrices Ck (k = 1, 2, …) vanish when power k exceeds the dimension of the Jordan block. You can verify with Mathematica that every time you multiply by C, superdiagonal of ones is shifted up.
The power of nilpotent matrix Cn is a matrix whose only nonzero elements are "1" in the upper superdiagonal in the upper right corner. Moreover, the power matrix Cm-1 (last term in Eq.(E2.2)) has a single nonzero entry of "1" in the upper right corner.
If |λi| < 1, then every term \( \displaystyle \quad \binom{k}{n} \lambda^{k-n}_i {\bf C}^n {\bf v} \quad \) in Eq.(E2.2) turns to zero as k → ∞ because exponential functions decreases faster than any polynomial.
If |λi| > 1, then every term in Eq.(E2.2) blows up for any vector v.
If |λi| = 1 and λi ≠ 1, then every term in Eq.(E2.2) oscillates, so the Jordan block Jiv diverges for any v. If u is an eigenvector corresponding to λi with |λi| = 1, then Jiu = λiv, so Ciu = 0. Expanding arbitrary vector v over eigenvectors (including generalized eigenvectors), we see that at least one part of it containing u oscillates. Therefore, the corresponding Jordan block does not converge.
If λi = 1 and m = 1, then the Jordan block becomes one-dimensional and it consists of one diagonal entry of 1. Then if v is an eigenvector corresponding to λi = 1, then Jiv = v.
Corollary 1: Power matrix Hkv converges to some limit vector as k → ∞ jf and only if all eigenvalues of matrix H, |λi| < 1, or matrix H has simple eigenvalue λ = 1 for each Jordan b;ock.
Following Householder (1958), the number ρ = max |λi| is called the spectral radius of matrix H
- Brigs, W.L., Multigrid Tutorial,
- Burden, R.L. and Faires, J.D., Numerical Analysis, ninth edition, Brows/Cole, Cengage, Boston, MA,
- Darve, E. and Wootters, M., Numerical Linear Algebra with Julia, SIAM, Philadelphia, 2021.
- Forsythe, G.E. and Wasow, W.R., Finite Difference Methods for Partial Differential Equations. New York: John Wiley & Sons, Inc., 1960.
- Loehr, N., Advanced Linear Algebra, CRC Press, 2014.
- Watkins, D.S., Fundamentals of Matrix Computations, Wiley; 3rd edition, 2010.
- White, R.E., Computational Linear Algebra with Applications and MATLAB Computations, CRC Press Boca Raton, FL, 2023. doi: 10.1201/9781003304128