ConceptComplete

Iterative Methods for Linear Systems

Iterative methods generate a sequence of approximations x(k)→xx^{(k)} \to x converging to the solution of Ax=bAx = b. They are essential for large sparse systems where direct methods are too expensive in memory or computation.


Classical Iterative Methods

Definition5.4Jacobi and Gauss-Seidel Methods

Write A=D−L−UA = D - L - U where DD is diagonal, −L-L is strictly lower, −U-U is strictly upper triangular. The Jacobi iteration is x(k+1)=D−1(L+U)x(k)+D−1bx^{(k+1)} = D^{-1}(L + U)x^{(k)} + D^{-1}b, i.e., xi(k+1)=1aii(bi−∑j≠iaijxj(k))x_i^{(k+1)} = \frac{1}{a_{ii}}\left(b_i - \sum_{j \neq i} a_{ij} x_j^{(k)}\right). The Gauss-Seidel iteration uses the latest values: xi(k+1)=1aii(bi−∑j<iaijxj(k+1)−∑j>iaijxj(k))x_i^{(k+1)} = \frac{1}{a_{ii}}\left(b_i - \sum_{j < i} a_{ij} x_j^{(k+1)} - \sum_{j > i} a_{ij} x_j^{(k)}\right).

Definition5.5Successive Over-Relaxation (SOR)

The SOR method with relaxation parameter ω∈(0,2)\omega \in (0, 2) is: xi(k+1)=(1−ω)xi(k)+ωaii(bi−∑j<iaijxj(k+1)−∑j>iaijxj(k))x_i^{(k+1)} = (1 - \omega) x_i^{(k)} + \frac{\omega}{a_{ii}}\left(b_i - \sum_{j < i} a_{ij} x_j^{(k+1)} - \sum_{j > i} a_{ij} x_j^{(k)}\right). For ω=1\omega = 1, SOR reduces to Gauss-Seidel. The iteration matrix is Mω=(D−ωL)−1[(1−ω)D+ωU]M_\omega = (D - \omega L)^{-1}[(1-\omega)D + \omega U]. The Kahan theorem states that SOR converges only if 0<ω<20 < \omega < 2.


Convergence Theory

RemarkSpectral Radius and Convergence

An iterative method x(k+1)=Mx(k)+cx^{(k+1)} = Mx^{(k)} + c converges for all initial guesses if and only if the spectral radius ρ(M)<1\rho(M) < 1. The asymptotic convergence rate is r=−log⁥10ρ(M)r = -\log_{10}\rho(M): roughly 1/r1/r iterations reduce the error by a factor of 10. For strictly diagonally dominant matrices (∣aii∣>∑j≠i∣aij∣|a_{ii}| > \sum_{j\neq i}|a_{ij}|), both Jacobi and Gauss-Seidel converge. For SPD matrices, Gauss-Seidel always converges.

ExampleOptimal SOR Parameter

For the discrete Laplacian on an n×nn \times n grid with mesh size h=1/(n+1)h = 1/(n+1), the Jacobi spectral radius is ρJ=cos⁥(πh)\rho_J = \cos(\pi h). The optimal SOR parameter is ω∗=21+1−ρJ2≈2−2πh\omega^* = \frac{2}{1 + \sqrt{1 - \rho_J^2}} \approx 2 - 2\pi h for small hh, giving ρ(Mω∗)=ω∗−1≈1−2πh\rho(M_{\omega^*}) = \omega^* - 1 \approx 1 - 2\pi h. This reduces the iteration count from O(1/h2)O(1/h^2) (Jacobi/Gauss-Seidel) to O(1/h)O(1/h).


Krylov Subspace Methods

Definition5.6Conjugate Gradient Method

For SPD AA, the conjugate gradient (CG) method generates iterates x(k)∈x(0)+Kk(A,r(0))x^{(k)} \in x^{(0)} + \mathcal{K}_k(A, r^{(0)}) that minimize ∄x−x(k)∄A\|x - x^{(k)}\|_A over the Krylov subspace Kk=span{r(0),Ar(0),
,Ak−1r(0)}\mathcal{K}_k = \mathrm{span}\{r^{(0)}, Ar^{(0)}, \ldots, A^{k-1}r^{(0)}\}. The algorithm requires only one matrix-vector product, two inner products, and three vector updates per iteration. CG converges in at most nn iterations in exact arithmetic, and satisfies ∄x−x(k)∄A≀2(Îș−1Îș+1)k∄x−x(0)∄A\|x - x^{(k)}\|_A \leq 2\left(\frac{\sqrt{\kappa} - 1}{\sqrt{\kappa} + 1}\right)^k \|x - x^{(0)}\|_A where Îș=λmax⁥/λmin⁥\kappa = \lambda_{\max}/\lambda_{\min} is the condition number.

RemarkPreconditioning

Preconditioning replaces Ax=bAx = b with M−1Ax=M−1bM^{-1}Ax = M^{-1}b where M≈AM \approx A is easy to invert. The preconditioned CG converges with rate governed by Îș(M−1A)â‰ȘÎș(A)\kappa(M^{-1}A) \ll \kappa(A). Common preconditioners include: incomplete Cholesky (M=L~L~TM = \tilde{L}\tilde{L}^T), algebraic multigrid (AMG), and domain decomposition methods. For well-designed preconditioners, the iteration count becomes O(1)O(1) independent of problem size.