ConceptComplete

Orthogonal Diagonalization

The spectral theorem states that symmetric matrices can be orthogonally diagonalized. This canonical form provides optimal coordinates for understanding the matrix's action.

TheoremSpectral Theorem for Real Symmetric Matrices

Let AA be a real n×nn \times n symmetric matrix. Then:

  1. All eigenvalues of AA are real
  2. AA has nn orthonormal eigenvectors
  3. AA is orthogonally diagonalizable: there exists an orthogonal matrix QQ and diagonal matrix Λ\Lambda such that: A=QΛQTA = Q\Lambda Q^T

where the columns of QQ are orthonormal eigenvectors and the diagonal entries of Λ\Lambda are the corresponding eigenvalues.

The complex version: A Hermitian matrix is unitarily diagonalizable: A=UΛUA = U\Lambda U^*.

This result is remarkable: every symmetric matrix, regardless of repeated eigenvalues, admits an orthonormal eigenvector basis. The orthogonality condition QT=Q1Q^T = Q^{-1} makes computations particularly elegant.

ExampleOrthogonal Diagonalization

Diagonalize A=[3113]A = \begin{bmatrix} 3 & 1 \\ 1 & 3 \end{bmatrix}.

Eigenvalues: det(AλI)=(3λ)21=0\det(A - \lambda I) = (3-\lambda)^2 - 1 = 0 gives λ1=4,λ2=2\lambda_1 = 4, \lambda_2 = 2.

Eigenvectors: For λ1=4\lambda_1 = 4: v1=(11)\mathbf{v}_1 = \begin{pmatrix} 1 \\ 1 \end{pmatrix}, normalized: q1=12(11)\mathbf{q}_1 = \frac{1}{\sqrt{2}}\begin{pmatrix} 1 \\ 1 \end{pmatrix}

For λ2=2\lambda_2 = 2: v2=(11)\mathbf{v}_2 = \begin{pmatrix} 1 \\ -1 \end{pmatrix}, normalized: q2=12(11)\mathbf{q}_2 = \frac{1}{\sqrt{2}}\begin{pmatrix} 1 \\ -1 \end{pmatrix}

Then: A=12[1111][4002]12[1111]=QΛQTA = \frac{1}{\sqrt{2}}\begin{bmatrix} 1 & 1 \\ 1 & -1 \end{bmatrix}\begin{bmatrix} 4 & 0 \\ 0 & 2 \end{bmatrix}\frac{1}{\sqrt{2}}\begin{bmatrix} 1 & 1 \\ 1 & -1 \end{bmatrix} = Q\Lambda Q^T

DefinitionSpectral Decomposition

The orthogonal diagonalization A=QΛQTA = Q\Lambda Q^T can be written as a sum: A=i=1nλiqiqiTA = \sum_{i=1}^n \lambda_i \mathbf{q}_i\mathbf{q}_i^T

where qi\mathbf{q}_i are the orthonormal eigenvectors and λi\lambda_i are eigenvalues. Each term λiqiqiT\lambda_i\mathbf{q}_i\mathbf{q}_i^T is a rank-one projection matrix.

This spectral decomposition expresses AA as a weighted sum of orthogonal projections onto eigenspaces.

TheoremProperties of Orthogonal Diagonalization

If A=QΛQTA = Q\Lambda Q^T with QQ orthogonal:

  1. Ak=QΛkQTA^k = Q\Lambda^k Q^T (easy to compute powers)
  2. det(A)=i=1nλi\det(A) = \prod_{i=1}^n \lambda_i (product of eigenvalues)
  3. tr(A)=i=1nλi\text{tr}(A) = \sum_{i=1}^n \lambda_i (sum of eigenvalues)
  4. AA is invertible iff all λi0\lambda_i \neq 0; then A1=QΛ1QTA^{-1} = Q\Lambda^{-1}Q^T
  5. The eigenvectors give principal axes; eigenvalues measure variance along these axes
Remark

Orthogonal diagonalization is optimal for computation: the orthogonality condition QT=Q1Q^T = Q^{-1} means we don't need to invert QQ explicitly. Moreover, orthogonal transformations preserve lengths and angles, making them numerically stable. The spectral decomposition perspective—viewing the matrix as a sum of scaled projections—is particularly powerful in applications like principal component analysis.