ConceptComplete

Eigenvalue and Eigenvector

An eigenvector of a linear transformation TT is a nonzero vector vv whose direction is preserved by TT: the output TvTv is a scalar multiple of vv. The corresponding scalar λ\lambda is the eigenvalue. Eigenvalues reveal the intrinsic scaling behavior of a linear map, independent of any choice of basis.


Definition

Definition5.1Eigenvalue and eigenvector

Let VV be a vector space over a field FF, and let T:VVT: V \to V be a linear transformation. A scalar λF\lambda \in F is an eigenvalue of TT if there exists a nonzero vector vVv \in V such that:

T(v)=λv.T(v) = \lambda v.

Such a vector v0v \neq 0 is called an eigenvector of TT corresponding to the eigenvalue λ\lambda.

For a matrix AMn×n(F)A \in M_{n \times n}(F), this becomes Av=λvAv = \lambda v, or equivalently (AλI)v=0(A - \lambda I)v = 0 with v0v \neq 0.

Definition5.2Eigenspace

The eigenspace of TT corresponding to an eigenvalue λ\lambda is:

Eλ=ker(TλI)={vV:T(v)=λv}.E_\lambda = \ker(T - \lambda I) = \{ v \in V : T(v) = \lambda v \}.

This is a subspace of VV (it always contains the zero vector). The dimension dimEλ\dim E_\lambda is the geometric multiplicity of λ\lambda.

ExampleEigenvalues of a 2x2 matrix

Let A=(3102)A = \begin{pmatrix} 3 & 1 \\ 0 & 2 \end{pmatrix}. To find eigenvalues, solve det(AλI)=0\det(A - \lambda I) = 0:

det(3λ102λ)=(3λ)(2λ)=0.\det \begin{pmatrix} 3 - \lambda & 1 \\ 0 & 2 - \lambda \end{pmatrix} = (3 - \lambda)(2 - \lambda) = 0.

So λ1=3\lambda_1 = 3 and λ2=2\lambda_2 = 2.

For λ1=3\lambda_1 = 3: (A3I)v=0(A - 3I)v = 0 gives (0101)v=0\begin{pmatrix} 0 & 1 \\ 0 & -1 \end{pmatrix} v = 0, so E3=span{(1,0)}E_3 = \operatorname{span}\{(1, 0)\}.

For λ2=2\lambda_2 = 2: (A2I)v=0(A - 2I)v = 0 gives (1100)v=0\begin{pmatrix} 1 & 1 \\ 0 & 0 \end{pmatrix} v = 0, so E2=span{(1,1)}E_2 = \operatorname{span}\{(-1, 1)\}.

ExampleRotation matrix (no real eigenvalues)

The rotation by θ\theta in R2\mathbb{R}^2: Rθ=(cosθsinθsinθcosθ)R_\theta = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}.

The characteristic polynomial is λ22cosθλ+1=0\lambda^2 - 2\cos\theta \cdot \lambda + 1 = 0, with discriminant 4cos2θ4=4sin2θ4\cos^2\theta - 4 = -4\sin^2\theta.

  • For θ0,π\theta \neq 0, \pi: the discriminant is negative, so there are no real eigenvalues. Over C\mathbb{C}, the eigenvalues are λ=e±iθ\lambda = e^{\pm i\theta}.
  • For θ=0\theta = 0: λ=1\lambda = 1 (double), every nonzero vector is an eigenvector.
  • For θ=π\theta = \pi: λ=1\lambda = -1 (double), Rπ=IR_\pi = -I.

Eigenvalues of specific transformations

ExampleDiagonal matrices

If A=diag(d1,d2,,dn)A = \operatorname{diag}(d_1, d_2, \ldots, d_n), then the eigenvalues are exactly d1,d2,,dnd_1, d_2, \ldots, d_n, with eigenvectors e1,e2,,ene_1, e_2, \ldots, e_n (the standard basis vectors).

For A=diag(2,1,5)A = \operatorname{diag}(2, -1, 5): eigenvalues 2,1,52, -1, 5 with eigenvectors (1,0,0)(1,0,0), (0,1,0)(0,1,0), (0,0,1)(0,0,1).

ExampleTriangular matrices

If AA is upper (or lower) triangular, the eigenvalues are the diagonal entries. This follows from det(AλI)\det(A - \lambda I) being the product of the diagonal entries of AλIA - \lambda I.

For A=(147028003)A = \begin{pmatrix} 1 & 4 & 7 \\ 0 & 2 & 8 \\ 0 & 0 & 3 \end{pmatrix}: eigenvalues are 1,2,31, 2, 3.

ExampleProjection operators

Let P:VVP: V \to V be a projection onto a subspace WW along a complement UU (so V=WUV = W \oplus U and PW=idP|_W = \operatorname{id}, PU=0P|_U = 0). Then:

  • λ=1\lambda = 1: eigenspace is WW (everything in WW is fixed by PP).
  • λ=0\lambda = 0: eigenspace is UU (everything in UU is killed by PP).
  • These are the only eigenvalues, since P2=PP^2 = P implies λ2=λ\lambda^2 = \lambda, so λ{0,1}\lambda \in \{0, 1\}.
ExampleNilpotent operators

A linear map N:VVN: V \to V is nilpotent if Nk=0N^k = 0 for some k1k \geq 1. If Nv=λvNv = \lambda v with v0v \neq 0, then Nkv=λkv=0N^k v = \lambda^k v = 0, so λk=0\lambda^k = 0, hence λ=0\lambda = 0.

The only eigenvalue of a nilpotent operator is 00. For example, N=(0100)N = \begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix} has N2=0N^2 = 0, eigenvalue λ=0\lambda = 0 with eigenspace span{(1,0)}\operatorname{span}\{(1, 0)\} (geometric multiplicity 11, algebraic multiplicity 22).


Key properties

TheoremLinear independence of eigenvectors

Eigenvectors corresponding to distinct eigenvalues are linearly independent. That is, if λ1,,λk\lambda_1, \ldots, \lambda_k are distinct eigenvalues of TT with eigenvectors v1,,vkv_1, \ldots, v_k, then v1,,vkv_1, \ldots, v_k are linearly independent.

ProofProof by induction

Base case k=1k = 1: a single eigenvector is nonzero, hence linearly independent.

Inductive step: suppose c1v1++ckvk=0c_1 v_1 + \cdots + c_k v_k = 0. Apply TT: c1λ1v1++ckλkvk=0c_1 \lambda_1 v_1 + \cdots + c_k \lambda_k v_k = 0. Subtract λk\lambda_k times the original: c1(λ1λk)v1++ck1(λk1λk)vk1=0c_1(\lambda_1 - \lambda_k)v_1 + \cdots + c_{k-1}(\lambda_{k-1} - \lambda_k)v_{k-1} = 0. By the inductive hypothesis, ci(λiλk)=0c_i(\lambda_i - \lambda_k) = 0 for i=1,,k1i = 1, \ldots, k-1. Since λiλk\lambda_i \neq \lambda_k, we get ci=0c_i = 0. Then ckvk=0c_k v_k = 0 gives ck=0c_k = 0.

ExampleThree distinct eigenvalues in R^3

Let A=(200130001)A = \begin{pmatrix} 2 & 0 & 0 \\ 1 & 3 & 0 \\ 0 & 0 & -1 \end{pmatrix}. The eigenvalues (diagonal of a triangular matrix) are 2,3,12, 3, -1.

Computing eigenvectors: v1=(1,1,0)v_1 = (-1, 1, 0) for λ=2\lambda = 2; v2=(0,1,0)v_2 = (0, 1, 0) for λ=3\lambda = 3; v3=(0,0,1)v_3 = (0, 0, 1) for λ=1\lambda = -1.

These three eigenvectors are linearly independent (they form a basis of R3\mathbb{R}^3), confirming the theorem.


Spectrum and trace/determinant relations

Definition5.3Spectrum

The spectrum of TT (or of the matrix AA) is the set of all eigenvalues:

σ(T)={λF:TλI is not invertible}.\sigma(T) = \{ \lambda \in F : T - \lambda I \text{ is not invertible} \}.

TheoremTrace and determinant from eigenvalues

If AMn×n(F)A \in M_{n \times n}(F) has eigenvalues λ1,,λn\lambda_1, \ldots, \lambda_n (counted with algebraic multiplicity, over the algebraic closure of FF), then:

tr(A)=λ1+λ2++λn,det(A)=λ1λ2λn.\operatorname{tr}(A) = \lambda_1 + \lambda_2 + \cdots + \lambda_n, \quad \det(A) = \lambda_1 \lambda_2 \cdots \lambda_n.

ExampleTrace and determinant verification

For A=(4213)A = \begin{pmatrix} 4 & 2 \\ 1 & 3 \end{pmatrix}: tr(A)=7\operatorname{tr}(A) = 7, det(A)=10\det(A) = 10.

Characteristic polynomial: λ27λ+10=(λ5)(λ2)=0\lambda^2 - 7\lambda + 10 = (\lambda - 5)(\lambda - 2) = 0, so λ1=5,λ2=2\lambda_1 = 5, \lambda_2 = 2.

Check: 5+2=7=tr(A)5 + 2 = 7 = \operatorname{tr}(A) and 52=10=det(A)5 \cdot 2 = 10 = \det(A).

Example3x3 verification

For A=(100020003)A = \begin{pmatrix} 1 & 0 & 0 \\ 0 & 2 & 0 \\ 0 & 0 & 3 \end{pmatrix}: eigenvalues 1,2,31, 2, 3.

tr(A)=1+2+3=6\operatorname{tr}(A) = 1 + 2 + 3 = 6, det(A)=123=6\det(A) = 1 \cdot 2 \cdot 3 = 6.


Eigenvalues of special matrices

ExampleIdempotent matrices

A2=AA^2 = A implies λ2=λ\lambda^2 = \lambda, so λ{0,1}\lambda \in \{0, 1\}.

For A=(1000)A = \begin{pmatrix} 1 & 0 \\ 0 & 0 \end{pmatrix}: eigenvalues 1,01, 0. For A=12(1111)A = \frac{1}{2}\begin{pmatrix} 1 & 1 \\ 1 & 1 \end{pmatrix}: eigenvalues 1,01, 0 (check: tr=1\operatorname{tr} = 1, det=0\det = 0).

ExampleInvolutions

A2=IA^2 = I implies λ2=1\lambda^2 = 1, so λ{1,1}\lambda \in \{1, -1\}.

For A=(0110)A = \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix} (permutation matrix): det(AλI)=λ21\det(A - \lambda I) = \lambda^2 - 1, eigenvalues 1,11, -1, eigenvectors (1,1)(1,1) and (1,1)(1,-1).

ExampleStochastic matrices

A (right) stochastic matrix has nonnegative entries with each row summing to 11. Then A1=1A \mathbf{1} = \mathbf{1} where 1=(1,1,,1)T\mathbf{1} = (1, 1, \ldots, 1)^T... actually, for a right stochastic matrix, columns sum to 11, so 1TA=1T\mathbf{1}^T A = \mathbf{1}^T. For a left stochastic matrix (rows sum to 11): A1=1A \mathbf{1} = \mathbf{1}, so λ=1\lambda = 1 is always an eigenvalue with eigenvector 1\mathbf{1}.

For A=(0.70.30.40.6)A = \begin{pmatrix} 0.7 & 0.3 \\ 0.4 & 0.6 \end{pmatrix}: eigenvalues λ1=1\lambda_1 = 1 (with eigenvector (3,4)T(3, 4)^T after scaling) and λ2=0.3\lambda_2 = 0.3.


Eigenvalues under matrix operations

TheoremEigenvalue relationships

Let AA have eigenvalue λ\lambda with eigenvector vv. Then:

  • AkA^k has eigenvalue λk\lambda^k with eigenvector vv (for any k1k \geq 1).
  • A+cIA + cI has eigenvalue λ+c\lambda + c with eigenvector vv.
  • If AA is invertible, A1A^{-1} has eigenvalue λ1\lambda^{-1} with eigenvector vv.
  • p(A)p(A) has eigenvalue p(λ)p(\lambda) for any polynomial pp.
ExampleEigenvalue of a matrix polynomial

Let A=(2103)A = \begin{pmatrix} 2 & 1 \\ 0 & 3 \end{pmatrix} with eigenvalues 2,32, 3.

For p(x)=x25x+6=(x2)(x3)p(x) = x^2 - 5x + 6 = (x-2)(x-3): p(A)p(A) has eigenvalues p(2)=0p(2) = 0 and p(3)=0p(3) = 0. Indeed, p(A)=A25A+6I=0p(A) = A^2 - 5A + 6I = 0 (this is the Cayley--Hamilton theorem in action).

For q(x)=x2+1q(x) = x^2 + 1: q(A)q(A) has eigenvalues q(2)=5q(2) = 5 and q(3)=10q(3) = 10.


Geometric and algebraic multiplicity

Definition5.4Algebraic and geometric multiplicity

For an eigenvalue λ\lambda of TT:

  • The algebraic multiplicity ma(λ)m_a(\lambda) is the multiplicity of λ\lambda as a root of the characteristic polynomial.
  • The geometric multiplicity mg(λ)=dimEλ=dimker(TλI)m_g(\lambda) = \dim E_\lambda = \dim \ker(T - \lambda I).

Always: 1mg(λ)ma(λ)1 \leq m_g(\lambda) \leq m_a(\lambda).

ExampleA defective matrix (m_g less than m_a)

A=(2102)A = \begin{pmatrix} 2 & 1 \\ 0 & 2 \end{pmatrix} has characteristic polynomial (λ2)2(\lambda - 2)^2, so λ=2\lambda = 2 with ma=2m_a = 2.

But A2I=(0100)A - 2I = \begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix}, so ker(A2I)=span{(1,0)}\ker(A - 2I) = \operatorname{span}\{(1, 0)\}, giving mg=1<2=mam_g = 1 < 2 = m_a.

This matrix is called defective: it cannot be diagonalized because there are not enough linearly independent eigenvectors.

ExampleEqual multiplicities

A=(3003)=3IA = \begin{pmatrix} 3 & 0 \\ 0 & 3 \end{pmatrix} = 3I has λ=3\lambda = 3 with ma=2m_a = 2 and mg=2m_g = 2 (every nonzero vector is an eigenvector). This matrix is diagonalizable (it is already diagonal).


Summary

RemarkEigenvalues as fundamental invariants

Eigenvalues encode essential information about a linear transformation:

  • The trace is the sum of eigenvalues, the determinant is the product.
  • Invertibility: AA is invertible iff 0σ(A)0 \notin \sigma(A).
  • Nilpotence: AA is nilpotent iff σ(A)={0}\sigma(A) = \{0\}.
  • The gap between geometric and algebraic multiplicity measures the "defect" of the transformation -- the failure to be diagonalizable.
  • Eigenvectors for distinct eigenvalues are always linearly independent, which is the foundation for diagonalization.