ConceptComplete

Orthogonality

Orthogonality is the generalization of perpendicularity to abstract inner product spaces. Two vectors are orthogonal when their inner product vanishes. This simple condition has profound consequences: orthogonal sets are always linearly independent, orthogonal decompositions split spaces into complementary pieces, and orthogonal complements provide canonical complements to subspaces.


Definition

Definition6.5Orthogonal vectors

Two vectors u,vu, v in an inner product space VV are orthogonal (written uvu \perp v) if:

u,v=0.\langle u, v \rangle = 0.

A set of vectors {v1,,vk}\{v_1, \ldots, v_k\} is an orthogonal set if vi,vj=0\langle v_i, v_j \rangle = 0 for all iji \neq j.

ExampleStandard basis is orthogonal

In Rn\mathbb{R}^n with the dot product, the standard basis {e1,,en}\{e_1, \ldots, e_n\} is orthogonal: ei,ej=δij\langle e_i, e_j \rangle = \delta_{ij} (the Kronecker delta).

In R3\mathbb{R}^3: e1e2=0e_1 \cdot e_2 = 0, e1e3=0e_1 \cdot e_3 = 0, e2e3=0e_2 \cdot e_3 = 0.

ExampleNon-standard orthogonal vectors

In R3\mathbb{R}^3: u=(1,1,0)u = (1, 1, 0) and v=(1,1,0)v = (1, -1, 0).

u,v=11+0=0\langle u, v \rangle = 1 - 1 + 0 = 0, so uvu \perp v.

But uu and vv are not standard basis vectors -- orthogonality just means their dot product vanishes.

ExampleOrthogonality of trigonometric functions

On C[π,π]C[-\pi, \pi] with f,g=ππf(x)g(x)dx\langle f, g \rangle = \int_{-\pi}^{\pi} f(x)g(x)\,dx:

sin(mx),sin(nx)=0\langle \sin(mx), \sin(nx) \rangle = 0 for mnm \neq n (by orthogonality of Fourier modes).

sin(mx),cos(nx)=0\langle \sin(mx), \cos(nx) \rangle = 0 for all m,nm, n (sine is odd, cosine is even).

1,cos(nx)=ππcos(nx)dx=0\langle 1, \cos(nx) \rangle = \int_{-\pi}^{\pi} \cos(nx)\,dx = 0 for n1n \geq 1.

This orthogonality is the foundation of Fourier analysis.


Properties of orthogonal sets

TheoremOrthogonal sets are linearly independent

If {v1,,vk}\{v_1, \ldots, v_k\} is an orthogonal set of nonzero vectors, then v1,,vkv_1, \ldots, v_k are linearly independent.

Proof

Suppose c1v1++ckvk=0c_1 v_1 + \cdots + c_k v_k = 0. Take the inner product with vjv_j:

0=c1v1++ckvk,vj=cjvj,vj.0 = \langle c_1 v_1 + \cdots + c_k v_k, v_j \rangle = c_j \langle v_j, v_j \rangle.

Since vj0v_j \neq 0, vj,vj>0\langle v_j, v_j \rangle > 0, so cj=0c_j = 0. This holds for all jj.

ExampleOrthogonality implies independence

v1=(1,1,0)v_1 = (1, 1, 0), v2=(1,1,0)v_2 = (1, -1, 0), v3=(0,0,1)v_3 = (0, 0, 1) in R3\mathbb{R}^3.

Check pairwise orthogonality: v1v2=0v_1 \cdot v_2 = 0, v1v3=0v_1 \cdot v_3 = 0, v2v3=0v_2 \cdot v_3 = 0 ✓.

These are automatically linearly independent and form a basis of R3\mathbb{R}^3.

TheoremPythagorean theorem

If uvu \perp v, then u+v2=u2+v2\|u + v\|^2 = \|u\|^2 + \|v\|^2.

More generally, if v1,,vkv_1, \ldots, v_k are pairwise orthogonal:

v1++vk2=v12++vk2.\|v_1 + \cdots + v_k\|^2 = \|v_1\|^2 + \cdots + \|v_k\|^2.

ExamplePythagorean theorem in R^3

u=(3,0,0)u = (3, 0, 0) and v=(0,4,0)v = (0, 4, 0): uvu \perp v, u+v=(3,4,0)=5=9+16=u2+v2\|u + v\| = \|(3, 4, 0)\| = 5 = \sqrt{9 + 16} = \sqrt{\|u\|^2 + \|v\|^2} ✓.

This is the classical Pythagorean theorem for right triangles.


Orthogonal complement

Definition6.6Orthogonal complement

For a subset SVS \subseteq V, the orthogonal complement of SS is:

S={vV:v,s=0 for all sS}.S^\perp = \{v \in V : \langle v, s \rangle = 0 \text{ for all } s \in S\}.

SS^\perp is always a subspace of VV, even if SS is not.

ExampleOrthogonal complement in R^3

Let W=span{(1,0,0)}W = \operatorname{span}\{(1, 0, 0)\} (the xx-axis) in R3\mathbb{R}^3.

W={(x,y,z):x=0}=span{(0,1,0),(0,0,1)}W^\perp = \{(x, y, z) : x = 0\} = \operatorname{span}\{(0, 1, 0), (0, 0, 1)\} (the yzyz-plane).

dimW+dimW=1+2=3=dimR3\dim W + \dim W^\perp = 1 + 2 = 3 = \dim \mathbb{R}^3.

ExampleOrthogonal complement of a plane

W=span{(1,1,0),(0,1,1)}W = \operatorname{span}\{(1, 1, 0), (0, 1, 1)\} in R3\mathbb{R}^3.

W={(x,y,z):x+y=0 and y+z=0}={(x,x,x):xR}=span{(1,1,1)}W^\perp = \{(x, y, z) : x + y = 0 \text{ and } y + z = 0\} = \{(x, -x, x) : x \in \mathbb{R}\} = \operatorname{span}\{(1, -1, 1)\}.

Check: (1,1,1)(1,1,0)=11+0=0(1, -1, 1) \cdot (1, 1, 0) = 1 - 1 + 0 = 0 ✓ and (1,1,1)(0,1,1)=01+1=0(1, -1, 1) \cdot (0, 1, 1) = 0 - 1 + 1 = 0 ✓.

ExampleOrthogonal complement and null space

The orthogonal complement of the row space of AA is the null space of AA:

Row(A)=Null(A).\operatorname{Row}(A)^\perp = \operatorname{Null}(A).

For A=(123456)A = \begin{pmatrix} 1 & 2 & 3 \\ 4 & 5 & 6 \end{pmatrix}: Row(A)=span{(1,2,3),(4,5,6)}\operatorname{Row}(A) = \operatorname{span}\{(1,2,3), (4,5,6)\} is a 22-dimensional subspace of R3\mathbb{R}^3.

Null(A)=span{(1,2,1)}\operatorname{Null}(A) = \operatorname{span}\{(1, -2, 1)\}, which has dimension 1=321 = 3 - 2. And (1,2,1)(1,2,3)=14+3=0(1, -2, 1) \cdot (1, 2, 3) = 1 - 4 + 3 = 0 ✓.


Fundamental properties of orthogonal complements

TheoremProperties of orthogonal complements

Let VV be a finite-dimensional inner product space and WW a subspace.

  1. V=WWV = W \oplus W^\perp (orthogonal direct sum).
  2. dimW+dimW=dimV\dim W + \dim W^\perp = \dim V.
  3. (W)=W(W^\perp)^\perp = W.
  4. W1W2    W2W1W_1 \subseteq W_2 \implies W_2^\perp \subseteq W_1^\perp.
  5. (W1+W2)=W1W2(W_1 + W_2)^\perp = W_1^\perp \cap W_2^\perp and (W1W2)=W1+W2(W_1 \cap W_2)^\perp = W_1^\perp + W_2^\perp.
ExampleDouble complement

W=span{(1,0,0)}W = \operatorname{span}\{(1, 0, 0)\} in R3\mathbb{R}^3: W=span{(0,1,0),(0,0,1)}W^\perp = \operatorname{span}\{(0,1,0), (0,0,1)\}, (W)={(x,y,z):y=0,z=0}=W(W^\perp)^\perp = \{(x,y,z): y = 0, z = 0\} = W ✓.

ExampleThe four fundamental subspaces

For AMm×n(R)A \in M_{m \times n}(\mathbb{R}), the four fundamental subspaces are:

  1. Col(A)Rm\operatorname{Col}(A) \subseteq \mathbb{R}^m (column space), dimension rr.
  2. Null(AT)Rm\operatorname{Null}(A^T) \subseteq \mathbb{R}^m (left null space), dimension mrm - r.
  3. Row(A)Rn\operatorname{Row}(A) \subseteq \mathbb{R}^n (row space), dimension rr.
  4. Null(A)Rn\operatorname{Null}(A) \subseteq \mathbb{R}^n (null space), dimension nrn - r.

The orthogonal complement relationships: Col(A)=Null(AT)\operatorname{Col}(A)^\perp = \operatorname{Null}(A^T) and Row(A)=Null(A)\operatorname{Row}(A)^\perp = \operatorname{Null}(A).

For A=(100100)A = \begin{pmatrix} 1 & 0 \\ 0 & 1 \\ 0 & 0 \end{pmatrix}: r=2r = 2, Col(A)=span{e1,e2}\operatorname{Col}(A) = \operatorname{span}\{e_1, e_2\}, Null(AT)=span{e3}\operatorname{Null}(A^T) = \operatorname{span}\{e_3\}, and R3=Col(A)Null(AT)\mathbb{R}^3 = \operatorname{Col}(A) \oplus \operatorname{Null}(A^T).


Orthogonal decomposition

TheoremOrthogonal decomposition theorem

If WW is a subspace of a finite-dimensional inner product space VV, then every vector vVv \in V can be uniquely written as:

v=w+w,wW,wW.v = w + w^\perp, \quad w \in W, \quad w^\perp \in W^\perp.

The component w=projW(v)w = \operatorname{proj}_W(v) is the orthogonal projection of vv onto WW.

ExampleProjecting onto a line

W=span{(1,1)}W = \operatorname{span}\{(1, 1)\} in R2\mathbb{R}^2, v=(3,1)v = (3, 1).

w=projW(v)=v,(1,1)(1,1)2(1,1)=42(1,1)=(2,2)w = \operatorname{proj}_W(v) = \frac{\langle v, (1,1) \rangle}{\|(1,1)\|^2}(1,1) = \frac{4}{2}(1,1) = (2, 2).

w=vw=(3,1)(2,2)=(1,1)w^\perp = v - w = (3, 1) - (2, 2) = (1, -1).

Check: (2,2)(1,1)=22=0(2, 2) \cdot (1, -1) = 2 - 2 = 0 ✓ and v=(2,2)+(1,1)v = (2, 2) + (1, -1) ✓.

ExampleProjecting onto a plane in R^3

W=span{(1,0,0),(0,1,0)}W = \operatorname{span}\{(1,0,0), (0,1,0)\} (the xyxy-plane), v=(3,4,5)v = (3, 4, 5).

projW(v)=(3,4,0)\operatorname{proj}_W(v) = (3, 4, 0), vprojW(v)=(0,0,5)Wv - \operatorname{proj}_W(v) = (0, 0, 5) \in W^\perp ✓.

ExampleProjecting polynomials

V=P2V = \mathcal{P}_2 (polynomials of degree 2\leq 2) with f,g=01f(x)g(x)dx\langle f, g \rangle = \int_0^1 f(x)g(x)\,dx.

W=span{1}W = \operatorname{span}\{1\}. The projection of f(x)=xf(x) = x onto WW:

projW(x)=x,11,11=1/21=12\operatorname{proj}_W(x) = \frac{\langle x, 1 \rangle}{\langle 1, 1 \rangle} \cdot 1 = \frac{1/2}{1} = \frac{1}{2}.

The orthogonal component is x12x - \frac{1}{2}, and x1/2,1=01(x1/2)dx=0\langle x - 1/2, 1 \rangle = \int_0^1 (x - 1/2)\,dx = 0 ✓.


Orthogonal matrices

Definition6.7Orthogonal matrix

A real n×nn \times n matrix QQ is orthogonal if QTQ=IQ^T Q = I, equivalently Q1=QTQ^{-1} = Q^T. The columns of QQ form an orthonormal basis.

For complex matrices, the analogous notion is unitary: UU=IU^*U = I where U=UˉTU^* = \bar{U}^T.

ExampleRotation matrices are orthogonal

Q=(cosθsinθsinθcosθ)Q = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}. Then QTQ=IQ^T Q = I ✓ (the columns (cosθ,sinθ)(\cos\theta, \sin\theta) and (sinθ,cosθ)(-\sin\theta, \cos\theta) are orthonormal).

Orthogonal matrices preserve inner products: Qu,Qv=(Qu)T(Qv)=uTQTQv=uTv=u,v\langle Qu, Qv \rangle = (Qu)^T(Qv) = u^T Q^T Q v = u^T v = \langle u, v \rangle.

ExampleReflection matrices are orthogonal

Reflection across the line through (1,0)(1, 0): Q=(1001)Q = \begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix}. Then QTQ=IQ^T Q = I and detQ=1\det Q = -1 (orthogonal but not a rotation).

Reflection across the line y=xy = x: Q=(0110)Q = \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix}, QTQ=IQ^T Q = I ✓.

Example3D orthogonal matrix

Q=13(212221122)Q = \frac{1}{3}\begin{pmatrix} 2 & -1 & 2 \\ 2 & 2 & -1 \\ -1 & 2 & 2 \end{pmatrix}. One can verify QTQ=IQ^TQ = I and detQ=1\det Q = 1, so QQ is a rotation in R3\mathbb{R}^3. The axis of rotation is the eigenvector for λ=1\lambda = 1.


Summary

RemarkOrthogonality as the central concept

Orthogonality provides the geometric backbone of inner product spaces:

  • Orthogonal sets are automatically linearly independent.
  • Every finite-dimensional inner product space decomposes as V=WWV = W \oplus W^\perp.
  • The four fundamental subspaces of a matrix are related by orthogonal complement.
  • Orthogonal (unitary) matrices preserve inner products and are the "symmetries" of the inner product.
  • The Gram--Schmidt process converts any basis to an orthogonal one, and orthogonal projections solve least-squares problems.