Orthogonal Projections and Least Squares
Orthogonal projections provide the best approximation of a vector by elements of a subspace. This geometric insight solves the fundamental problem of least squares approximation.
Let be a finite-dimensional subspace of inner product space , and let . The orthogonal projection is the unique closest point in to :
for any with .
The error vector is orthogonal to .
This theorem is the theoretical foundation for least squares: the best fit is achieved when the residual is orthogonal to the approximation space.
Consider the overdetermined system where is with (more equations than unknowns). The least squares solution minimizes and satisfies the normal equations:
If has full column rank, the solution is unique: .
The projection matrix projects onto the column space of .
Fit a line to data points .
Set up system:
The least squares solution minimizes the sum of squared residuals .
Solving normal equations: gives the regression coefficients.
Let be a subspace of finite-dimensional inner product space . Every can be uniquely written as:
where and . This decomposition satisfies:
A matrix is an orthogonal projection matrix if and only if:
- (idempotent)
- (symmetric)
The matrix projects onto the orthogonal complement of the range of .
Orthogonal projection is pervasive in applications: data fitting (regression), signal processing (filtering), computer graphics (shadows), and quantum mechanics (measurement). The geometric insight—that the best approximation is achieved when the error is perpendicular—translates into the algebraic condition captured by the normal equations.