Lagrange Multipliers
Lagrange multipliers provide a method for finding extrema of a function subject to constraints. The key idea: at a constrained extremum, the gradient of the objective function is parallel to the gradient of the constraint. This reduces constrained optimization to solving a system of equations. Lagrange multipliers are ubiquitous in economics, physics, and machine learning.
Statement
To find the extrema of subject to the constraint , solve the system
for and (the Lagrange multiplier). Solutions are candidates for constrained extrema.
At a constrained extremum, must be perpendicular to the constraint manifold (otherwise, moving along the manifold would increase or decrease ). Since is also perpendicular to the manifold, and are parallel: .
Examples
Objective: . Constraint: .
Lagrange condition: gives , so and . Thus , so either or .
- If : , so , giving with (maximum).
- If : , giving (minimum).
Minimize subject to (ellipsoid). Lagrange: . This gives , , . Solutions occur when one variable is nonzero and others vanish, giving minimum distance along principal axes.
Multiple constraints
To optimize subject to , solve
Minimize subject to and . Lagrange: . Solving gives critical points.
Justification via Implicit Function Theorem
The Implicit Function Theorem guarantees that the constraint locally defines a manifold . To optimize on , we need (the gradient restricted to ). Since directions tangent to are perpendicular to , the condition ensures .
Summary
Lagrange multipliers for constrained optimization:
- Solve with .
- Geometric: parallel to at extrema.
- Multiple constraints: .
- Justification: Implicit Function Theorem.
See Extrema and Implicit Function Theorem.