ConceptComplete

Extrema of Multivariable Functions

Finding extrema (maxima and minima) of multivariable functions is central to optimization, economics, and physics. Critical points occur where the gradient vanishes, and the second derivative test (using the Hessian matrix) classifies them as local maxima, minima, or saddle points. Constrained optimization uses Lagrange multipliers.


Critical points

Definition9.1Critical point

A point a\mathbf{a} is a critical point of f:RnRf : \mathbb{R}^n \to \mathbb{R} if f(a)=0\nabla f(\mathbf{a}) = \mathbf{0} or f(a)\nabla f(\mathbf{a}) does not exist.

Theorem9.1First derivative test

If ff has a local extremum (maximum or minimum) at an interior point a\mathbf{a} and ff is differentiable at a\mathbf{a}, then f(a)=0\nabla f(\mathbf{a}) = \mathbf{0}.

ExampleCritical points of f(x, y) = x² - y²

f=(2x,2y)=0\nabla f = (2x, -2y) = \mathbf{0} gives (x,y)=(0,0)(x, y) = (0, 0) as the only critical point. However, (0,0)(0, 0) is a saddle point, not an extremum.


Second derivative test

Definition9.2Hessian matrix

The Hessian matrix of ff at a\mathbf{a} is

Hf(a)=(2fx122fx1xn2fxnx12fxn2)a.Hf(\mathbf{a}) = \begin{pmatrix} \frac{\partial^2 f}{\partial x_1^2} & \cdots & \frac{\partial^2 f}{\partial x_1 \partial x_n} \\ \vdots & \ddots & \vdots \\ \frac{\partial^2 f}{\partial x_n \partial x_1} & \cdots & \frac{\partial^2 f}{\partial x_n^2} \end{pmatrix}_{\mathbf{a}}.

Theorem9.2Second derivative test

Let a\mathbf{a} be a critical point of ff with f(a)=0\nabla f(\mathbf{a}) = \mathbf{0}. Let H=Hf(a)H = Hf(\mathbf{a}).

  • If HH is positive definite, then a\mathbf{a} is a local minimum.
  • If HH is negative definite, then a\mathbf{a} is a local maximum.
  • If HH has both positive and negative eigenvalues, then a\mathbf{a} is a saddle point.
  • If HH is indefinite (test inconclusive), higher-order tests are needed.
Examplef(x, y) = x² + y²

f=(2x,2y)=0\nabla f = (2x, 2y) = \mathbf{0} at (0,0)(0, 0). The Hessian is

H=(2002),H = \begin{pmatrix} 2 & 0 \\ 0 & 2 \end{pmatrix},

which is positive definite (eigenvalues 2,2>02, 2 > 0). Thus (0,0)(0, 0) is a local minimum.


Lagrange multipliers

Theorem9.3Lagrange multipliers

To optimize f(x)f(\mathbf{x}) subject to the constraint g(x)=0g(\mathbf{x}) = 0, solve

f=λg\nabla f = \lambda \nabla g

for x\mathbf{x} and λ\lambda (the Lagrange multiplier). The solutions are candidates for constrained extrema.

ExampleMaximize xy subject to x² + y² = 1

Lagrange condition: (y,x)=λ(2x,2y)(y, x) = \lambda (2x, 2y). This gives y=2λxy = 2\lambda x and x=2λyx = 2\lambda y, so x=4λ2xx = 4\lambda^2 x. If x0x \neq 0, then λ2=1/4\lambda^2 = 1/4, so λ=±1/2\lambda = \pm 1/2.

For λ=1/2\lambda = 1/2: y=xy = x, so 2x2=12x^2 = 1, giving (x,y)=(±1/2,±1/2)(x, y) = (\pm 1/\sqrt{2}, \pm 1/\sqrt{2}) with f=1/2f = 1/2 (maximum).

For λ=1/2\lambda = -1/2: y=xy = -x, giving f=1/2f = -1/2 (minimum).


Summary

Extrema of multivariable functions:

  • Critical points: f=0\nabla f = \mathbf{0}.
  • Second derivative test: Hessian determines local behavior.
  • Lagrange multipliers for constrained optimization.

See Implicit Function Theorem for related results.