TheoremComplete

Sufficiency and the Rao-Blackwell Theorem

Sufficient statistics capture all the information in a sample relevant to estimating a parameter, and the Rao-Blackwell theorem shows how to improve any estimator by conditioning on a sufficient statistic.


Sufficient Statistics

Definition

A statistic T=T(X1,,Xn)T = T(X_1, \ldots, X_n) is sufficient for θ\theta if the conditional distribution of the sample (X1,,Xn)(X_1, \ldots, X_n) given TT does not depend on θ\theta. Intuitively, once TT is known, the remaining sample information is "noise" that carries no information about θ\theta.

Theorem8.6Fisher-Neyman Factorization

A statistic T(X)T(\mathbf{X}) is sufficient for θ\theta if and only if the joint density (or mass function) factors as f(x;θ)=g(T(x),θ)h(x)f(\mathbf{x}; \theta) = g(T(\mathbf{x}), \theta) \cdot h(\mathbf{x}) where gg depends on x\mathbf{x} only through T(x)T(\mathbf{x}) and hh does not depend on θ\theta.

ExampleSufficient statistics for common families
  • Normal N(μ,σ2)N(\mu, \sigma^2): (Xˉ,(XiXˉ)2)(\bar{X}, \sum(X_i - \bar{X})^2) is sufficient for (μ,σ2)(\mu, \sigma^2)
  • Poisson Poi(λ)\text{Poi}(\lambda): Xi\sum X_i is sufficient for λ\lambda
  • Exponential families: the natural sufficient statistic is T(x)=t(xi)T(\mathbf{x}) = \sum t(x_i)

The Rao-Blackwell Theorem

Theorem8.7Rao-Blackwell Theorem

Let θ^\hat{\theta} be an unbiased estimator of θ\theta and TT a sufficient statistic. Define θ~=E[θ^T]\tilde{\theta} = E[\hat{\theta} | T]. Then:

  1. θ~\tilde{\theta} is a function of TT only (does not depend on the full sample)
  2. E[θ~]=θE[\tilde{\theta}] = \theta (unbiased)
  3. Var(θ~)Var(θ^)\operatorname{Var}(\tilde{\theta}) \leq \operatorname{Var}(\hat{\theta}) for all θ\theta, with equality iff θ^\hat{\theta} is already a function of TT

The Rao-Blackwell theorem says that conditioning any unbiased estimator on a sufficient statistic never increases the variance and typically strictly reduces it.


Complete Statistics and MVUE

Definition

A sufficient statistic TT is complete if for any measurable function gg, E[g(T)]=0E[g(T)] = 0 for all θ\theta implies g(T)=0g(T) = 0 almost surely. Completeness ensures uniqueness: there is only one unbiased estimator that is a function of TT.

RemarkLehmann-Scheffe theorem

If TT is a complete sufficient statistic and θ~=g(T)\tilde{\theta} = g(T) is an unbiased estimator of θ\theta, then θ~\tilde{\theta} is the unique minimum variance unbiased estimator (MVUE). The Lehmann-Scheffe theorem thus provides a constructive method: find a complete sufficient statistic, then find an unbiased function of it.