Sufficiency and the Rao-Blackwell Theorem
Sufficient statistics capture all the information in a sample relevant to estimating a parameter, and the Rao-Blackwell theorem shows how to improve any estimator by conditioning on a sufficient statistic.
Sufficient Statistics
A statistic is sufficient for if the conditional distribution of the sample given does not depend on . Intuitively, once is known, the remaining sample information is "noise" that carries no information about .
A statistic is sufficient for if and only if the joint density (or mass function) factors as where depends on only through and does not depend on .
- Normal : is sufficient for
- Poisson : is sufficient for
- Exponential families: the natural sufficient statistic is
The Rao-Blackwell Theorem
Let be an unbiased estimator of and a sufficient statistic. Define . Then:
- is a function of only (does not depend on the full sample)
- (unbiased)
- for all , with equality iff is already a function of
The Rao-Blackwell theorem says that conditioning any unbiased estimator on a sufficient statistic never increases the variance and typically strictly reduces it.
Complete Statistics and MVUE
A sufficient statistic is complete if for any measurable function , for all implies almost surely. Completeness ensures uniqueness: there is only one unbiased estimator that is a function of .
If is a complete sufficient statistic and is an unbiased estimator of , then is the unique minimum variance unbiased estimator (MVUE). The Lehmann-Scheffe theorem thus provides a constructive method: find a complete sufficient statistic, then find an unbiased function of it.