The Law of Large Numbers
The Law of Large Numbers (LLN) formalizes the intuitive notion that averages of many independent observations converge to the expected value, providing the mathematical justification for the frequency interpretation of probability.
Weak Law
Let be i.i.d. random variables with finite mean . Then the sample mean converges in probability to : That is, for every , .
When is finite, the WLLN follows immediately from Chebyshev's inequality: .
Strong Law
Let be i.i.d. random variables with (finite first moment). Then That is, .
The strong law is a deeper result than the weak law: it says that not only does get close to with high probability, but the entire sequence of averages converges to for almost every outcome.
Applications
To estimate for a random variable , generate i.i.d. samples and compute . The SLLN guarantees almost surely. The CLT gives the error estimate: , so doubling accuracy requires quadrupling .
The empirical distribution function satisfies (the Glivenko-Cantelli theorem). This "fundamental theorem of statistics" says the empirical CDF converges uniformly to the true CDF — a far-reaching generalization of the SLLN.
The condition cannot be removed. If are i.i.d. Cauchy random variables (), then and in fact has the same Cauchy distribution as for all — the average does not concentrate at all.