ConceptComplete

Law of Large Numbers and CLT - Core Definitions

Limit theorems describe the behavior of sequences of random variables, providing the theoretical foundation for statistical inference.

Convergence Concepts

Definition

Let X1,X2,…X_1, X_2, \ldots be a sequence of random variables and XX a random variable.

Convergence in Probability: Xnβ†’PXX_n \xrightarrow{P} X if for all Ο΅>0\epsilon > 0: lim⁑nβ†’βˆžP(∣Xnβˆ’X∣>Ο΅)=0\lim_{n \to \infty} P(|X_n - X| > \epsilon) = 0

Convergence in Distribution: Xnβ†’dXX_n \xrightarrow{d} X if: lim⁑nβ†’βˆžFXn(x)=FX(x)\lim_{n \to \infty} F_{X_n}(x) = F_X(x) at all continuity points of FXF_X.

Convergence Almost Surely: Xnβ†’a.s.XX_n \xrightarrow{a.s.} X if: P(lim⁑nβ†’βˆžXn=X)=1P\left(\lim_{n \to \infty} X_n = X\right) = 1

Relationships: Almost sure convergence β‡’\Rightarrow convergence in probability β‡’\Rightarrow convergence in distribution.

Example

Let Xn=X+1/nX_n = X + 1/n where XX is any random variable.

Then Xnβ†’a.s.XX_n \xrightarrow{a.s.} X (and thus in probability and distribution) since: ∣Xnβˆ’X∣=1/nβ†’0Β withΒ probabilityΒ 1|X_n - X| = 1/n \to 0 \text{ with probability 1}

Law of Large Numbers

Theorem

Let X1,X2,…X_1, X_2, \ldots be IID with E[Xi]=ΞΌE[X_i] = \mu and Var(Xi)=Οƒ2<∞\text{Var}(X_i) = \sigma^2 < \infty. Then: XΛ‰n=1nβˆ‘i=1nXiβ†’PΞΌ\bar{X}_n = \frac{1}{n}\sum_{i=1}^n X_i \xrightarrow{P} \mu

Proof Sketch: By Chebyshev's inequality: P(∣XΛ‰nβˆ’ΞΌβˆ£>Ο΅)≀Var(XΛ‰n)Ο΅2=Οƒ2/nΟ΅2β†’0P(|\bar{X}_n - \mu| > \epsilon) \leq \frac{\text{Var}(\bar{X}_n)}{\epsilon^2} = \frac{\sigma^2/n}{\epsilon^2} \to 0 β–‘

Theorem

Under the same conditions as WLLN: Xˉn→a.s.μ\bar{X}_n \xrightarrow{a.s.} \mu

The strong law states that sample averages converge to the population mean with probability 1, justifying the use of sample means as estimators.

Example

Monte Carlo Integration: Estimate ∫01g(x) dx\int_0^1 g(x) \, dx by generating uniform (0,1)(0,1) samples U1,…,UnU_1, \ldots, U_n: I^n=1nβˆ‘i=1ng(Ui)β†’a.s.E[g(U)]=∫01g(x) dx\hat{I}_n = \frac{1}{n}\sum_{i=1}^n g(U_i) \xrightarrow{a.s.} E[g(U)] = \int_0^1 g(x) \, dx

For g(x)=Ο€x1βˆ’x2g(x) = \pi x \sqrt{1-x^2} (quarter circle), this estimates Ο€/8\pi/8.

Remark

The law of large numbers provides the theoretical justification for the frequentist interpretation of probability: long-run relative frequencies converge to probabilities. It's the foundation for empirical scienceβ€”averages of repeated measurements converge to true values.