ConceptComplete

Modes of Convergence

There are several notions of convergence for sequences of random variables, each with different strengths and applications. Understanding their relationships is essential for rigorous probability theory.


Four Modes of Convergence

Definition

Let Xn,XX_n, X be random variables on a probability space (Ω,F,P)(\Omega, \mathcal{F}, P).

  1. Almost sure convergence (Xna.s.XX_n \xrightarrow{a.s.} X): P(limnXn=X)=1P(\lim_{n\to\infty} X_n = X) = 1
  2. Convergence in probability (XnPXX_n \xrightarrow{P} X): For all ϵ>0\epsilon > 0, limnP(XnX>ϵ)=0\lim_{n\to\infty} P(|X_n - X| > \epsilon) = 0
  3. Convergence in LpL^p (mean) (XnLpXX_n \xrightarrow{L^p} X): limnE[XnXp]=0\lim_{n\to\infty} E[|X_n - X|^p] = 0
  4. Convergence in distribution (XndXX_n \xrightarrow{d} X): limnFXn(x)=FX(x)\lim_{n\to\infty} F_{X_n}(x) = F_X(x) at all continuity points of FXF_X

Relationships

Theorem7.5Implications Between Modes

The following implications hold: Xna.s.X    XnPX    XndXX_n \xrightarrow{a.s.} X \implies X_n \xrightarrow{P} X \implies X_n \xrightarrow{d} X XnLpX    XnPX    XndXX_n \xrightarrow{L^p} X \implies X_n \xrightarrow{P} X \implies X_n \xrightarrow{d} X None of the reverse implications hold in general, though:

  • XnPXX_n \xrightarrow{P} X implies there exists a subsequence with Xnka.s.XX_{n_k} \xrightarrow{a.s.} X
  • XndcX_n \xrightarrow{d} c (a constant) implies XnPcX_n \xrightarrow{P} c
ExampleConvergence in probability but not a.s.

The "typewriter sequence": on [0,1][0,1] with Lebesgue measure, let Xn=1[k/2m,(k+1)/2m]X_n = \mathbf{1}_{[k/2^m, (k+1)/2^m]} where n=2m+kn = 2^m + k. Then XnP0X_n \xrightarrow{P} 0 (since the intervals shrink) but XnX_n does not converge a.s. to 00 (every point ω\omega is eventually covered by infinitely many intervals).


Slutsky's Theorem

Theorem7.6Slutsky's Theorem

If XndXX_n \xrightarrow{d} X and YnPcY_n \xrightarrow{P} c (a constant), then:

  1. Xn+YndX+cX_n + Y_n \xrightarrow{d} X + c
  2. YnXndcXY_n X_n \xrightarrow{d} cX
  3. Xn/YndX/cX_n / Y_n \xrightarrow{d} X/c (if c0c \neq 0)
RemarkThe continuous mapping theorem

If XndXX_n \xrightarrow{d} X and gg is a continuous function, then g(Xn)dg(X)g(X_n) \xrightarrow{d} g(X). Combined with Slutsky's theorem, this allows one to derive the limiting distribution of complex statistics from simpler ones — a technique used extensively in asymptotic statistics.