TheoremComplete

The Poisson Limit Theorem

The Poisson limit theorem (or law of rare events) states that the sum of many independent rare events converges in distribution to a Poisson distribution. This fundamental result explains why the Poisson distribution arises ubiquitously in applications and justifies the Poisson process as a model for random arrivals.


Statement of the theorem

Theorem2.1Poisson limit theorem (binomial approximation)

Let XnBinomial(n,pn)X_n \sim \text{Binomial}(n, p_n) with npnλn p_n \to \lambda as nn \to \infty (where λ>0\lambda > 0 is fixed). Then

P(Xn=k)eλλkk!as n,\mathbb{P}(X_n = k) \to e^{-\lambda} \frac{\lambda^k}{k!} \quad \text{as } n \to \infty,

for each k=0,1,2,k = 0, 1, 2, \ldots. In other words, XndPoisson(λ)X_n \xrightarrow{d} \text{Poisson}(\lambda).

Intuition: Suppose nn is large and pnp_n is small, with npnλn p_n \approx \lambda fixed. Then the binomial distribution (sum of nn Bernoulli trials, each with success probability pnp_n) is well-approximated by a Poisson distribution with parameter λ\lambda.

ExampleDefects in manufacturing

A factory produces n=10,000n = 10{,}000 items per day. Each item is defective with probability p=0.0005p = 0.0005, independently. The number of defects XX is Binomial(10,000,0.0005)\text{Binomial}(10{,}000, 0.0005). Since np=5np = 5, we have XPoisson(5)X \approx \text{Poisson}(5):

P(X=0)e50.0067,P(X=1)5e50.0337.\mathbb{P}(X = 0) \approx e^{-5} \approx 0.0067, \quad \mathbb{P}(X = 1) \approx 5 e^{-5} \approx 0.0337.

The Poisson approximation is much simpler than the binomial formula and highly accurate for large nn and small pp.


Proof

We prove convergence using the binomial probability mass function.

Step 1: Recall

P(Xn=k)=(nk)pnk(1pn)nk.\mathbb{P}(X_n = k) = \binom{n}{k} p_n^k (1-p_n)^{n-k}.

Step 2: Write

(nk)=n(n1)(nk+1)k!=nkk!(11n)(12n)(1k1n).\binom{n}{k} = \frac{n(n-1)\cdots(n-k+1)}{k!} = \frac{n^k}{k!} \left(1 - \frac{1}{n}\right)\left(1 - \frac{2}{n}\right) \cdots \left(1 - \frac{k-1}{n}\right).

Step 3: Substitute pn=λ/n+o(1/n)p_n = \lambda/n + o(1/n) (since npnλnp_n \to \lambda):

P(Xn=k)=nkk!(λn)k(1pn)nk=0k1(1n).\mathbb{P}(X_n = k) = \frac{n^k}{k!} \left(\frac{\lambda}{n}\right)^k (1-p_n)^{n-k} \prod_{\ell=0}^{k-1} \left(1 - \frac{\ell}{n}\right).

Step 4: Simplify:

P(Xn=k)=λkk!(1λn)nk=0k1(1n)+o(1).\mathbb{P}(X_n = k) = \frac{\lambda^k}{k!} \left(1 - \frac{\lambda}{n}\right)^{n-k} \prod_{\ell=0}^{k-1} \left(1 - \frac{\ell}{n}\right) + o(1).

Step 5: Take the limit as nn \to \infty:

limn(1λn)n=eλ,limn(1λn)k=1,=0k1(1n)1.\lim_{n \to \infty} \left(1 - \frac{\lambda}{n}\right)^n = e^{-\lambda}, \quad \lim_{n \to \infty} \left(1 - \frac{\lambda}{n}\right)^{-k} = 1, \quad \prod_{\ell=0}^{k-1} \left(1 - \frac{\ell}{n}\right) \to 1.

Hence,

limnP(Xn=k)=λkk!eλ.\lim_{n \to \infty} \mathbb{P}(X_n = k) = \frac{\lambda^k}{k!} e^{-\lambda}.


General version: Poisson approximation for rare events

Theorem2.2General Poisson limit theorem

Let X1(n),X2(n),,Xn(n)X_1^{(n)}, X_2^{(n)}, \ldots, X_n^{(n)} be independent Bernoulli random variables with P(Xi(n)=1)=pi(n)\mathbb{P}(X_i^{(n)} = 1) = p_i^{(n)}. Suppose:

  1. i=1npi(n)λ\sum_{i=1}^n p_i^{(n)} \to \lambda as nn \to \infty.
  2. max1inpi(n)0\max_{1 \leq i \leq n} p_i^{(n)} \to 0 as nn \to \infty (individual probabilities vanish).

Then Sn=i=1nXi(n)dPoisson(λ)S_n = \sum_{i=1}^n X_i^{(n)} \xrightarrow{d} \text{Poisson}(\lambda).

This is the law of rare events: the sum of many independent, rare events (each with small probability) is approximately Poisson. The events need not be identically distributed.

ExampleTypos in a book

A book has n=500n = 500 pages. On page ii, the probability of a typo is pip_i (which may vary by page, depending on complexity). Suppose i=1500pi=10\sum_{i=1}^{500} p_i = 10 and maxipi<0.05\max_i p_i < 0.05. Then the total number of typos is approximately Poisson(10)\text{Poisson}(10).


Connection to the Poisson process

Theorem2.3Poisson process from binomial approximation

Consider a time interval [0,t][0, t] divided into nn subintervals of length h=t/nh = t/n. In each subinterval, an event occurs with probability pn=λh+o(h)=λt/n+o(1/n)p_n = \lambda h + o(h) = \lambda t/n + o(1/n), independently. As nn \to \infty, the number of events in [0,t][0, t] converges in distribution to Poisson(λt)\text{Poisson}(\lambda t).

This theorem provides a discrete approximation to the Poisson process: by partitioning time into many small intervals and allowing at most one event per interval (with probability proportional to the interval length), we obtain a Poisson process in the limit.

RemarkHistorical context

The Poisson distribution was introduced by Siméon Denis Poisson in 1837 as a limiting case of the binomial distribution. The Poisson limit theorem was later generalized by von Mises (1919) and Prokhorov (1953) to the law of rare events for non-identically distributed summands.


Rate of convergence

Theorem2.4Total variation distance bound

Let XBinomial(n,p)X \sim \text{Binomial}(n, p) with λ=np\lambda = np. Then

L(X)Poisson(λ)TVmin(p,2pλ).\|\mathcal{L}(X) - \text{Poisson}(\lambda)\|_{TV} \leq \min(p, 2p\lambda).

Here, μνTV=12kμkνk\|\mu - \nu\|_{TV} = \frac{1}{2} \sum_k |\mu_k - \nu_k| is the total variation distance.

This bound shows that the approximation is accurate when pp is small (the "rare event" regime). For example, if p0.01p \leq 0.01 and λ=np10\lambda = np \leq 10, the error is at most 0.010.01.

ExampleNumerical comparison

For n=100n = 100, p=0.05p = 0.05, λ=5\lambda = 5:

| kk | Binomial | Poisson | Error | |-----|----------|---------|-------| | 0 | 0.0059 | 0.0067 | 0.0008 | | 1 | 0.0312 | 0.0337 | 0.0025 | | 5 | 0.1800 | 0.1755 | 0.0045 | | 10 | 0.0167 | 0.0181 | 0.0014 |

The approximation is excellent across the range.


Applications

ExampleRadioactive decay

A radioactive substance has NN atoms. In a small time interval [0,t][0, t], each atom decays independently with probability p=λt/Np = \lambda t / N (where λ\lambda is the decay rate per atom). The number of decays is approximately Poisson(λt)\text{Poisson}(\lambda t) for large NN and small tt.

This justifies modeling radioactive decay as a Poisson process: the number of decay events in any time interval [0,t][0, t] is Poisson(λt)\text{Poisson}(\lambda t).

ExamplePhone calls

A telephone exchange receives calls from n=10,000n = 10{,}000 subscribers. In a one-minute interval, each subscriber calls with probability p=0.001p = 0.001, independently. The total number of calls is approximately Poisson(10)\text{Poisson}(10). This is the classical model for telephone traffic (Erlang, 1909).


Summary

The Poisson limit theorem explains why the Poisson distribution is ubiquitous:

  • Rare events: Sum of many independent, rare events \to Poisson.
  • Binomial approximation: Binomial(n,p)Poisson(np)\text{Binomial}(n, p) \approx \text{Poisson}(np) for large nn, small pp.
  • Poisson process: Discretizing time and taking limits yields the Poisson process.
  • Applications: Defects, arrivals, decays, calls, typos, accidents, etc.

The theorem is both a practical approximation tool and a conceptual foundation for continuous-time stochastic models.