ConceptComplete

Common Distributions - Core Definitions

Probability distributions form families characterized by parameters that control their shape, location, and scale. Understanding these distribution families is essential for statistical modeling.

The Normal (Gaussian) Distribution

Definition

XN(μ,σ2)X \sim \mathcal{N}(\mu, \sigma^2) has PDF: f(x)=12πσ2exp((xμ)22σ2),xRf(x) = \frac{1}{\sqrt{2\pi\sigma^2}} \exp\left(-\frac{(x-\mu)^2}{2\sigma^2}\right), \quad x \in \mathbb{R}

Parameters: μ\mu (mean), σ2>0\sigma^2 > 0 (variance)

The normal distribution is the most important continuous distribution, arising naturally from the Central Limit Theorem. The standard normal ZN(0,1)Z \sim \mathcal{N}(0,1) has CDF denoted Φ(z)\Phi(z).

Properties:

  • Symmetric about μ\mu
  • Bell-shaped curve
  • 68-95-99.7 rule: Approximately 68%, 95%, 99.7% of mass within 1, 2, 3 standard deviations
  • MGF: MX(t)=eμt+σ2t2/2M_X(t) = e^{\mu t + \sigma^2 t^2/2}

The Binomial Distribution

Definition

XBinomial(n,p)X \sim \text{Binomial}(n,p) represents successes in nn independent Bernoulli(p)(p) trials: P(X=k)=(nk)pk(1p)nk,k=0,1,,nP(X = k) = \binom{n}{k} p^k (1-p)^{n-k}, \quad k = 0, 1, \ldots, n

Parameters: n{1,2,3,}n \in \{1,2,3,\ldots\} (trials), p[0,1]p \in [0,1] (success probability)

Properties:

  • E[X]=npE[X] = np, Var(X)=np(1p)\text{Var}(X) = np(1-p)
  • MGF: MX(t)=(pet+1p)nM_X(t) = (pe^t + 1-p)^n
  • Sum of independent Binomial(ni,p)(n_i, p) is Binomial(ni,p)(\sum n_i, p) (same pp)
  • Normal approximation: For large nn, XN(np,np(1p))X \approx \mathcal{N}(np, np(1-p))

The Poisson Distribution

Definition

XPoisson(λ)X \sim \text{Poisson}(\lambda) models rare events occurring at rate λ>0\lambda > 0: P(X=k)=eλλkk!,k=0,1,2,P(X = k) = \frac{e^{-\lambda} \lambda^k}{k!}, \quad k = 0, 1, 2, \ldots

Parameter: λ>0\lambda > 0 (rate)

Properties:

  • E[X]=Var(X)=λE[X] = \text{Var}(X) = \lambda
  • MGF: MX(t)=eλ(et1)M_X(t) = e^{\lambda(e^t - 1)}
  • Sum of independent Poisson(λi)(\lambda_i) is Poisson(λi)(\sum \lambda_i)
  • Approximates Binomial(n,p)(n,p) when nn large, pp small, np=λnp = \lambda

The Exponential Distribution

Definition

XExponential(λ)X \sim \text{Exponential}(\lambda) models waiting times with rate λ>0\lambda > 0: f(x)=λeλx,x0f(x) = \lambda e^{-\lambda x}, \quad x \geq 0

Parameter: λ>0\lambda > 0 (rate)

Properties:

  • E[X]=1/λE[X] = 1/\lambda, Var(X)=1/λ2\text{Var}(X) = 1/\lambda^2
  • CDF: F(x)=1eλxF(x) = 1 - e^{-\lambda x} for x0x \geq 0
  • Memoryless: P(X>s+tX>s)=P(X>t)P(X > s+t | X > s) = P(X > t)
  • MGF: MX(t)=λλtM_X(t) = \frac{\lambda}{\lambda - t} for t<λt < \lambda
  • Minimum of independent Exponential(λi)(\lambda_i) is Exponential(λi)(\sum \lambda_i)

The Gamma Distribution

Definition

XGamma(α,λ)X \sim \text{Gamma}(\alpha, \lambda) generalizes the exponential: f(x)=λαΓ(α)xα1eλx,x0f(x) = \frac{\lambda^{\alpha}}{\Gamma(\alpha)} x^{\alpha-1} e^{-\lambda x}, \quad x \geq 0

Parameters: α>0\alpha > 0 (shape), λ>0\lambda > 0 (rate)

where Γ(α)=0tα1etdt\Gamma(\alpha) = \int_0^{\infty} t^{\alpha-1} e^{-t} dt

Properties:

  • E[X]=α/λE[X] = \alpha/\lambda, Var(X)=α/λ2\text{Var}(X) = \alpha/\lambda^2
  • Exponential(λ)=Gamma(1,λ)(\lambda) = \text{Gamma}(1, \lambda)
  • Sum of nn independent Exponential(λ)(\lambda) is Gamma(n,λ)(n, \lambda)
  • Chi-squared(ν)=Gamma(ν/2,1/2)(\nu) = \text{Gamma}(\nu/2, 1/2)
Remark

These distributions are not arbitrary—they arise naturally from physical processes. Normal from sums (CLT), Poisson from rare events, Exponential from memoryless waiting times, and Gamma from sums of exponentials. Understanding their origins helps in choosing appropriate models.