ConceptComplete

Expectation and Variance - Examples and Constructions

Computing expectations and variances for standard distributions reveals patterns and provides reference values for applications.

Common Discrete Distributions

Bernoulli(p)(p): E[X]=p,Var(X)=p(1p)E[X] = p, \quad \text{Var}(X) = p(1-p)

Binomial(n,p)(n,p): Sum of nn independent Bernoulli(p)(p) trials: E[X]=np,Var(X)=np(1p)E[X] = np, \quad \text{Var}(X) = np(1-p)

Geometric(p)(p): Number of trials until first success: E[X]=1p,Var(X)=1pp2E[X] = \frac{1}{p}, \quad \text{Var}(X) = \frac{1-p}{p^2}

Poisson(λ)(\lambda): E[X]=λ,Var(X)=λE[X] = \lambda, \quad \text{Var}(X) = \lambda

Remarkably, for Poisson the mean equals the variance!

Example

For Binomial(100,0.3)(100, 0.3): E[X]=100×0.3=30E[X] = 100 \times 0.3 = 30 Var(X)=100×0.3×0.7=21\text{Var}(X) = 100 \times 0.3 \times 0.7 = 21 σ=214.58\sigma = \sqrt{21} \approx 4.58

By Chebyshev: At least 75% of outcomes lie in [302(4.58),30+2(4.58)]=[20.84,39.16][30 - 2(4.58), 30 + 2(4.58)] = [20.84, 39.16].

Common Continuous Distributions

Uniform(a,b)(a,b): E[X]=a+b2,Var(X)=(ba)212E[X] = \frac{a+b}{2}, \quad \text{Var}(X) = \frac{(b-a)^2}{12}

Exponential(λ)(\lambda): E[X]=1λ,Var(X)=1λ2E[X] = \frac{1}{\lambda}, \quad \text{Var}(X) = \frac{1}{\lambda^2}

Normal(μ,σ2)(\mu, \sigma^2): E[X]=μ,Var(X)=σ2E[X] = \mu, \quad \text{Var}(X) = \sigma^2

The parameters directly give the mean and variance!

Gamma(α,λ)(\alpha, \lambda): Generalizes exponential (α=1\alpha = 1): E[X]=αλ,Var(X)=αλ2E[X] = \frac{\alpha}{\lambda}, \quad \text{Var}(X) = \frac{\alpha}{\lambda^2}

Example

Standardization: For any XX with mean μ\mu and variance σ2\sigma^2, define: Z=XμσZ = \frac{X - \mu}{\sigma}

Then E[Z]=0E[Z] = 0 and Var(Z)=1\text{Var}(Z) = 1. This standardized variable has mean 0 and variance 1.

If XN(μ,σ2)X \sim \mathcal{N}(\mu, \sigma^2), then ZN(0,1)Z \sim \mathcal{N}(0,1) (standard normal).

Computing Variance via MGF

Example

For XPoisson(λ)X \sim \text{Poisson}(\lambda), the MGF is: MX(t)=eλ(et1)M_X(t) = e^{\lambda(e^t - 1)}

Derivatives: MX(t)=λeteλ(et1)M_X'(t) = \lambda e^t \cdot e^{\lambda(e^t - 1)} MX(t)=λeteλ(et1)(λet+1)M_X''(t) = \lambda e^t \cdot e^{\lambda(e^t - 1)} \cdot (\lambda e^t + 1)

At t=0t = 0: E[X]=MX(0)=λE[X] = M_X'(0) = \lambda E[X2]=MX(0)=λ(λ+1)=λ2+λE[X^2] = M_X''(0) = \lambda(\lambda + 1) = \lambda^2 + \lambda Var(X)=E[X2](E[X])2=λ2+λλ2=λ\text{Var}(X) = E[X^2] - (E[X])^2 = \lambda^2 + \lambda - \lambda^2 = \lambda

Sums of Independent Random Variables

Theorem: If X1,,XnX_1, \ldots, X_n are independent: E[i=1nXi]=i=1nE[Xi]E\left[\sum_{i=1}^n X_i\right] = \sum_{i=1}^n E[X_i] Var(i=1nXi)=i=1nVar(Xi)\text{Var}\left(\sum_{i=1}^n X_i\right) = \sum_{i=1}^n \text{Var}(X_i)

Example

If X1,,X100X_1, \ldots, X_{100} are independent Exponential(2)(2) random variables: S=i=1100XiS = \sum_{i=1}^{100} X_i

has: E[S]=100×12=50E[S] = 100 \times \frac{1}{2} = 50 Var(S)=100×14=25\text{Var}(S) = 100 \times \frac{1}{4} = 25 σS=5\sigma_S = 5

In fact, SGamma(100,2)S \sim \text{Gamma}(100, 2).

Remark

Linearity of expectation holds universally, even for dependent variables. Variance additivity requires independence. These formulas are the workhorses of probability calculations—they allow us to build complex distributions from simple building blocks.