Random Variables and Distributions - Main Theorem
The Law of the Unconscious Statistician provides a fundamental method for computing expectations of transformed random variables without explicitly finding the distribution of the transformation.
Law of the Unconscious Statistician (LOTUS)
Let be a random variable and a function. Then:
Discrete case: If has PMF :
Continuous case: If has PDF :
provided the sum or integral converges absolutely.
The theorem states that we can compute directly from the distribution of without first deriving the distribution of .
Proof Sketch (Continuous Case): Let . By definition:
Through change of variables and the transformation formula for PDFs, this equals: β‘
Let (standard normal). Find .
Using LOTUS:
By integration by parts or recognizing this as the variance of a standard normal:
Without LOTUS, we would need to first find the distribution of (a chi-squared distribution), then compute its expectationβmuch more work!
Applications of LOTUS
Computing Moments: For the -th moment of :
\sum_x x^k p_X(x) & \text{(discrete)} \\ \int_{-\infty}^{\infty} x^k f_X(x) \, dx & \text{(continuous)} \end{cases}$$ **Variance**: Using $\text{Var}(X) = E[X^2] - (E[X])^2$: $$\text{Var}(X) = E[X^2] - \mu^2 = \int (x-\mu)^2 f_X(x) \, dx$$ <Example> For $X \sim \text{Exponential}(\lambda)$ with $f_X(x) = \lambda e^{-\lambda x}$ for $x \geq 0$: $$E[X] = \int_0^{\infty} x \cdot \lambda e^{-\lambda x} \, dx = \frac{1}{\lambda}$$ $$E[X^2] = \int_0^{\infty} x^2 \cdot \lambda e^{-\lambda x} \, dx = \frac{2}{\lambda^2}$$ $$\text{Var}(X) = \frac{2}{\lambda^2} - \frac{1}{\lambda^2} = \frac{1}{\lambda^2}$$ </Example> ## Moment Generating Function <Definition> The **moment generating function** (MGF) of $X$ is: $$M_X(t) = E[e^{tX}]$$ when this expectation exists for $t$ in some neighborhood of 0. </Definition> By LOTUS: $$M_X(t) = \begin{cases} \sum_x e^{tx} p_X(x) & \text{(discrete)} \\ \int_{-\infty}^{\infty} e^{tx} f_X(x) \, dx & \text{(continuous)} \end{cases}$$ The MGF "generates moments": $E[X^k] = M_X^{(k)}(0)$ (the $k$-th derivative at 0). <Example> For $X \sim \mathcal{N}(\mu, \sigma^2)$: $$M_X(t) = e^{\mu t + \sigma^2 t^2/2}$$ Taking derivatives: - $E[X] = M_X'(0) = \mu$ β - $E[X^2] = M_X''(0) = \mu^2 + \sigma^2$, so $\text{Var}(X) = \sigma^2$ β </Example> <Remark> LOTUS is one of the most frequently used results in probability. Its name derives from the fact that early statisticians would apply it "unconsciously"βcomputing $E[g(X)]$ without realizing they were using a theorem. The MGF is particularly powerful because it uniquely determines a distribution and simplifies calculations involving sums of independent random variables. </Remark>