← back to probability

Law of Large Numbers

Grinstead & Snell · GFDL · PDF

The law of large numbers says: the sample mean converges to the true mean as the sample grows. Flip a fair coin a million times. The fraction of heads will be close to 0.5. Not because outcomes balance out, but because the noise gets drowned by the signal.

Chebyshev's inequality

Before proving the LLN, Grinstead and Snell establish the key tool: Chebyshev's inequality. For any random variable X with mean mu and variance sigma^2: P(|X - mu| ≥ k sigma) ≤ 1/k^2. The probability of being far from the mean is bounded by the variance. No distributional assumptions needed.

μ 1.0 0.0 n (sample size) 10 100 1000 10000 μ = 0.5 all paths converge to μ Sample mean trajectories (fair coin)
Scheme

Weak law of large numbers

Let X_1, X_2, ..., X_n be independent with the same distribution, mean mu, and finite variance sigma^2. The sample mean S_n/n satisfies: for any epsilon > 0, P(|S_n/n - mu| ≥ epsilon) ≤ sigma^2 / (n epsilon^2). As n grows, this bound goes to zero. The sample mean converges in probability to the true mean.

Scheme

Bernoulli's theorem

The LLN was first proved by Jacob Bernoulli in 1713, specifically for coin flips. His version: the fraction of heads in n flips of a coin with probability p converges to p. This is the special case of the weak LLN applied to Bernoulli trials (Ch 5). Bernoulli called it the "golden theorem." It took him twenty years to prove.

Scheme

What LLN does not say

The LLN does not say outcomes "balance out." After 10 heads in a row, the next flip is still 50/50. What happens is dilution: those 10 heads become a smaller fraction of the total as n grows. The law works by addition, not correction. The wpgambler's fallacy confuses these two.

Scheme

Notation reference

Notation Scheme Meaning
S_n / n(/ sum n)Sample mean
P(|X-μ| ≥ kσ) ≤ 1/k²(chebyshev-bound k)Chebyshev's inequality
σ²/(nε²)(lln-bound var n eps)Weak LLN convergence bound
X_n → μ in probabilitybound → 0 as n → ∞Convergence in probability
Neighbors

Probability chapters

  • 🎰 Ch 7 — sums of random variables (the machinery behind LLN)
  • 🎰 Ch 6 — expected value and variance (the quantities LLN connects)
  • 🎰 Ch 5 — Bernoulli distribution (the original LLN setting)

Foundations (Wikipedia)

Ready for the real thing? Read Grinstead & Snell, Chapter 8.

jkThe Handshake