The Bedrock of Statistics
The law of large numbers is perhaps the most important theorem in probability. It states a simple but profound truth: as you collect more data, the average converges to the true expected value. Flip a fair coin 10 times and you might get 70% heads. Flip it 10,000 times and you'll be within a fraction of a percent of 50%. This convergence is not just likely — it's mathematically guaranteed for any distribution with a finite mean.
Convergence at 1/√n
The rate of convergence follows a precise mathematical law: the standard error of the mean equals σ/√n, where σ is the standard deviation and n is the sample size. This means quadrupling your data halves the uncertainty. To get 10× more precision, you need 100× more data. This square-root scaling is why political polls survey thousands (not millions) of people — 1,000 respondents give about 3% margin of error.
Watching Convergence
The simulation plots the running average as samples accumulate. Initially the line is volatile, jumping with each new sample. As more data arrives, the oscillations shrink and the average settles toward the dashed true-mean line. The shaded confidence bounds (±σ/√n) narrow visibly, funneling the average toward convergence. Try different distributions to see that convergence works for all of them.
Why It Matters Everywhere
The law of large numbers is why insurance companies can predict total claims, why casinos always win in the long run, and why scientific experiments are repeatable. It's the mathematical foundation for polling, clinical trials, quality control, and machine learning. Without it, collecting data would be pointless — no amount of evidence would bring us closer to the truth.