13. The central limit theorem#
Recommended reference: Wasserman [Was04], Sections 5.3–5.4.
The central limit theorem is a very important result in probability theory. It tell us that when we have \(n\) independent and identically distributed random variables, the distribution of their average (up to suitable shifting and rescaling) approximates a normal distribution as \(n\to\infty\).
13.1. Sample mean#
Consider a sequence of independent random variables \(X_1, X_2, \ldots\) distributed according to the same distribution function (which can be discrete or continuous). We say \(X_1, X_2, \ldots\) are independent and identically distributed (or i.i.d.).
Note
We saw the notion of independence of two random variables in Definition 11.6. For a precise definition of independence of an arbitrary number of random variables, see Section 2.9 in Wasserman [Was04].
Definition 13.1
The \(n\)-th sample mean of \(X_1,X_2,\ldots\) is
Note that \(\overline{X}_n\) is itself again a random variable.
13.2. The law of large numbers#
As above, consider i.i.d. samples \(X_1, X_2, \ldots\), say with distribution function \(f\). We assume that \(f\) has finite mean \(\mu\). The law of large numbers says that the \(n\)-th sample average is likely to be close to \(\mu\) for sufficiently large \(n\).
Theorem 13.1 (Law of large numbers)
For every \(\epsilon>0\), the probability
tends to \(0\) as \(n\to\infty\).
Note
The above formulation is known as the weak law of large numbers; there are also stronger versions, but the differences are not important here.
Fig. 13.1 illustrates the law of large numbers for the average of the first \(n\) out of 1000 dice rolls.