What is the difference between the weak and strong law of large numbers?

34.9k Views Asked by At

I don't really understand exactly what the difference between the weak and strong law of large numbers is. The weak law says

\begin{align*} \lim_{n \rightarrow \infty} \mathbb{P}[\mid \bar{X}_n - \mu \mid \leq \epsilon ] = 1, \end{align*}

while the strong law reads as

\begin{align*} \mathbb{P}[\lim_{n \rightarrow \infty} \bar{X}_n = \mu ] = 1 \end{align*}

Isn't this a very subtle difference? Since I can chose my $\epsilon$ arbitrarily small I can write for $n \rightarrow \infty$

\begin{align*} \mid \bar{X}_n - \mu \mid \leq \epsilon \\ - \epsilon \leq \bar{X}_n - \mu \leq \epsilon \\ \mu - \epsilon \leq \bar{X}_n \leq \mu + \epsilon \end{align*}

Which of course means that as $\epsilon \approx 0$ should be the same as $\lim_{n \rightarrow \infty} \bar{X}_n = \mu$. So, in what sense are those conditions actually "different"?

Regarding the weak law I'd like to know if these are actually the same:

\begin{align*} \lim_{n \rightarrow \infty} \mathbb{P}[\mid \bar{X}_n - \mu \mid > \epsilon] = \mathbb{P}[ \mid \lim_{n \rightarrow \infty} \bar{X}_n - \mu \mid > \epsilon] \end{align*}

I ask because the weak law always gets written like the l.h.s. but the strong law always has $\lim_{n \rightarrow \infty}$ inside the probability operator ..

5

There are 5 best solutions below

3
On

The weak law of large numbers refers to convergence in probability, whereas the strong law of large numbers refers to almost sure convergence.

We say that a sequence of random variables $\{Y_n\}_{n=1}^{\infty}$ converges in probability to a random variable $Y$ if, for all $\epsilon>0$, $\lim_n P(|Y_n-Y|>\epsilon)=0$.

We say that a sequence of random variables $\{Y_n\}_{n=1}^{\infty}$ converges almost surely to a random variable $Y$ if $\lim_n Y_n(\omega)=Y(\omega)$ for almost every $\omega$, that is, $P(\{\omega:\lim_nY_n(\omega)=Y(\omega)\})=1$.

Almost sure convergence implies convergence in probability, but the converse is not true (that is why the laws of large numbers are called strong and weak respectively). To see that the converse is not true, just consider discrete random variables $Y_n$ satisfying $P(Y_n=1)=1/n$ and $P(Y_n=0)=1-1/n$. Given $0<\epsilon<1$, $P(|Y_n|\leq\epsilon)=p(Y_n=0)=1-1/n\rightarrow 1$, so $Y_n\rightarrow 0$ in probability. However, as $\sum_n P(Y_n=1)=\infty$, by Borel-Cantelly lemma we have that, for almost every $\omega$, $Y_n(\omega)=1$ for infinitely many $n$'s. The sequence $\{Y_n\}$ does not converge almost surely.

Concerning your reasoning, the fact that $\lim_nP(|\bar{X}_n-\mu|\leq\epsilon)=1$ does not imply that, for large $n$, $|\bar{X}_n-\mu|\leq\epsilon$. In my previous example, you do not have $|Y_n|\leq\epsilon$ for every large $n$, as $Y_n=1$ for infinitely many $n$'s.

0
On

For your last question (are those actually the same). Well we can show that they are the same using Lebesgue's dominated convergence theorem, only if we already know that almost surely $X_n$ has a limit. So we need the law of large numbers to prove the law of large numbers using your trick.

0
On

In general, integration and limits do not commute.

Since $\mathbb{P}$ is an integral under the Kolmogorov axioms, it's not something we can assume a priori.

0
On

The relative position of the symbol for probability with respect to the symbol for limit in their respective definitions:

$$ \begin{cases} \text{WLLN}: \color{blue}{\displaystyle \lim_{n\to \infty}}\color{red}{\Pr}\left(\vert \bar X_n - \mathbb E[X] \vert < \varepsilon \right) = 1\\[3ex] \text{SLLN}: \color{red}{\Pr}\left(\color{blue}{\displaystyle \lim_{n\to \infty}} \vert \bar X_n - \mathbb E[X]\vert =0\right) = 1 \end{cases}$$

Hence, in the WLLN, we are contemplating a sequence of probabilities indexed by the sample size: $\{\Pr_1, \Pr_2,\dots,\Pr_n \}:$ for any infinitesimally small $\varepsilon$ we may choose, the probability that the difference between sample mean and population mean (or expectation of the random variable) is even smaller than the chosen $\varepsilon$ will form a sequence of probability values, which will increase with $n$ because the sample size will approach the population size. This sequence converges to $1.$

In the SLLN, the concept is a sequence of differences between the sample mean of increasing size and the population mean. This difference vanishes to $0$ as $n$ goes to infinity with probability $1,$ or almost surely.

In a way, it is the difference between the assurance that something does happen (SLLN) versus the assurance that what we are after will happen with increasing probability (WLLN), accounting for the fact that SLLN $\implies$ WLLN, but not the other way around.

0
On

The weak law describes how a sequence of probabilities converges, and the strong law describes how a sequence of random variables behaves in the limit.