Limit laws: weak law of large numbers
Published:
This is the first of a series of entries where we will explore several limit laws for sequences of random variables. The setting is going to be the standard for probability theory: fix a measure space Ω, a sigma-algebra B and a probability measure P. By random variables we mean measurable functions X:Ω→R. The distribution of X is defined as F(x)=P(ω:X(ω)≤x), so P(a≤X≤b)=F(b)−F(a). If the measure X−1P(A)=P(X−1(A)) is absolutely continuous with respect to P, we denote the Radon-Nikodym derivative by f=dX−1PdP and hence we can compute probabilities as integrals of this function with respect to the measure P:
P(X∈A)=∫Af(x)dLeb(x).We call f the density of the random variable X. The expectation (mean) of X is defined by μ=E(X)=∫ΩXdP=∫RxdF(x) where the second integral is in the sense of Lebesgue-Stieltjes, and the variance by σ2=∫ωX2dP−(E(X))2=E(X−μ)2.
Now we suppose we have a sequence of iid random variables X1,X2,… with distribution F. We are interested in describing the limit behavior of the averages
Snn=X1+⋯+Xnn,where Sn=X1+⋯+Xn. Intuitively, the averaging of the independent observations of the same distribution should smooth out and regularize the random fluctuations of the random variables around the mean. This is the content of the weak law of large numbers:
Theorem (Weak law of large numbers): Assume that the iid sequence (Xn) has finite variance. Then (Xn) converges in distribution to μ.
Recall that convergence in distribution means that for any ε>0,
P(|Snn−μ|>ε)→0as n→∞.
This result can be interpreted in the following way: we fix a precision threshold ε, and as we take more measurements of the same distributions, the probability that the average is close to the mean within ε precision becomes arbitrarily small.
We have formulated here the result in a slightly more restricted formulation, as it hold under more general conditions where the variance may not necessarily exist, but in that case the proof is more involved. When the variance is finite, the result can be proven using a concentration inequality:
Theorem (Chebyshev’s inequality): Assume that the random variable X has finite variance σ2 and mean μ. Then for every positive number ε>0 we have that
P(|X−μ|≥ε)≤σ2ε2.This inequality follows immediately from another concentration inequality:
Theorem (Markov’s inequality): Assume that the non-negative random variable X has finite mean μ. Then for every positive number a>0 we have that
P(X≥a)≤μa.Proof:
Let I(X≥a) the indicator function of the set (X≥a). Then
a⋅I(X≥a)≤Xalmost surely. Taking expectations we get
P(X≥a)≤μaas we wanted.◻
The proof of Chebyshev’s inequality follows immediately by using the random variable (X−μ)2 and the threshold ε2. Now the weak law of large numbers follows easily from Chebyshev’s inequality if we note that the variance of Sn/n is equal to σ2/n. Then, given δ>0, take n0 such that n0≥σ2/(δε2), then
P(|Snn−μ|>ε)≤σ2nε2≤δfor all n≥n0, which proves the statement of the WLLN. ◻
Although this version of the WLLN is not the most general, it is interesting because it shows how concentration inequalities can be used to derive limit laws for sums.
Leave a Comment