This theorem is widely acclaimed in statistics, along with the Law of Large Numbers. Despite being frequently discussed and applied, many encounter its proof only upon studying mathematical statistics. However, the proof itself is interesting, making the theorem even more valuable.
Firstly, assume that the moment generating function M(t)=E(etY),−h<t<h of Y:=nσXn−μ exists. By defining a new function m(t):=E[et(X−μ)]=e−μtM(t),
M(t)======E(etnσXn−μ)E(etσn∑i=1nXi−nμ)E(etσnX1−μ)E(etσnX2−μ)⋯E(etσnXn−μ)E(etσnX−μ)E(etσnX−μ)⋯E(etσnX−μ){E(etσnX−μ)}n{m(σnt)}n,−h<σnt<h
Taylor’s theorem: If the function f(x) is continuous at [a,b] and differentiable up to n times at (a,b), then for x0∈(a,b), there exists ξ∈(a,b) such that f(x)=k=0∑n−1k!(x−x0)kf(k)(x0)+n!(x−x0)nf(n)(ξ) is satisfied.
Applying Taylor’s theorem to n=2 reveals that there exists ξ satisfying either (−t,0) or (0,t).
Hence, m(t) can be expressed as
m(t)=m(0)+m′(0)t+2m′′(ξ)t2
Meanwhile,
⎩⎨⎧m(0)=1m′(0)=E(X−μ)=0m′′(0)=E[(X−μ)2]=σ2
thus m(t)=1+2m′′(ξ)t2. Here comes the trick: by adding and then subtracting 2σ2t2 on the right side,
m(t)=1+2σ2t2+2[m′′(ξ)−σ2]t2
In other words,
M(t)={m(σnt)}n={1+2nt2+2nσ2[m′′(ξ)−σ2]t2}n
According to Taylor’s theorem, since ξ lies between (−σnt,0) and (0,σnt), when n→∞, ξ→0; thus, m′′(ξ)→m′′(0)=σ2. By eliminating the terms that converge to