MGFs, the Normal, and the CLT
MGFs, the Normal, and the CLT¶
Let Z be standard normal. Then the mgf of Z is given by
MZ(t) = et2/2 for all tTo see this, just work out the integral:
MZ(t) = ∫∞−∞etz1√2πe−12z2dz= ∫∞−∞1√2πe−12(z2−2tz)dz= et2/2∫∞−∞1√2πe−12(z2−2tz+t2)dz= et2/2∫∞−∞1√2πe−12(z−t)2dz= et2/2because the integral is 1. It is the normal (t,1) density integrated over the whole real line.
Normal (μ,σ2)¶
It's a good idea to first note that moment generating functions behave well under linear transformations.
MaX+b(t) = E(et(aX+b)) = ebtE(eatX) = ebtMX(at)Since a normal (μ,σ2) variable can be written as σZ+μ where Z is standard normal, its m.g.f. is
MσZ+μ(t) = eμtMZ(σt) = eμt+σ2t2/2Details aside, what this formula is saying is that if a moment generating function is exp(c1t+c2t2) for any constant c1 and any positive constant c2, then it is the moment generating function of a normally distributed random variable.
Sums of Independent Normal Variables¶
We can now show that sums of independent normal variables are normal.
Let X have normal (μX,σ2X) distribution, and let Y independent of X have normal (μY,σ2Y) distribution. Then
MX+Y(t) = eμXt+σ2Xt2/2⋅eμYt+σ2Yt2/2 = e(μX+μY)t+(σ2X+σ2Y)t2/2That's the m.g.f. of the normal distribution with mean μX+μY and variance σ2X+σ2Y.
"Proof" of the Central Limit Theorem¶
Another important reason for studying mgf's is that they can help us identify the limit of a sequence of distributions.
The main example of convergence that we have seen is the Central Limit Theorem. Now we can indicate a proof.
Let X1,X2,… be i.i.d. random variables with expectation μ and SD σ. For every n≥1 let Sn=X1+X2+⋯+Xn.
The Central Limit Theorem says that for large n, the distribution of the standardized sum
S∗n = Sn−nμ√nσis approximately standard normal.
To show this, we will assume a major result whose proof is well beyond the scope of this class. Suppose Y1,Y2,… are random variables and we want to show that the the distribution of the Yn's converges to the distribution of some random variable Y. The result says that it is enough to show that the mgf's of the Yn's converge to the mgf of Y.
The result requires a careful statement and the proof requires considerable attention to detail. We won't go into that in this course. Instead we'll just point out that it should seem reasonable. Since mgf's determine distributions, it's not difficult to accept that if two mgf's are close to each other then the corresponding distributions should also be close to each other.
Let's use this result to "prove" the CLT. The quotes are because we will use the above result without proof, and also because the argument below involves some hand-waving about approximations.
First, write the standardized sum in terms of the standardized X's.
S∗n = Sn−nμ√nσ = n∑i=11√n(Xi−μσ) = n∑i=11√nX∗iwhere for each i, the random variable X∗i is Xi in standard units.
The random variables X∗i are i.i.d., so let MX∗ denote the mgf of any one of them. By the linear transformation property proved above, the mgf of each 1√nX∗i is given by
M1√nX∗i(t) = MX∗(t/√n)Therefore
MS∗n(t) = (MX∗(t/√n))n= (1 + t√n⋅E(X∗)1! + t2n⋅E(X∗2)2! + t3n3/2⋅E(X∗3)3! + ⋯)n≈ (1 + t22n)n for large nby ignoring small terms and using the fact that for any standardized random variable X∗ we have E(X∗)=0 and E(X∗2)=1.
Thus for large n,
MS∗n(t) ≈ (1 + t22n)n → et22 as n→∞The limit is the moment generating function of the standard normal distribution.