MGFs, the Normal, and the CLT

Interact

MGFs, the Normal, and the CLT

Let Z be standard normal. Then the mgf of Z is given by

MZ(t) = et2/2   for all t

To see this, just work out the integral:

MZ(t) = etz12πe12z2dz= 12πe12(z22tz)dz= et2/212πe12(z22tz+t2)dz= et2/212πe12(zt)2dz= et2/2

because the integral is 1. It is the normal (t,1) density integrated over the whole real line.

Normal (μ,σ2)

It's a good idea to first note that moment generating functions behave well under linear transformations.

MaX+b(t) = E(et(aX+b)) = ebtE(eatX) = ebtMX(at)

Since a normal (μ,σ2) variable can be written as σZ+μ where Z is standard normal, its m.g.f. is

MσZ+μ(t) = eμtMZ(σt) = eμt+σ2t2/2

Details aside, what this formula is saying is that if a moment generating function is exp(c1t+c2t2) for any constant c1 and any positive constant c2, then it is the moment generating function of a normally distributed random variable.

Sums of Independent Normal Variables

We can now show that sums of independent normal variables are normal.

Let X have normal (μX,σ2X) distribution, and let Y independent of X have normal (μY,σ2Y) distribution. Then

MX+Y(t) = eμXt+σ2Xt2/2eμYt+σ2Yt2/2 = e(μX+μY)t+(σ2X+σ2Y)t2/2

That's the m.g.f. of the normal distribution with mean μX+μY and variance σ2X+σ2Y.

"Proof" of the Central Limit Theorem

Another important reason for studying mgf's is that they can help us identify the limit of a sequence of distributions.

The main example of convergence that we have seen is the Central Limit Theorem. Now we can indicate a proof.

Let X1,X2, be i.i.d. random variables with expectation μ and SD σ. For every n1 let Sn=X1+X2++Xn.

The Central Limit Theorem says that for large n, the distribution of the standardized sum

Sn = Snnμnσ

is approximately standard normal.

To show this, we will assume a major result whose proof is well beyond the scope of this class. Suppose Y1,Y2, are random variables and we want to show that the the distribution of the Yn's converges to the distribution of some random variable Y. The result says that it is enough to show that the mgf's of the Yn's converge to the mgf of Y.

The result requires a careful statement and the proof requires considerable attention to detail. We won't go into that in this course. Instead we'll just point out that it should seem reasonable. Since mgf's determine distributions, it's not difficult to accept that if two mgf's are close to each other then the corresponding distributions should also be close to each other.

Let's use this result to "prove" the CLT. The quotes are because we will use the above result without proof, and also because the argument below involves some hand-waving about approximations.

First, write the standardized sum in terms of the standardized X's.

Sn = Snnμnσ = ni=11n(Xiμσ) = ni=11nXi

where for each i, the random variable Xi is Xi in standard units.

The random variables Xi are i.i.d., so let MX denote the mgf of any one of them. By the linear transformation property proved above, the mgf of each 1nXi is given by

M1nXi(t) = MX(t/n)

Therefore

MSn(t) = (MX(t/n))n= (1 + tnE(X)1! + t2nE(X2)2! + t3n3/2E(X3)3! + )n (1 + t22n)n   for large n

by ignoring small terms and using the fact that for any standardized random variable X we have E(X)=0 and E(X2)=1.

Thus for large n,

MSn(t)  (1 + t22n)n  et22  as n

The limit is the moment generating function of the standard normal distribution.

 

results matching ""

    No results matching ""