This week we will see how moment generating functions (MGFs) lead to the Central Limit Theorem. For more detail, read Evans and Rosenthal, Chapter 4.
Last week we had a theorem that said that a random variable’s distribution is completely determined by its MGF. Another similar fact is the following:
If is a sequence of random variables with MGFs that converge to the MGF of a random variable , then the probability distribution of converges to the probability distribution of .
We worked out the following examples in class.
What is the MGF for a Binom(n,p) random variable?
Suppose that has a distribution for some fixed . What is the MGF for ?
Use the fact that to find the limit of the MGFs from #2.
What probability distribution corresponds to the MGF in #3?
Suppose that is any random variable with mean and variance . What are the first three moments , , and ?
Find the first 3 terms of the Maclaurin series for the MGF .
If are i.i.d. RVs with mean 0 and variance 1, then what are the mean and variance of ?
Let . Use homework 1, problem 5, to find the MGF for in terms of the shared MGF of the ’s.
Use Taylor’s Theorem to show that on any bounded interval containing , there is a constant such that
Last time we showed that if is a random variable with mean 0 and variance 1, that on any bounded interval containing 0 there is a constant such that
We can simplify the notation for this by using Big-O Notation.
If and are functions defined on an interval and there is a constant such that for all , then we say that .
In many applications, Big-O notation is used on intervals of the form to deal with the asymptotic behavior of functions near infinity. For our purposes, the interval will typically be a neighborhood around 0, that is, an interval of the form , so Big-O notation helps understand how a function behaves near zero.
If has derivatives up to order at , then in an interval around 0.
Re-write the inequality for above using Big-O notation.
Write the MGF for using Big-O notation to collect all the terms of order 3 or higher.
Suppose that are i.i.d. random variables with mean 0 and variance 1. Show that the MGF for converges to the MGF for a Norm(0,1) random variable as .
In order to do #3, you will need to calculate limits of the form One approach is to take the natural logarithm of the limit, and then use L’Hospital’s rule. Do this, and show that the limit is , no matter what is.
This argument leads to:
If are i.i.d. random variables with mean and variance , then the probability distribution of converges to the distribution as .
We finished the class by mentioning that MGFs are not always defined. An alternative proof of the central limit theorem uses characteristic functions instead of MGFs. These are similar to MGFs, but are defined using the imaginary number as . Characteristic functions are the Fourier transforms of the probability density function, and are always defined for every probability distribution, unlike MGFs.
We finished our discussion of MGFs today by looking at two examples.
The Laplace distribution has probability density function:
Using the definition of a MGF, the MGF for a Laplace distribution is:
Depending on the value of , that integral might be finite, or infinite, as we saw in the graphs below:
In particular, if , then the right side of does not decay, so the integral is infinite. The same problem happens on the left side when .
Calculate integral to find the MGF of the Laplace distribution.
Write the MGF of the Laplace distribution as a Maclaurin series. What are the 1st through 6th moments?
The reason that the Laplace distribution is not defined when is too big is that the tails of the distribution do not decay fast enough. You can get even worse behavior if you choose a distribution with tails that converge even slower.
The Cauchy distribution has probability density function:
Show that the MGF for the Cauchy distribution is not defined for any .
Change the Desmos graph above for the Laplace MGF to a graph showing the function that you would have to integrate to get the Cauchy MGF. Does the graph make it clear (if you zoom out far enough) why the integral won’t converge to a finite area?