STT 861 Theory of Prob and STT I Lecture Note - 14
2017-12-06
Proof of central limit theorem; multivariate normal distribution and its example; review of the second half of the semester, Poisson approximation vs. normal approximation to binomial distribution; miscellaneous notes of convergence.
Portal to all the other notes
- Lecture 01 - 2017.09.06
- Lecture 02 - 2017.09.13
- Lecture 03 - 2017.09.20
- Lecture 04 - 2017.09.27
- Lecture 05 - 2017.10.04
- Lecture 06 - 2017.10.11
- Lecture 07 - 2017.10.18
- Lecture 08 - 2017.10.25
- Lecture 09 - 2017.11.01
- Lecture 10 - 2017.11.08
- Lecture 11 - 2017.11.15
- Lecture 12 - 2017.11.20
- Lecture 13 - 2017.11.29
- Lecture 14 - 2017.12.06 -> This post
Lecture 14 - Dec 06 2017
Proof of CLT (Central Limit Theorem)
This is the last lecture of this course.
Let be i.i.d distribution. , . WOLOG (without loss of generality), , .
It is sufficient to show that MGF of MGF of for all near 0.
where .
We will use the following fundamental property of MGF.
In our case: .
Then expand using Taylor’s formula.
(This is for near 0. )
Therefore,
for small .
Next,
When ,
which is the MGF of N(0,1).
Calculation of the MGF of :
(this part is in the homework 6.)
Multivariate Normal
Quick review from a new example.
Example 1
Let be a normal vector. Assume and . Find the MGF of the entire vector.
Let . Since is multivariate normal, then is normal. We just need to compute the expectation and variance of .
where .
We have proved that
where , . Finally, the full form of MGF of is:
We can identify what it means for multivariate normal vectors to be independent of each other.
with length , then if is independent of ,
This also implies that if a normal vector has a block diagonal covariance matrix, then the corresponding two subvectors are independent.
Really Important Example - Example 2
Let be i.i.d . Let (empirical mean)
- what is the covariance ?
Answer:
- Forget the above one. Next, what is (this is the actual important question)?
this proves that the bivariate is bivariate normal since the original is multivariate normal and also and are independent because their covariance is 0.
- Next part of the example, let empirical variance . What’s the relation between and ?
Since and are independent for fixed , this is still true of and .
Therefore, the entire vector is independent of . In fact, any function of is independent of . In particular is independent of .
Also,
This is all i.i.d.
This is Chi-Square distribution .
Review of the Second Part of the Semester
When to use Poisson and when to use Normal to approximate a series of Bernoulli trials?
Poisson Approximation Case
i.i.d. For example, vote for Dr. Jill Stein (or Gary Johnson), otherwise.
is small. What if where is the # of trials?
Then ,
Since average # of successes is constant then . Range .
Note: . If then still holds.
Normal Approximation Case
it works for Binomial and many other partial sums’’ (like Negative Binomial).
In Binomial case, assume and is fixed. here, not constant.
Let
Let
then , by CLT.
Note: Unlike Poisson approximation, CLT (Normal approximation) works for any i.i.d as long as .
in distribution for all
( is a sequence of random variables, is a random variable)
Quick notes:
-
If is not continuous, the convergence does not need to hold for those ’s where is not continuous.
-
If there exists such that for all then does not converges in distribution, because there is no r.v. such that all the time.
-
Convergence in probability: in probability if and all live in the same probability space because we need to evaluate . And we ask, for any no matter how small.
-
is really small when .
-
almost surely if .