My Avatar

LanternD's Castle

An electronics enthusiast - survive technically

STT 861 Theory of Prob and STT I Lecture Note - 8

2017-10-25

Proof of the biased sample mean; Example 3.5.3 in the text book; Normal distribution, joint normal distribution (multivariate), Gamma distribution.

Portal to all the other notes

Lecture 08 - Oct 25 2017

For video record

(This part is similar to previous note. We recorded a video for it so the professor talked about it once again.)

Basically it is the prove that sample variance σ^2=1ni=1n(xix¯)2 biased.

Recall,

x¯=1ni=1nxi σ^2=1ni=1n(xix¯)2

Note that x¯ is the expectation of a r.v. X^ which takes the value xi with probability = 1n.

Therefore

x¯=E(X^)

and

Var(X^)=σ^2

Now recall the formula:

E((Xc))2=Var(X)+(E(X)c)2 Var(X)=E((Xc)2)(E(X)c)2

Use this with X=X^ defined above:

σ^2=1ni=1n(xic)2(x¯c)

Now the question is: what is the bias of σ^2? For this we replace each xi by Xi, where Xi’s are i.i.d with mean =μ and variance=σ2. The resulting expression is

σ^2(X)=1ni=1n(XiX¯)2

We are interested in E(σ^2(X)). We want to know if this =σ2 or not.

Now use formula above with C=μ and x=X and we take E on both side.

E(σ^2)=E(1ni=1n(Xiμ)2)E((X¯μ)2)=E(1ni=1n(Xiμ)2)Var(X¯)=1ni=1nVar(Xi)1nVar(xi)=1ni=1nVar(Xi)σ2n=σ21nσ2

We proved that E(σ^2)=(11nσ2). It is biased with 1nσ2.

Example 3.5.3, Page 100

(X,Y) has density f(x,y) is 2 if 0yy1, and 0 otherwise.

First compute the joint CDF of (X,Y),

General definition

F(u,v)=P(Xu,Yv)=x(yf(x,y)dy)dx=0x(0yf(x,y)dy)dx

If uv, we integrate y between 0 and x,

0vf(x,y)dy=0min(v,x)dy=2min(v,x)

Now we integrates between 0 and u with respect to x.

0u2min(v,x)dx=0v2min(v,x)dx+uv2min(v,x)dx=v22v(uv)

So we have proved that when u>v,

F(u,v)=v2+2v(uv)

If u<v, we still compute the same integral

0u2min(v,x)dx=u2

(0<u<v)

This proven F(u,v)=u2 when u<v.

Compute the marginal density of x and y.

In general, marginal density fX(x)=xF(x,y) when fixing y. Similarly, fY(y)=xF(x,y) where x is fixed.

If x>y,

fX(x)=2y(constant)

If x<y, then fX(x)=2x.

fX(x)=2x

Special continuous distributions - Chapter 4

Definition: X is a normal with parameters 0 and 1 if it has this density

f(x)=12πex2/2

Facts:

Notation: XN(0,1).

Definition: X is normal with parameters μ and σ2 if it has the density

f(x)=12πσ2exp(12σ2(xμ)2)

Facts:

Notation: XN(μ,σ)

Proof: comes from the case N(0,1) by using change of variables.

Fact: let XN(0,1), let μ and σ be fixed. Let Y=μ+σX. Then YN(μ,σ2).

Fact: let X and X be two independent r.v.’s respectively, N(μ,σ2), N(μ,σ2), then

Y=X+XN(μ+μ,σ2+σ2)

Moral of the story: the class of normal r.v.’s is stable by linear combination.

Example 1

let XN(1,1), YN(0,4), ZN(2,1). Assume they are independent. V=2X+3+Y4Z. Find E(V), Var(V).

A: By the above moral, V should be normal.

E(V)=13, Var(V)=24

Example 2

Let XN(μ,σ2), let

Z=xμσ

so E(Z)=0, Var(Z)=1. Because Z is a linear transformation of X.

Joint normal distribution (multivariate normal)

Definition: The vector (X1,X2,,Xn) is normal with mean μRn and covariance matrix C. (here μ={μi},i=1,2,,n, C=ciji,j=1n, and E(xi)=μi, cov(xi,xj)=cij and sometimes people use the letter Q or Σ instead of C), if its joint density is

f(x1,x2,...,xn)=1(2π)n/21detCexp(12(xμ)TC1(xμ))

(Notice: x and μ are column vectors)

The stuff in exp() is

(xμ)TC1(xμ)=j=1ni=1n(xiμi)Cij1(xjμj)

Sample case, n=1, C=(σ2), so det(C)=σ2 and 1det(C)=1σ2

So

(xμ)TC1(xμ)=(xμ)2σ2

This matches f(x).

when n=2 with independent x1 and x2.

det(C)=σ12σ22 f(x1,x2)=12πσ12σ22exp(j=12i=12(xiμi)Cij(xjμj))=12πσ12σ22exp(12i=j=12(xiμi)1σi2(xiμi))=12πσ12σ22exp(12((x1μ1)2σ12+(x2μ2)2σ22))

In general, if x1,x2,,xn are independent normals, N(μi,σi2),i=1,2,,n, then

f(x1,x2,...,xn)=1(2π)n/2i=1nσiexp(12||(xiμiσi)vector||eucli2)=1(2π)n/2i=1nσiexp(12i=1n(xiμiσi)2)

We recognize that if Cov(xi,xj)=0 and (xi,xj) are bivariate normal, then xi and xj are independent.

Normally, Cov(xi,xj) (can not infer) xi and xj are independent. But it does when (xi,xj) is bivariate normal.

Question: let’s create a multivariate normal vector Y with covariance matrix C to create a vector X=(x1,x2,,xn), where all xi’s are i.i.d N(0,1) (standard normals’’).

(It’s easy to use Box-Mueller transformation to do so).

How to create a Y from this X?

Answer: Recall the notion of square-root of a matrix (C is positive definite).

(Linear algebra: MTM=MMT=C)

There exists a matrix M which is C’’.

Consider

Z=MX

We know (by stability by linear combinations) that Z is multivariate normal.

Note

E(Z)=ME(X)=0

What about Cov(Z)?

Qij=Cov(Zi,Zj)=E[ZiZj]=E((k=inMikxk)(l=1nMjlxl))=k=1nl=1nMikMjlE(xkxl)=k=1nMikMjl(E(xkxl)=1 where i=j)=(MMT)ij

So we see that Y=Z=X.

Exercise: For n=2, find a square-root for a 2×2 matrix. There should be a place in the book where the author does that in hiding for the purpose of creating a bivariate normal.

Gamma distribution - Chapter 4.3

Definition: A r.v. X is Gamma with shape parameter α and scale parameter θ if it has this density

f(x)=c(xθ)α1θ1exp(xθ)

where the constant c=Γ(α), if α=n is an integer, Γ=(n1)!

Notation: XΓ(α,θ).

Fact:

XΓ(α1+α2++αn,θ)

Consequently, apply the above with α1=α2==αn=1, the sum of n i.i.d r.v.’s exp(λ=1θ) is a r.v. Γ(n,θ).

Specifically, this proves exactly that the nth arrival’’ time of a Pos(λ) process is a Γ(n,1λ) r.v.



Disqus Comment 0