4 Information and Asymptotics

Sketch Proofs of Results

We will now prove Theorem 6 and Theorem 7. It is important to follow the steps involved to help your understanding of likelihood concepts.

Lemma 8: Multivariate Central Limit Theorem.
Suppose that Y is a d-dimensional vector variable with mean vector μ and variance-covariance matrix Σ with finite values on the diagonal. If Y1,,Yn is an IID sequence of vector random variables having the same distribution as Y, and if

Sn=i=1nYi

(meaning a vector componentwise sum) then

n-1/2[Sn-E(Sn)]=n-1/2[Sn-nμ]MVNd(𝟎,Σ)

as

n.

Lemma 9: If YMVNd(𝟎,Σ) then YTΣ-1Yχd2.

Proof.

First recall that if RVs X1,,Xd are iid N(0,1), then X12++Xd2χd2 by definition.

Letting X=Σ-1/2Y, we have

  • E(X)=E(Σ-1/2Y)=𝟎

    (vector linear combination of zero mean r.v. – using result at start of chapter)

  • Var(X)=Var(Σ-1/2Y)=Σ-1/2Var(Y)(Σ-1/2)T=Σ-1/2Σ(Σ-1/2)T=Id

    (using variance result at start of chapter)

  • (vector) linear combination of normal r.v. is also normal

So XMVNd(0,I).

Hence YTΣ-1Y=XTX=X12++Xd2χd2. ∎

Lemma 10: Asymptotic distribution of the true score.
Under the regularity conditions,

  • E{U(θ0)}=𝟎

  • Var{U(θ0)}=E{IO(θ0)}=IE(θ0).

  • Asymptotically as n, U(θ0)N(0,IE(θ0)).

Example 4.1:  Normal Data, ctd.
XiN(μ,σ2)
, so θ=(μ,σ). In this case,

U(θ)={1σ2i=1n(xi-μ),-nσ+1σ3i=1n(xi-μ)2}T

and

IO(θ)=[nσ22σ3(xi-μ)2σ3(xi-μ)3σ4(xi-μ)2-nσ2]

so

IO(θ^)=[nσ^2002nσ^2] and IO(θ^)-1=[σ^2n00σ^22n].

Since

E{i=1n(Xi-μ)}=i=1n{E(Xi)-μ}=0

and

E{i=1n(Xi-μ)2}=i=1nE{(Xi-μ)2}=nσ2,

it follows that

IE(θ)=[nσ2002nσ2] and IE(θ)-1=[σ2n00σ22n].

Example 4.2:  Gamma Distribution, ctd.
XiGamma(α,β)
. In this case

U(θ)={nlogβ+i=1nlogxi-nγ(α),nαβ-i=1nxi}T,

where as before γ(α):=Γ(α)Γ(α) and

IO(θ)=IE(θ)=[nγ(α)-n/β-n/βnα/β2].

It follows that

IO(θ)-1=IE(θ)-1=Δ-1[nαβ2nβnβnγ(α)]

where the determinant Δ is

Δ=(nβ)2(αγ(α)-1).

Notice that in Example 4.4 IE(θ^)=IO(θ^), and in Example 4.4 IE=IO, but this is the exception rather than the rule.