5 Parameter Functions

Asymptotic Results

Obtaining a point estimate of ϕ is straightforward due to the invariance property of the likelihood function: if θ^ is the maximum likelihood estimator of θ, then ϕ^=g(θ^) is the maximum likelihood estimator of ϕ.

For finding confidence intervals, as in previous chapters, there are two basic techniques, each based on asymptotic approximation: the first uses the asymptotic normality of the maximum likelihood estimator; the second is based on the asymptotic χ2 distribution of the deviance function.

Theorem 11: Asymptotic distribution of a function of the MLE, multiparameter case.
For a regular estimation problem, if ϕ0=g(θ0) is the true parameter value, then as n:

ϕ^N(ϕ0,g(θ0)TIE(θ0)-1g(θ0)),

where

g(θ0)={θ1g(θ0),,θdg(θ0)}T.

As a consequence we have

Var(ϕ^)=Var(g(θ^))g(θ0)TIE(θ0)-1g(θ0).

For practical problems we use alternatives to IE(θ0) and g(θ0) which are asymptotically equivalent:

ϕ^ N(ϕ0,g(θ^)TIO(θ^)-1g(θ^))
ϕ^ N(ϕ0,g(θ^)TIE(θ^)-1g(θ^)).

When d=1, g(θ0)=g(θ0) and IE(θ0)-1 is a scalar (1×1 matrix), so that

Var(ϕ^)=g(θ0)IE(θ0)-1g(θ0)=[g(θ0)]2IE(θ0)-1

Thus, Theorem 11 is a multi-parameter generalisation of that in Theorem 3.

We omit the proof (which follows similar lines to the proof of Theorem 6).

Theorem 11 leads to a very simple method of constructing confidence intervals: an approximate (1-α)×100% confidence interval for ϕ is

(ϕ^-zα/2se(ϕ^),ϕ^+zα/2se(ϕ^)),

where se(ϕ^)=[Var(ϕ^)]1/2.

This method of finding confidence intervals is relatively easy, but suffers the same drawback of the corresponding method in previous chapters: the interval will be symmetric about ϕ^, and this may lead to some values included in the interval having lower likelihood than some of the excluded values.

An alternative approach is based on the deviance function. In this case, however, the treatment of the nuisance parameters is rather different.

Theorem 12: For a regular estimation problem, if ϕ0=g(θ0) is the true parameter value, then as n:

D*(ϕ0)=2[(θ^)-P(ϕ0)]χ12,

where D* is the profile deviance function.

P(ϕ), termed the profile log-likelihood, is defined by

P(ϕ)=maxλ(ϕ,λ).

If ϕϕ0 then D*(ϕ) as n.

Thus, to evaluate the profile (log)likelihood, first the value of ϕ is fixed, and then the likelihood is maximised over all possible values of λ. Thus, P(ϕ) is the skyline of the log-likelihood surface viewed looked from the ϕ axis.

Theorem 12 then tells us that the corresponding profile deviance function has an approximate χ12 distribution, and this, in turn, can form the basis of confidence interval construction. To evaluate the profile deviance two maximisations of the likelihood are required, one over all of θ and the other over λ holding ϕ fixed.