1 Modelling and Statistical Inference

Maximum likelihood estimation and relative likelihood

By definition, L(θ) or (θ) give measures of how likely θ is to be the true parameter value for the population, since (up to proportionality) L(θ) is the probability of obtaining the observed data if θ had been the true parameter.

Consequently, values of θ with large (log)likelihood are more likely to be correct than values with low (log)likelihood.

This leads to the principle of maximum likelihood estimation, whereby we estimate θ by the maximum likelihood estimator (MLE), defined to be any value θ^ of θ that maximises L(θ) (or (θ)).

Example 1.2:  IID Poisson data.
Find θ^ for independent data x=x1,,xn from a Poisson(θ) distribution.

For θΘ with Θ=(0,), f(x|θ)=e-θθxx!  x=0,1,2,. Then

L(θ)=i=1ne-θθxixi!i=1ne-θθxi.
(θ) = log{i=1ne-θθxi}(+const)
(θ) = -nθ+nx¯log(θ)
(θ) = -n+nx¯θ
(θ^) = -n+nx¯θ^=0
θ^ = x¯.

We should also check that the second derivative is negative. Differentiating is a common way to find maxima (for continuous parameter space) but not the only way. It won’t always work and we may have to resort to ‘brute force’ approaches.

We define the first derivative of the log-likelihood as the score function:

S(θ)=(θ)=θlogL(θ)

so the MLE θ^ is the solution of the score equation S(θ)=0.

The likelihood function can also be used to interpret the plausibility of values of θ other than the MLE. Define the relative likelihood of θ as

RL(θ)=L(θ)L(θ^).

Example 1.3:  Coin tossing.
Suppose a coin is tossed n=10 times and r=6 heads are observed. Then if θ represents the probability of a head on a single coin toss, the number of heads is Binomial(10,θ), and the likelihood function is

L(θ) = θr(1-θ)n-r,
(θ) = rlogθ+(n-r)log(1-θ),
(θ) = rθ-n-r1-θ,

and this gives

θ^=rn=0.6.

Now we can consider the relative likelihood of θ=0.5 (i.e. the coin is fair) as

RL(0.5)=L(θ=0.5)L(θ=0.6)=0.56×0.540.66×0.44=9.77×10-411.94×10-4=0.82.

One could subjectively interpret the relative likelihood; however, we will develop theoretical tools to compare values of θ with different likelihoods later.

For this development, it is useful to introduce the deviance, which is a relative likelihood-like measure:

D(θ)=2{(θ^)-(θ)}.

Exercise: what is the relationship between the relative likelihood and the deviance?

Observed and expected information

There are two ways to measure the information about θ contained in the sample: the observed information IO(θ) and the expected information IE(θ). The observed information is based on the observed data

IO(θ)=-′′(θ;x)=-′′(θ)

and the expected information, sometimes called Fisher information, is averaged over all possible data sets (expectation over X)

IE(θ)=E{-2θ2(θ|X)}.

These quantities also represent the curvature of the likelihood function. Particularly of interest is the curvature at the MLE, IO(θ^). High information (large curvature) at the MLE corresponds to a tight peak, and indicates less uncertainty about θ0.

Example 1.4:  IID Poisson data, ctd.
Find the observed and expected information for independent data x1,,xn from a Poisson(θ) distribution. How do they change with n?

Recall (θ)=-nθ+nx¯log(θ). So

(θ) = -n+nx¯θ
′′(θ) = -nx¯θ2

Thus IO(θ)=nx¯/θ2.

Now as E(X¯)=E(i=1nXi/n)=i=1nθ0/n=θ0,

′′(θ0;X) = -nX¯θ02
IE(θ0) = nE(X¯)θ02=nθ0θ02=nθ0

Note that both IO(θ) and IE(θ) are multiples of n.

The pair (θ^,IO(θ^)) provide a useful two-dimensional summary of any (log) likelihood function. Indeed, consider approximating (θ) by a second-order Taylor expansion about θ^:

(θ) (θ^)+(θ-θ^)(θ^)+12(θ-θ^)2′′(θ^)
(θ) (θ^)+12(θ-θ^)2′′(θ^).

This can be written as

log(RL(θ))-12(θ-θ^)2IO(θ^),

so this is a quadratic approximation for the log relative likelihood depending only on our pair (θ^,IO(θ^)). If this approximation is accurate around the MLE, we call the likelihood function regular.

Confidence Regions

As frequentists, however, we must consider that the data, x, which we have used to construct our MLE θ^(x), is a realisation of the random variables X. Another draw from these random variables would result in different data, x, and hence a different MLE, θ^(x).

Therefore, θ^(X) is itself a random variable, and must have a distribution: we call this the sampling distribution of θ^.

The fact that the MLE has a distribution makes it seem somewhat foolish simply to quote the MLE as a point estimate of the true value of the parameter θ0. This motivates the idea of a confidence region for θ0, which is defined as any rule for constructing a region C with the following property:

P[C(X)θ0]=1-α,

where 1-α is referred to as the confidence level, and is specified in advance.

Where confidence regions are 1-dimensional and of the form C(x)=(θl(x),θu(x)) they are usually called confidence intervals.

This is a precise and important definition. Any specific confidence region C(x) we produce depends upon the observed data x. It either contains, or does not contain, the true parameter θ0.

We cannot make a probability statement about this specific confidence region in isolation because there is no repeated sampling. However, there is a collection of confidence regions that would be produced for all the different samples x we could draw from the random variables X.

The probability of these intervals covering the true value θ0 is 1-α, where we define probability as a limiting proportion over the samples.

It is rarely possible to construct confidence regions exactly; hence, it is usual to rely on asymptotic results (i.e. results with n) and then assume that these results hold approximately for our observed n (see e.g. Example 2.2).

The Cramér-Rao bound and MVUEs

Before framing our discussion in more formal asymptotic theory, we first we first recall a result which gives us a bound on how good an inference method can be.

Theorem (CRLB): Cramér-Rao bound for variance of estimators.
Let T(𝐗) be an estimator with expected value E[T(𝐗)]=g(θ), where g(θ) is a differentiable function. Then

Var[T(𝐗)][g(θ0)]2[IE(θ0)]-1.

In particular, if T(𝐗) is an unbiased estimator of θ, we have that g(θ)=θ and g(θ)=1. Thus,

Var[T(𝐗)][IE(θ0)]-1.

This states that an estimator has minimum variance if it achieves the bounds on the righthand side of the two expressions above.

In particular, if the estimator is unbiased, if its variance is the reciprocal of the Fisher information of a model, it is a minimum variance unbiased estimator (MVUE).

Example 1.5:  The sample mean estimator of the Poisson parameter.
Let x1,,xn be independent data from a Poisson(θ) distribution. Is X¯ a MVUE for the Poisson parameter θ?

  • Firstly, we note that E(X¯)=E(i=1nXi/n)=i=1nθ/n=θ, thus X¯ is unbiased as an estimator for θ.

  • The variance of X¯ is

    Var(X¯) = Var(i=1nXi/n)
    = i=1n1n2Var(θ)=nθ/n2=θ/n.
  • Lastly, recall from Example 1.4 that IE(θ)=nθ.

Since the variance of X¯ is equal to 1/IE(θ), according to the CRLB Theorem, it is a minimum variance estimator (because it achieves the Cramér-Rao lower bound).

From above, it is unbiased too, so is a MVUE.