Home page for accesible maths 3.2 One sample tests

Style control - access keys in brackets

Font (2 3) - + Letter spacing (4 5) - + Word spacing (6 7) - + Line spacing (8 9) - +

3.2.1 Justification of the one-sample t-test

The general concept of conducting an hypothesis test is to

  1. 1.

    Calculate a test statistic;

  2. 2.

    Compare the test statistic to a sampling distribution, either using a critical value or a p-value.

This is a valid method for testing all sorts of parameters from different statistical models. The test statistic and sampling distribution change according to the parameter being tested. You will see further examples throughout the course.

We give a justification for the use of both the t statistic and Students-t distribution. First recall the null hypothesis,

H0:μ=μ0.

Since an unbiased estimator of μ is the sample mean X¯, it seems sensible to compare the observed sample mean x¯ to the hypothesised population mean μ0. The question is then:

When μ=μ0, how far from μ0 can we reasonably expect x¯ to be?

To answer this we need to look at the sampling distribution of the estimator X¯. This will show us, for a given sample size n, how much sampling variability we can expect in the sample mean. We start with the case where the population variance σ2 is known, since in this case we can use the Central Limit Theorem to help prove the sampling distribution of the test statistic.

Theorem 1.

If X1,,Xn are IID random variables, with mean μ and variance σ2< then as n,

X¯Normal(μ,σ2n).

This is an exact result if X1,,Xn are Normal random variables, and so holds for all n. Otherwise the limiting result follows by application of the Central Limit Theorem.

If H0 is true, we can adjust this to,

X¯Normal(μ0,σ2n).

By the usual standardisation technique,

Z=X¯-μ0σ/nNormal(0,1).

This then gives us both a test statistic Z and its sampling distribution but only if we know the population variance σ2.

When the population variance is unknown, it is replaced by the sample variance s2, but this is itself a realisation of a random variable S2 (different samples of size n will each have a different sample variance). Deriving the sampling distribution of

T=X¯-μS/n

is therefore considerably harder, as it is the ratio of two related random variables X¯ and S2. In fact, the t-distribution was first tabulated as the sampling distribution for this test statistic, by William S. Gossett (work published in 1908) as part of his work in quality control for the Guinness brewery in Dublin. Gossett found that the Normal distribution was no longer an appropriate sampling distribution for T, due to the extra uncertainty introduced when estimating S (especially for small samples). His derivation of the t-distribution had three steps:

  1. 1.

    Showing that the sampling distribution of S is χn2;

  2. 2.

    Showing that S is independent of X¯;

  3. 3.

    Showing that the ratio of a Normal random variable to the square root of a χn2 random variable follows a tn-1-distribution.

The entire proof takes up a page of A4 paper. To cover this thoroughly is beyond the scope of this course.