Home page for accesible maths 9.1 Preview

Style control - access keys in brackets

Font (2 3) - + Letter spacing (4 5) - + Word spacing (6 7) - + Line spacing (8 9) - +

Sequences of Normal random variables

Suppose X1,X2, are independent and identically distributed Normal random variables, XiN(μ,σ2). What can we say about X¯n?

Let us first recall the important ‘convolution’ property of the normal distribution, Theorem 6.4.2 (see also Chapter 8, Example 8.2.1). If Xi (i=1,,n) are independent and each has a Normal distribution then so does X1+X2.

By induction, therefore, Sn=i=1nXi also has a Normal distribution.

Thus X¯n=1ni=1nXi has a Normal distribution too.

This is the reason why the normal distribution plays such a central role in probability theory. For instance it is because of this property that it is the normal distribution that appears in the Central Limit Theorem (see later) and not anything else.

From the previous subsection we know the expectation and variance of X¯n, and hence:

X¯nN(μ,σ2/n). (9.1)
Example 9.1.1.

Find, in terms of the CDF of the standard Normal, Φ,

𝖯(|X¯n-μ|>0.01),

and hence find limn𝖯(|X¯n-μ|>0.01).

Solution. 

𝖯(|X¯n-μ|>0.01) =𝖯(X¯n-μ>0.01)+𝖯(X¯n-μ<-0.01)
=2𝖯(X¯n-μ<-0.01)

by symmetry – draw it. So

𝖯(|X¯n-μ|>0.01) =2Pr(X¯n-μσ/n<-0.01nσ)
=2Φ(-0.01nσ).

Hence

limn𝖯(|X¯n-μ|>0.01)=2limnΦ(-0.01nσ)=0.

Clearly the same argument holds for any ϵ>0, so for the mean of n IID normal random variables

𝖯(|X¯n-μ|>ϵ)0.

Crudely put, for large n, X¯nμ (probably). This type of convergence is called ‘convergence in probability’; i.e. X¯nμ in probability. You will cover this type of convergence (and other types) in the third year module on probability.

The Weak Law of Large Numbers (see next section) will show that this property often holds for averages of random variables even when they do not have a Normal distribution.

What about deviations from μ? From (9.1),

X¯n-μN(0,σ2/n),

and as n, the distribution of X¯n-μ tends to N(0,0), a point mass at 0, which we knew anyway. To obtain a useful distribution we must rescale, dividing X¯n-μ by its SD:

X¯n-μσ/nN(0,1),

or equivalently

n(X¯n-μ)σN(0,1),

and hence

𝖯(n(X¯n-μ)σz)=Φ(z).

If the discrepancy from μ is scaled by the standard deviation of Xi and stretched by n then it has a standard Normal distribution.

If the Xi have a distribution other than the Normal then the average of any n of them is, in general, not Normal, however the Central Limit Theorem will show that, if the same rescaling is applied then for large enough n the distribution (and hence the cdf) of n(X¯n-μ)/σ can be made as close as we like to a standard Normal: for any z,

𝖯(n(X¯n-μ)σz)Φ(z).

This type of convergence is called ‘convergence in distribution’; i.e. for any Z𝖭(0,1), n(X¯n-μ)/σZ in distribution. You will cover this type of convergence (and other types) in the third year module on probability.