Given the joint distribution of we may want to find the (marginal) distribution of or alone. The marginal distribution tells us about the behaviour of one random variable alone, i.e. irrespective of the other. We have been studying such distributions in the earlier chapters on univariate variables.
If we have the cdf the marginal cdfs are obtained as follows:
,
.
because the marginal event is the same as the joint event and the event is the same as the event , as illustrated on Figure Figure 5.2 (First Link, Second Link).
Just as, for a discrete RV, the the marginal pmfs are obtained by summing over the other variable, so, for a continuous RV, the marginal pdfs are obtained by integrating over the other variable.
If and are continuous random variables their marginal pdfs are
: For continuous random variables and we have
and by differentiating both sides wrt. we get
Similarly for .
Find the marginal pdf of for the joint distribution given in Example 5.5.2.
Solution. pdf: The joint pdf is
Hence
for .
cdf: for and , so
Differentiating gives
The random variables have joint distribution function
for , , for . Find the marginal distributions of and and identify their forms.
Solution. Since ,
for By symmetry for . Both marginal distributions are therefore
The random variables have joint pdf
for and . Find the marginal distributions of and and identify their forms.
Solution.
for (since is a density on ), giving By symmetry
Recall that, formally, we say that two random variables and are independent if the events and are independent for all sets and , i.e.
for all sets and .
We have seen that when and are both discrete, they are independent if and only if their joint pmf can be factorised as a product of the marginal pmfs.
Similarly, when and are both continuous they are independent if and only if their joint pdf can be factorised as a product of the marginal pdfs.
Two continuous random variables and are independent if and only if
If and are independent then whatever the values of and , take and . Then
This is true for all , and so we may differentiate both sides wrt. and to obtain
If the joint pdf factorises we get for arbitrary sets and ,
∎
Factorisation To check for independence: if we have the joint pdf (or pmf) it is enough to check that it can be factorised as a function of times a function of :
and that the range of does not depend on (see CW question). We do not have to show that the functions and are themselves densities. Also if the range of does not depend on then the range of does not depend on , so we only need to check one of the two possibilities.
If the range of does not depend on (and vice versa) we say that and are variationally independent.
The figure below illustrates the joint density
where the function is one when and zero otherwise, for four different regions . In which cases are and independent?
Unnumbered Figure: First link, Second Link
Unnumbered Figure: First link, Second Link
Solution. TL: ind, TR: not ind, BL: not ind., BR: ind.
Given a joint pdf, a standard way to prove independence is to show factorisation and variational independence. To disprove independence, a counterexample to either suffices. This is straightforward for variational independence, but disproving factorisation is less obvious. The following method is recommended. An alternative is to show that a conditional distribution is not the same as a marginal distribution, but that usually involves more work.
Two point method: Note that can be factorised as a function of times a function of if and only if for all , , , :
Since, in the case of independence, both sides equal .
This is particularly useful for proving that a given joint pdf does not factorise as above. Simply find , and , such that the two sides above are different.
Are the following pairs of random variables independent?
for ,
for ,
for
Solution.
Independent: variationally independent and , so the joint density factorises.
Not independent: factorises BUT the range of depends on .
Not independent: variationally independent BUT with and we have
Note: given variational independence, we first try to factorise ; when we cannot, we look for a counter-example.
Fewer sets and : as in the proof of Theorem 5.6.1, setting and shows that if and are independent then for all , . It turns out (we will not prove this) that for any pair of random variables, whether discrete, continuous or more complicated, ‘ for all ’ is equivalent to and being independent (i.e. one need only consider a subset of the possible sets and ).
Setting and shows that the independence of and implies for all ; again, it can be shown that independence is equivalent to the factorisation of the survivor functions.
Let and be independent exponential random variables with parameters and respectively. Find .
Solution. By independence, for , . So
Suppose we know the joint distribution of but then we find out the value of one of the random variables. What can we say about the other random variable?
We consider the conditional distributions , i.e. the distribution of given that , and , i.e. the distribution of given that .
Recall that when and were discrete random variables the conditional pmfs were:
Similarly when and are continuous random variables the conditional pdfs are
Note that since we can only condition on possible values, we don’t have to worry about zero’s in the denominators: the marginal pmf/pdf has to be positive for the value to occur.
Also note that the conditional pdfs are themselves valid pdfs: they are non-negative and they integrate to . For instance,
Similarly, conditional pmfs sum to .
When the variables are independent discrete RVs then for all , , recall that
,
.
Similarly if are independent continuous RVs then for all , ,
,
.
These results conform with intuition as when and are independent knowing the value of should tell us nothing about and vice versa.
The converse is also true: If the conditional distribution of given is independent of or, equivalently, the conditional distribution of given is independent of , then and are independent.
A piece of string of unit length is tied at one end to a hook. The string is cut at a (uniform) random distance from the hook. The piece remaining is then cut again at a (uniform) random distance from the hook. Given that the remaining length tied to the hook has length , find the pdf of the position of the first cut.
Unnumbered Figure: Link
Solution. Model with and We know for , and for .
Now .
We know and otherwise.
So we need .
for Hence
Continuous random variables and have joint pdf
Find
the conditional pdf of given ,
.
Solution.
Since we need the marginal pdf .
for Hence for
When we have so
Let be a bivariate rv.
The joint cdf is . .
For a discrete rv, the joint pmf is .
For a continuous rv, the joint pdf is .
For discrete rvs and , the marginal pmf of , is , and the conditional pmf of given is .
For continuous rvs and , the marginal pdf of is , and the conditional pdf of given is .
and are independent if and only if the events and are independent for all sets and : for all , .
An equivalent, but easier to check, condition for independence (of discrete or continuous rvs) is: . For discrete rvs, independence is also equivalent to , whereas for continuous rvs it is equivalent to . When just checking factorisation within the range where the rvs are non-zero, variational independence must also be verified.
Lack of independence can be shown using the two-point method; showing that for some . Alternatively, show that for some .
We have already encountered the expectation and variance of a univariate random variable, . In this chapter we examine the corresponding measures for multivariate random variables. We also investigate a particularly useful expectation: the moment generating function.
We know how to obtain expectations for univariate random variables. The definition extends easily to bivariate random variables. The expectation of any function is given by:
In the rest of this section results are given for the continuous random variable case only, however these extend immediately to discrete random variables.
Moments of either variable alone can be obtained from the joint distribution or from the relevant marginal.
and, more generally, for a function ,
Similarly for and any function (including ),
Using linearity of integrals we also have for any functions and
In particular
regardless of the joint distribution of .
If and are independent we also have for any functions and
In particular, if and are independent, then
Firstly we note that for dependent random variables , in general. For example, setting gives
the difference between the two being
More subtly, even when , and need not be independent.
Let and . Find and .
Solution. , so . Also
since for an odd integer. So .
The joint distribution of is illustrated on Figure 6.1. Clearly the variables and are strongly related, as given we know exactly.
Find the expected value of if . Does this result depend on other features of the joint distribution of ?
Solution. . No other assumptions are needed.
The random variables have joint pdf
Find , and . Does ?
Solution.
Unnumbered Figure: Link
Expectations for conditional random variables are defined in the obvious way. Conditional expectations are given by
,
.
is a function , say, of (a real number). If we have not yet seen then this becomes a function of the random variable . i.e. is a random variable because it is a function of the random variable .
Sometimes conditioning provides an easy way to obtain the expectations of the marginal variables. Consider the random variable , which is a function of . Just as , so the expectation of is
Now consider , which is a random variable, since it is a function of the random variable .
Intuitively, by conditioning on the unknown it becomes an unknown constant as far as the expectation is concerned and so it can be taken outside the expectation.
The rvs and follow a distribution specified by and .
Write down and .
Find and .
Find .
Solution.
and .
and
Note that .
The conditional variances are given by
If and are independent the conditional distributions are the same as the marginal distributions ( and ), so that in particular
,
,
,
.
We have seen that the marginal expectations can be obtained from the conditional expectations. We can also obtain the marginal variances from the conditional expectations and variances by the following formula:
These formulae are particularly useful when a random variable is given as a mixture of distributions. This is most easily illustrated by an example.
Let be a Poisson random variable, and given takes the value let be Binomial-distributed, i.e. Binomial. Find the expectation and variance of .
Solution. From properties of the Binomial distribution we have
,
.
Hence, using properties of the Poisson distribution we obtain
In fact, it can be shown that .
The moment generating function or mgf of a random variable is defined through
for all real values of for which the expectation exists.
Moment generating functions can be manipulated in many ways to reveal properties of the underlying probability distributions. They often help in mathematical proofs of probability theorems, and will be used for this purpose in Chapter 9.
Find the mgf of the random variable following the exponential distribution with parameter ; sketch the mgf when .
Solution. , for . Hence,
for . Note that is only defined for , since only in that case does the integral exist. Hence, for the mgf looks like:
Unnumbered Figure: Link
Quiz: Now consider a general rv: can the mgf be negative? No; it is the expectation of a non-negative quantity.
If mgf is defined in some neighbourhood of the origin, , the following properties are satisfied:
The mgf determines uniquely the distribution of the rv . That is, if two rvs have the same mgf then they have the same cdf.
If , for real and non-zero real number, .
Moments about the origin can be obtained by differentiating the mgf with respect to and then evaluating the mgf at zero, i.e.
Hence the name!
Let be independent rvs with mgf respectively. Then,
Proof uses ideas from complex analysis (see Math215).
If , then
Since then , etc.; but so
,
,
,
and so on.
by independence, so . ∎
From Part 4, by induction, if are independent random variables:
Using its mgf, find the expectation and the variance of the random variable following the exponential distribution with parameter .
Solution. Consider the first two derivatives of the mgf:
, .
Hence, , and
The mgf of a Normal random variable We first consider . Then
by completing the squares. Hence by unit integrability of the N(t,1) density.
So if then by Property 2,
For instance, if then
,
,
,
,
.
In particular so and as mentioned in Chapter 3.
Unfortunately the mgf is not defined for some rvs.
Let , then
which is not defined as, if the integrand as and if the integrand as .
The sum of two independent Normal random variables is also Normal. Let and be two independent random variables, then
This is called the convolution property of the Normal distribution. There are several proofs of it; the above is the simplest and starts to show the power of mgfs.
(Exam2016) For some , let with probability , and be independent of each other. Let , and . You may take as given that the moment generating function (mgf) of is .
Find the mgf of , .
Solution. Since ,
Find the mgf of , . Be sure to specify the range of and make clear why this range applies.
Solution. This is (provided ); see Example 6.4.1 for detail and reason.
Show that, subject to the same range condition on , .
Solution.
which gives the required result.
Find the mgf of ; what (if any) condition on the range of applies?
Find and interpret the result heuristically with reference to your answer to an earlier part of this question. (Hint: recall that .)
Solution.
Need
For any , for large enough , . So,
This is the mgf of , so as (in some sense).
Find the mgf of and interpret the result with reference to your answer to an earlier part of this question. (Hint: use the tower property of expectations: .)
Solution.
This is like but with replaced by . So has the same distribution as the difference between two random variables. Or, equivalently, the difference between two random variables has the same distribution as the product of a and the square-root of an