An important tool in the study of the relationships that exist between two jointly distributed random variables,
the last two equations hold, respectively, in the cases in which
Example 7A. Sampling from an urn of random composition. Let a random sample of size
since the conditional probability law of
The mean number of white balls in the sample drawn is then equal to
Now
Example 7B. The conditional mean of jointly normal random variables. Two random variables,
Consequently, the conditional mean of
in which we define the constants
Similarly,
From (7.7) it is seen that the conditional mean of a random variable
The conditional mean of one random variable, given another random variable, represents one possible answer to the problem of prediction . Suppose that a prospective father of height
for any function
From (7.7) it is seen that in the case in which the random variables are jointly normally distributed the problem of computing the conditional mean
It may happen that the joint probability law of the random variables
Solving for the values of
Therefore,
Comparing (7.7) and (7.13), one sees that the best linear predictor
We can readily compute the mean square error of prediction achieved with the use of the best linear predictor. We have
From (7.14) one obtains the important conclusion that the closer the correlation between two random variables is to 1, the smaller the mean square error of prediction involved in predicting the value of one of the random variables from the value of the other.
The Phenomenon of“Spurious” Correlation . Given three random variables
(or in some similar way) as functions of
Example 7C. Do storks bring babies? Let
then represent, respectively, the number of storks per woman and the number of babies born per woman in the area. If the correlation coefficient
Theoretical Exercises
In the following exercises let
7.1. The best linear predictor , denoted by
where
in which we define
7.2. The residual of
Show that
Next show that the variance of the predictor is given by
The positive quantity
is called the multiple correlation coefficient between
7.3. The partial correlation coefficient of
in which
7.4. (Continuation of example 7A). Show that
Exercises
7.1. Let
Answer
(i)
7.2. Find the conditional mean of
7.3. Let
Find the mean square error of prediction achieved by the use of (i) the best linear predictor, (ii) the best predictor.
7.4. Let
Answer
(i)