Joint Probability Distributions
Joint Probability Distributions
The objective of statistics is to learn about population characteristics: this was first mentioned in the Introductory Section. An EXPERIMENT is then any process which generates data. It is easy to imagine circumstances where an experiment (say a few interviewers asking couples walking down the road) generates observations, for example the weekly income of husbands and the weekly income of wives in a particular population of husbands and wives. One possible use of such sample data is to investigate the relationship between the observed values of the two variables. In the Section Correlation and Regression we discussed how to use correlation and regression to summarise the extent and nature of any linear relationship between these observed values. It has to be said that the discussion of relationships between variables in this section does but scratch the surface of a very large topic. Subsequent courses in Econometrics take the analysis of relationships between two or more variables much further.
If these two pieces of information generated by the experiment are considered to be the values of two random variables defined on the SAMPLE SPACE of an experiment (see this [[ProbabilityIntro | Section]]), then the discussion of random variables and probability distributions needs to be extended to the multivariate case.
Let [math]X[/math] and [math]Y[/math] be the two random variables: for simplicity, they are considered to be discrete random variables. The outcome of the experiment is a pair of values [math]\left( x,y\right)[/math]. The probability of this outcome is a joint probability which can be denoted
[math]\Pr \left( X=x\cap Y=y\right) ,[/math]
emphasising the analogy with the probability of a joint event [math]\Pr \left( A\cap B\right)[/math], or, more usually, by
[math]\Pr \left( X=x,Y=y\right) .[/math]
The collection of these probabilities, for all possible combinations of [math]x[/math] and [math]y[/math], is the joint probability distribution of [math]X[/math] and [math]Y, [/math] denoted
[math]p\left( x,y\right) =\Pr \left( X=x,Y=y\right) .[/math]
The Axioms of Probability discussed in this Section carry over to imply
[math]0\leqslant p\left( x,y\right) \leqslant 1,[/math]
[math]\sum_{x}\sum_{y}p\left( x,y\right) =1,[/math]
where the sum is over all [math]\left( x,y\right)[/math] combinations.
Examples
Example 1
Let [math]H[/math] and [math]W[/math] be the random variables representing the population of weekly incomes of husbands and wives, respectively, in some country. There are only three possible weekly incomes, £0, £100 or £ 200. The joint probability distribution of [math]H[/math] and [math]W[/math] is represented as a table:
[math]i[/math] | ||||||||||||
Variable | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 |
Weight [math]=Y_i[/math] | 155 | 150 | 180 | 135 | 156 | 168 | 178 | 160 | 132 | 145 | 139 | 152 |
Height [math]=X_i[/math] | 70 | 63 | 72 | 60 | 66 | 70 | 74 | 65 | 62 | 67 | 65 | 68 |
[math]H[/math] | ||||
Probabilities | [math]0[/math] | [math]1[/math] | [math]2[/math] | |
Values of [math]W:[/math] | [math]0[/math] | [math]0.05[/math] | [math]0.15[/math] | [math]0.10[/math] |
[math]1[/math] | [math]0.10[/math] | [math]0.10[/math] | [math]0.30[/math] | |
[math]2[/math] | [math]0.05[/math] | [math]0.05[/math] | [math]0.10[/math] |
Then we can read off, for example, that
[math]\Pr \left( H=0,W=0\right) =0.05,[/math]
or that in this population, [math]5\%[/math] of husbands and wives have each a zero weekly income.