Skip to content

08. Joint Distribution

\gdef \N{\mathbb{N}} \gdef \Z{\mathbb{Z}} \gdef \Q{\mathbb{Q}} \gdef \R{\mathbb{R}} \gdef \C{\mathbb{C}} \gdef \setcomp#1{\overline{#1}} \gdef \sseq{\subseteq} \gdef \pset#1{\mathcal{P}(#1)} \gdef \covariant{\operatorname{Cov}} \gdef \of{\circ} \gdef \p{^{\prime}} \gdef \pp{^{\prime\prime}} \gdef \ppp{^{\prime\prime\prime}} \gdef \pn#1{^{\prime\times{#1}}}

Definition

Given 2 Random Variables XX and YY, their joint distribution is the probability function PX,Y(k,l)P_{X,Y}(k,l) for all possible values k,lk,l of XX and YY respectively.

Example:
Balls numbered 1,2,31,2,3 are put in a jar. Two are selected without replacement. Let X=X= number on the first selected ball, Y=Y= number on the second selected ball.
The joint dist. function can be written as a table:

X\Y 1 2 3
1 0 16\frac16 16\frac16
2 16\frac16 0 16\frac16
3 16\frac16 16\frac16 0

The upper right square represents PX,Y(1,3)P_{X,Y}(1,3), and by the table, we can see that it is equal to 131st ball122nd ball=16\underbrace{\frac13}_{\text{1st ball}}\cdot\underbrace{\frac12}_{\text{2nd ball}}=\frac16
The sum of all probabilities in a joint distribution function must be 1.
In case that was too easy to understand, here’s the same statement in MathSpeak:

\(klPX,Y(k,l) =1\sum_k\sum_lP_{X,Y}(k,l)\ =1\)
The expected value of XYX\cdot Y is equal to the sum of a…
Ok I tried, but I can’t write this in English, here’s the algorithm to do it:

Python
1
2
3
4
5
6
7
def expected_value():
    expected=0
    for x_value in possible_values_of_X: # For every row in table
        for y_value in possible_values_of_Y: # For every cell in row
            expected+=x_value*y_value*P(X=x_value,Y=y_value)
            # Where P(X=x_value,Y=y_value) is the value in the cell
    return expected

And here’s the MathSpeak version:
\(E[YX]=klklPX,Y(k,l)E[YX]=\sum_k\sum_lkl\cdot P_{X,Y}(k,l)\)

Given the joint distribution function, we can find Px(k)P_x(k) and PY(l)P_Y(l) for any kk or ll.
In the above example, PX(2)=PX,Y(2,1)+PX,Y(2,2)+PX,Y(2,3)=16+0+16=13P_X(2)=P_{X,Y}(2,1)+P_{X,Y}(2,2)+P_{X,Y}(2,3)=\frac16+0+\frac16=\frac13

Marginal Probabilities

In general:
PX(k)=P_X(k)= the sum of entries in row kk in the table.
PY(l)=P_Y(l)= the sum of entries in col ll in the table.

These probabilities are called marginal probabilities and are typically written in the margin of the joint dist. table.

The sum of marginal probabilities for any random variable is always 1.

The above table, with added marginal probabilities:

X\Y 1 2 3 PXP_X
1 0 16\frac16 16\frac16 13\frac13
2 16\frac16 0 16\frac16 13\frac13
3 16\frac16 16\frac16 0 13\frac13
PYP_Y 13\frac13 13\frac13 13\frac13

X,Y are independent if and only if their joint distribution table forms a multiplication table, where each cell is equal to the product of the relevant marginal probabilities.
The above table, therefore, is not independent.

If a 0 appears anywhere inside a joint dist. table, then X,YX,Y must be dependent, because marginal probabilities are never zero.