Week: 1 Statistics 2

Publish Date: June 10, 2024

Maringal Probability Mass Function (PMF) of Two Discrete Random Variables

In probability theory and statistics, the marginal probability mass function (PMF) of two discrete random variables is a function that gives the probability of each possible value of one of the variables, without reference to the other variable. It essentially represents the distribution of each variable individually, ignoring the other variable.

The marginal PMF is obtained by summing (or integrating, in the case of continuous variables) the joint PMF over all possible values of the other variable.


Formula

For two discrete random variables and , the marginal PMF of is calculated as:

Where:

Similarly, the marginal PMF of is calculated as:


Example

In this example, we’ll consider two fair six-sided dice. Let’s denote the outcomes of the first die as and the outcomes of the second die as .

The joint PMF of and can be represented in a table:

To find the marginal PMF of , we sum the probabilities of all outcomes for each value of :

Similarly, for other values of , we would get , , , , as each die has an equal probability of landing on any number from 1 to 6.

Similarly, to find the marginal PMF of , we sum the probabilities of all outcomes for each value of . We would obtain the same result as for , i.e., .

Thus, both and have marginal PMFs that represent a uniform distribution, each with a probability of for each possible outcome.


FAQ

What is a marginal PMF?

A marginal PMF is a probability mass function that gives the probabilities of individual values of a single random variable, ignoring the values of other variables.

How is a marginal PMF computed from a joint PMF?

To compute a marginal PMF from a joint PMF, you sum (or integrate, in the case of continuous variables) the probabilities over all possible values of the other variables, leaving only the variable of interest.

Why are marginal PMFs useful?

Marginal PMFs are useful because they allow us to analyze the behavior of individual random variables without considering the complexities of their joint distributions with other variables.