Discrete Uniform
A very simple story, closely connected to the naive definition of probability, describes picking a random number from some finite set of possibilities.
Story Discrete Uniform Distribution
Let be a finite, nonempty set of numbers. Choose one of these numbers uniformly at random (i.e., all values in are equally likely). Call the chosen number . Then is said to have the Discrete Uniform distribution with parameter ; we denote this by . The PMF of is
for (and otherwise), since a PMF must sum to . As with questions based on the naive definition of probability, questions based on a Discrete Uniform distribution reduce to counting problems. Specifically, for and any , we have
Example Random Slips of Paper
There are 100 slips of paper in a hat, each of which has one of the numbers written on it, with no number appearing more than once. Five of the slips are drawn, one at a time. First consider random sampling with replacement (with equal probabilities).
(a) What is the distribution of how many of the drawn slips have a value of at least written on them?
(b) What is the distribution of the value of the j-th draw (for )?
(c) What is the probability that the number 100 is drawn at least once?
Now consider random sampling without replacement (with all sets of five slips equally likely to be chosen).
(d) What is the distribution of how many of the drawn slips have a value of at least 80 written on them?
(e) What is the distribution of the value of the j-th draw (for )?
(f) What is the probability that the number 100 is drawn at least once?
Solution:
(a) By the story of the Binomial, the distribution is .
(b) Let be the value of the jth draw. By symmetry, .
(c) Taking complements,
By the naive definition of probability, this is
(d) By the story of the Hypergeometric, the distribution is .
(e) Let be the value of the jth draw. By symmetry, . Here learning any gives information about the other values (so are not independent), but symmetry still holds since, unconditionally, the jth slip drawn is equally likely to be any of the slips.
(f) The events are disjoint since we are now sampling without replacement, so
Cumulative Distribution Functions
Another function that describes the distribution of an r.v. is the cumulative distribution function (CDF). Unlike the PMF, which only discrete r.v.s possess, the CDF is defined for all r.v.s.
Definition
The cumulative distribution function (CDF) of an r.v. is the function given by . When there is no risk of ambiguity, we sometimes drop the subscript and just write (or some other letter) for a CDF.
The next example demonstrates that for discrete r.v.s, we can freely convert between CDF and PMF.
Example
Let . The following figure shows the PMF and CDF of .
-
From PMF to CDF: To find , which is the CDF evaluated at 1.5, we sum the PMF over all values of the support that are less than or equal to 1.5:
Similarly, the value of the CDF at an arbitrary point is the sum of the heights of the vertical bars of the PMF at values less than or equal to .
-
From CDF to PMF: The CDF of a discrete r.v. consists of jumps and flat regions. The height of a jump in the CDF at is equal to the value of the PMF at . For example, in the above figure, the height of the jump in the CDF at 2 is the same as the height of the corresponding vertical bar in the PMF; this is indicated in the figure with curly braces. The flat regions of the CDF correspond to values outside the support of , so the PMF is equal to 0 in those regions.
Valid CDFs satisfy the following criteria.
Theorem: Valid CDFs
Any CDF has the following properties.
- Increasing: If , then .
- Right-continuous: As in the above figure, the CDF is continuous except possibly for having some jumps. Wherever there is a jump, the CDF is continuous from the right. That is, for any , we have .
- Convergence to and in the limits:
The converse is true too: it turns out that given any function meeting these criteria, we can construct a random variable whose CDF is . To recap, we have now seen three equivalent ways of expressing the distribution of a random variable. Two of these are the PMF and the CDF: we know these two functions contain the same information, since we can always figure out the CDF from the PMF and vice versa. Generally the PMF is easier to work with for discrete r.v.s, since evaluating the CDF requires a summation. A third way to describe a distribution is with a story that explains (in a precise way) how the distribution can arise. We used the stories of the Binomial and Hypergeometric distributions to derive the corresponding PMFs. Thus the story and the PMF also contain the same information, though we can often achieve more intuitive proofs with the story than with PMF calculations.