Poisson
The last famous discrete distribution that we'll discuss in this unit is the Poisson, which is an extremely popular distribution for modeling discrete data. We'll introduce its PMF, mean, and variance, and then discuss its story in more detail.
Definition: Poisson Distribution
An r.v. has the Poisson distribution with parameter , where , if the PMF of is
We write this as .
This is a valid PMF because of the Taylor series .
Example Poisson Expectation and Variance
Let . Then the mean and variance are both equal to . For the* mean*, we have
To get the variance, we first find . By LOTUS,
Using the same method we used to get the variance of a Geometric r.v., we can obtain
so
The figure below shows the PMF and CDF of the and distributions from to . It appears that the mean of the is around 2 and the mean of the is around 5, consistent with our findings above. The PMF of the is highly skewed, but as grows larger, the skewness is reduced and the PMF becomes more bell-shaped.
The Poisson distribution is often used in situations where we are counting the number of successes in a particular region or interval of time, and there are a large number of trials, each with a small probability of success. For example, the following random variables could follow a distribution that is approximately Poisson.
- The number of emails you receive in an hour. There are a lot of people who could potentially email you in that hour, but it is unlikely that any specific person will actually email you in that hour. Alternatively, imagine subdividing the hour into milliseconds. There are seconds in an hour, but in any specific millisecond it is unlikely that you will get an email.
- The number of chips in a chocolate chip cookie. Imagine subdividing the cookie into small cubes; the probability of getting a chocolate chip in a single cube is small, but the number of cubes is large.
- The number of earthquakes in a year in some region of the world. At any given time and location, the probability of an earthquake is small, but there are a large number of possible times and locations for earthquakes to occur over the course of the year.
The parameter is interpreted as the rate of occurrence of these rare events; in the examples above, could be 20 (emails per hour), 10 (chips per cookie), and 2 (earthquakes per year). In applications similar to the ones above, we can approximate the distribution of the number of events that occur by a Poisson distribution.
Poisson Approximation
Let be events with , where is large, the are small, and the are independent or weakly dependent. Let count how many of the occur. Then is approximately , with .
The Poisson paradigm is also called the law of rare events. The interpretation of ''rare" is that the are small, not that is small. For example, in the email example, the low probability of getting an email from a specific person in a particular hour is offset by the large number of people who could send you an email in that hour.
In the examples we gave above, the number of events that occur isn't exactly Poisson because a Poisson random variable has no upper bound, whereas how many of occur is at most , and there is a limit to how many chocolate chips can be crammed into a cookie. But the Poisson distribution often gives good approximations. Note that the conditions for the Poisson paradigm to hold are fairly flexible: the trials can have different success probabilities, and the trials don't have to be independent, though they should not be very dependent. So there are a wide variety of situations that can be cast in terms of the Poisson paradigm. This makes the Poisson a popular model, or at least a starting point, for data whose values are nonnegative integers (called count data in statistics).
The Poisson approximation greatly simplifies obtaining a good approximate solution to the birthday problem, and makes it possible to obtain good approximations to various variations which would be very difficult to solve exactly.
Example Birthday Problem Continued
If we have people and make the usual assumptions about birthdays, then each pair of people has probability of having the same birthday, and there are pairs. By the Poisson paradigm the distribution of the number of birthday matches is approximately , where . Then the probability of at least one match is
For , and , which agrees with the earlier result that we need 23 people to have a 50-50 chance of a matching birthday. Note that even though is fairly small, the relevant quantity in this problem is actually , which is the total number of ''trials'' for a successful birthday match, so the Poisson approximation still performs well.
Example Near-birthday Problem
What if we want to find the number of people required in order to have a 50-50 chance that two people would have birthdays within one day of each other (i.e., on the same day or one day apart)? Unlike the original birthday problem, this is difficult to obtain an exact answer for, but the Poisson paradigm still applies.
The probability that any two people have birthdays within one day of each other is (choose a birthday for the first person, and then the second person needs to be born on that day, the day before, or the day after). Again there are possible pairs, so the number of within-one-day matches is approximately where . Then a calculation similar to the one above tells us that we need or more. This was a quick approximation, but it turns out that is the exact answer!
Another useful property of the Poisson distribution is that the sum of independent Poissons is Poisson.
Theorem: Sum of Independent Poissons
If , , and is independent of , then .
Proof:
The story of the Poisson distribution provides intuition for this result. If there are two different types of events occurring at rates and , independently, then the overall event rate is .