Conditional Probability and Bayes' Rule 2

145 阅读2分钟

Conditional Probabilities are Probabilities

When we condition on an event EE, we update our beliefs to be consistent with this knowledge, effectively putting ourselves in a universe where we know that EE occurred. Within our new universe, however, the laws of probability operate just as before. Conditional probability satisfies all the properties of probability! Therefore, any of the results we have derived about probability are still valid if we replace all unconditional probabilities with probabilities conditional on EE.

For example, here are conditional forms of Bayes' rule and the law of total probability. These are obtained by taking the ordinary forms of Bayes' rule and LOTP and adding EE to the right of the vertical bar everywhere.

Theorem: Bayes' Rule with Extra Conditioning

Provided that P(AE)>0P(A \cap E) > 0 and P(BE)>0P(B \cap E) > 0, we have

P(AB,E)=P(BA,E)P(AE)P(BE).P(A|B,E) = \frac{P(B|A,E) P(A|E)}{P(B|E)}.

Theorem: LOTP with Extra Conditioning

Let A1,,AnA_1,\dots,A_n be a partition of SS. Provided that P(AiE)>0P(A_i \cap E) > 0 for all ii, we have

P(BE)=i=1nP(BAi,E)P(AiE).P(B|E) = \sum_{i=1}^n P(B|A_i, E) P(A_i|E).

Independence of Events

We have now seen several examples where conditioning on one event changes our beliefs about the probability of another event. The situation where events provide no information about each other is called independence.

Definition: Independence of Two Events

Events AA and BB are independent if

P(AB)=P(A)P(B).P(A \cap B) = P(A) P(B).

If P(A)>0P(A) > 0 and P(B)>0P(B) > 0, then this is equivalent to

P(AB)=P(A),P(A|B) = P(A),

and also equivalent to P(BA)=P(B).P(B|A) = P(B).

In words, two events are independent if we can obtain the probability of their intersection by multiplying their individual probabilities. Alternatively, AA and BB are independent if learning that BB occurred gives us no information that would change our probabilities for AA occurring (and vice versa).

Note that independence is a symmetric relation: if AA is independent of BB, then BB is independent of AA.

Independence is completely different from disjointness. If AA and BB are disjoint, then P(AB)=0P(A \cap B) = 0, so disjoint events can be independent only if P(A)=0P(A) = 0 or P(B)=0P(B) = 0. Knowing that AA occurs tells us that BB definitely did not occur, so AA clearly conveys information about BB.

We also often need to talk about independence of three or more events.

Definition: Independence of Three Events

Events AA, BB, and CC are said to be independent if all of the following equations hold:

P(AB)=P(A)P(B),P(AC)=P(A)P(C),P(BC)=P(B)P(C),P(ABC)=P(A)P(B)P(C).\begin{align*} P(A \cap B) &= P(A) P(B), \\ P(A \cap C) &= P(A) P(C), \\ P(B \cap C) &= P(B) P(C), \\ P(A \cap B \cap C) &= P(A) P(B) P(C). \end{align*}

If the first three conditions hold, we say that AA, BB, and CC are pairwise independent. Pairwise independence does not imply independence: it is possible that just learning about AA or just learning about BB is of no use in predicting whether CC occurred, but learning that both AA and BB occurred could still be highly relevant for CC. Here is a simple example of this distinction.

Example Pairwise Independence doesn't Imply Independence

image.png

We can define independence of any number of events similarly. Intuitively, the idea is that knowing what happened with any particular subset of the events gives us no information about what happened with the events not in that subset.

Conditional independence is defined analogously to independence.

Definition: Conditional Independence

Events AA and BB are said to be conditionally independent given event EE if P(ABE)=P(AE)P(BE)P(A \cap B|E) = P(A|E) P(B|E).

Example Conditional Independence doesn't Imply Independence

image.png

Example Independence doesn't Imply Conditional Independence

image.png