Machine Learning
Bayes’ rules, Conditional probability, Chain rule

Bayes’ rules, Conditional probability, Chain rule

  • Tutorial

In the previous tutorial you got introduced to basic probability and the rules dealing with it. Now we are equipped with the ability to calculate probability of events when they are not dependent on any other events around them. But this definitely creates a practical limitation as many events are contingent on each other in reality. This tutorial dealing with conditional probability and bayes' theorem will answer these limitations.

Conditional Probability
Conditional probability as the name suggests, comes into play when the probability of occurrence of a particular event changes when one or more conditions are satisfied (these conditions again are events). Speaking in technical terms, if X and Y are two events then the conditional probability of X w.r.t Y is denoted by $$P(X|Y)$$. So when we talk in terms of conditional probability, just for an example, we make a statement like "The probability of event X given that Y has already occurred".

What if X and Y are independent events?
From the definition of independent events, the occurrence of event X is not dependent on event Y. Therefore $$P(X|Y) = P(X)$$.

What if X and Y are mutually exclusive?
As X and Y are disjoint events, the probability that X will occurs when Y has already occurred is 0. Therefore, $$P(X|Y) = 0$$

Product rule
Product rule states that,

\begin{equation} P(X \cap Y) = P(X|Y)*P(Y) \end{equation}

So the joint probability that both X and Y will occur is equal to the product of two terms:

Probability that event Y occurs.
Probability that X occurs given that Y has already occurred.

From the product rule, the following can be inferred,

$$X \subseteq Y$$ implies $$P(X|Y) = P(X)/P(Y)$$
$$Y \subseteq X$$ implies $$P(X|Y) = 1$$

Note: The distributive, associative and De Morgans laws are valid for probability calculation too. The following can be inferred using these laws,
$$P(X \cup Y|Z) = P(X|Z) + P(Y|Z) - P(X \cap Y|Z)$$
$$P(X^{c}|Z) = 1-P(X|Z)$$

Chain rule
Generalizing the product rule leads to the chain rule. Let $$E_{1}, E_{2},....E_{n}$$ be n events. The joint probability of all the n events is given by,

\begin{equation} P(\bigcap_{i=1,..,n}E_{i}) = P(E_{n}|\bigcap_{i=1,..,n-1}E_{i})*P(\bigcap_{i=1,..,n-1}E_{i}) \end{equation}

The chain rule can be used iteratively to calculate the joint probability of any no.of events.

Bayes' theorem
From the product rule, $$P(X \cap Y) = P(X|Y)P(Y)$$ and $$P(Y \cap X) = P(Y|X)P(X)$$. As $$P(X \cap Y)$$ and $$P(Y \cap X)$$ are both same, \begin{equation} P(Y|X) = \frac{P(X|Y)*P(Y)}{P(X)} \end{equation}

where, $$P(X) = P(X \cap Y) + P(X \cap Y^{c})$$ from sum rule.

View All Notifications