Core Probability Reference
Definition: Empirical Definition of Probability
The probability of any event $E$ can be defined as:$$ \p(E) = \lim_{n \rightarrow \infty} \frac {\text{count}(E)} {n} $$
Where $\text{count}(E)$ is the number of times that $E$ occured in $n$ experiments.
Definition: Core Identities
For an event $E$ and a sample space $S$
$0 ≤ \p(E) ≤ 1$ | All probabilities are numbers between 0 and 1. |
$\p(S) = 1$ | All outcomes must be from the Sample Space. |
$\P(E) = 1 - \P(E^\c)$ | The probability of an event from its complement. |
Definition: Probability of Equally Likely Outcomes
If $S$ is a sample space with equally likely outcomes, for an event $E$ that is a subset of the outcomes in $S$: $$ \begin{align} \p(E) &= \frac{\text{number of outcomes in $E$}}{\text{number of outcomes in $S$}} = \frac{|E|}{|S|} \end{align} $$
The probability of $E$ given that (aka conditioned on) event $F$ already happened: $$ \p(E |F) = \frac{\p(E \and F)}{\p(F)} $$
If two events $E$ and $F$ are mutually exclusive then the probability of $E$ or $F$ occurring is: $$ \p(E \or F) = \p(E) + \p(F) $$
For $n$ events $E_1, E_2, \dots E_n$ where each event is mutually exclusive of one another (in other words, no outcome is in more than one event). Then: $$ \p(E_1 \or E_2 \or \dots \or E_n) = \p(E_1) + \p(E_2) + \dots + \p(E_n) = \sum_{i=1}^n \p(E_i) $$
For any two events $E$ and $F$: $$ \p(E \or F) = \p(E) + \p(F) − \p(E \and F) $$
For three events, $E$, $F$, and $G$ the formula is: $$ \begin{align} \p(E \or F \or G) =& \text{ }\p(E) + \p(F) + \p(G) \\ & −\p(E \and F) − \p(E \and G)−P(F \and G) \\ & +\p(E \and F \and G) \end{align} $$
For more than three events see the chapter of probability of or.
If two events: $E$, $F$ are independent then the probability of $E$ and $F$ occurring is: $$ \p(E \and F) = \p(E) \cdot \p(F) $$
For $n$ events $E_1, E_2, \dots E_n$ that are independent of one another: $$ \p(E_1 \and E_2 \and \dots \and E_n) = \prod_{i=1}^n \p(E_i) $$
For any two events $E$ and $F$: $$ \p(E \and F) = \p(E | F) \cdot \p(F) $$
For $n$ events $E_1, E_2, \dots E_n$: $$ \begin{align} \p(E_1 \and E_2 \and \dots \and E_n) = &\p(E_1) \cdot \p(E_2|E_1) \cdot \p(E_3 |E_1 \and E_2) \dots \\ &\p(E_n|E_1 \dots E_{n−1}) \end{align} $$
For any two events $E$ and $F$: $$ \begin{align} \p(E) &= \p(E \and F) + \p(E \and F\c)\\ &=\p(E | F) \p(F) + \p(E | F\c) \p(F\c) \end{align} $$
For mutually exclusive events: $B_1, B_2, \dots B_n$ such that every outcome in the sample space falls into one of those events: $$ \begin{align} \p(E) &= \sum_{i=1}^n \p(E \and B_i) && \text{Extension of our observation}\\ &= \sum_{i=1}^n \p(E | B_i) \p(B_i) && \text{Using chain rule on each term} \end{align} $$
The most common form of Bayes' Theorem is Bayes' Theorem Classic: $$ \p(B|E) = \frac{\p(E | B) \cdot \p(B)}{\p(E)} $$
Bayes' Theorem combined with the Law of Total Probability: $$ \p(B|E) = \frac{\p(E | B) \cdot \p(B)}{\p(E|B)\cdot \p(B) + \p(E|B\c) \cdot \p(B\c)} $$