Probability
Learning Objectives
Learning objectives of the Probability section.
Summary Table
Summary of the Axiomatic Probability section.
Concept | Description | Example |
---|---|---|
Sample Space (\(S\)) | Set of all possible outcomes of an experiment | \(S = \{Heads, Tails\}\) |
Event (\(A\)) | A subset of the sample space | \(A = \{1, 3, 5\}\) when rolling a die |
Outcome | A single result from the sample space | \(Heads \in \{Heads, Tails\}\) |
Axiom 1: Nonnegativity | Probability of any event is ≥ 0 | \(P(A) \geq 0\) |
Axiom 2: Normalization | Probability of the sample space is 1 | \(P(S) = 1\) |
Axiom 3: Additivity | For disjoint events, the probability of their union is the sum of parts | \(P(A \cup B) = P(A) + P(B)\) if \(A \cap B = \emptyset\) |
Conditional Probability | Probability of \(A\) given \(B\) has occurred | \(P(A|B) = \frac{P(A \cap B)}{P(B)}\) |
Product Rule | Probability of intersection using conditional probability | \(P(A \cap B) = P(A|B)P(B)\) |
Total Probability Theorem | Compute probability over partitions of the sample space | \(P(B) = \sum_i P(B|A_i)P(A_i)\) |
Bayes’ Theorem | Reverse conditional probability using prior and likelihood | \(P(S|W) = \frac{P(W|S)P(S)}{P(W|S)P(S) + P(W|NS)P(NS)}\) |
Independent Events | Events that do not affect each other’s probabilities | \(P(A \cap B) = P(A)P(B)\) if \(A \perp B\) |
Conditioning and Independence | Independence may break down when conditioning on a third event | \(A \perp B\) might not imply \(A \perp B | C\) |
Summary of the Random Variables section.
Concept | Discrete Random Variables | Continuous Random Variables |
---|---|---|
Sample Space | Countable outcomes | Uncountable outcomes |
Domain of Variable | \(x \in \mathbb{Z}\) | \(x \in \mathbb{R}\) |
Mapping | \(X: \text{Outcome} \rightarrow x \in \mathbb{Z}\) | \(X: \text{Event} \rightarrow x \in \mathbb{R}\) |
Probability Function | Probability Mass Function (PMF):\(p_X(x) = P(X = x)\) | Probability Density Function (PDF):\(f_X(x) \geq 0\) |
Total Probability | \(\sum_x p_X(x) = 1\) | \(\int_{-\infty}^{\infty} f_X(x) \, dx = 1\) |
CDF (Cumulative Distribution) | \(F_X(x) = \sum_{k \leq x} p_X(k)\) | \(F_X(x) = \int_{-\infty}^{x} f_X(u) \, du\) |
Probability of Exact Value | \(P(X = x) = p_X(x) > 0\)possible | \(P(X = x) = 0\) for any exact \(x\) |
Expectation | \(\mathbb{E}[X] = \sum_x x \, p_X(x)\) | \(\mathbb{E}[X] = \int_{-\infty}^{\infty} x \, f_X(x) \, dx\) |
Variance | \(\text{Var}[X] = \sum_x (x - \mathbb{E}[X])^2 \, p_X(x)\) | \(\text{Var}[X] = \int_{-\infty}^{\infty} (x - \mathbb{E}[X])^2 \, f_X(x) \, dx\) |
Joint Distribution | Joint PMF:\(p_{X,Y}(x, y) = P(X = x, Y = y)\) | Joint PDF:\(f_{X,Y}(x, y)\) |
Marginal Distribution | \(p_X(x) = \sum_y p_{X,Y}(x, y)\) | \(f_X(x) = \int_{-\infty}^{\infty} f_{X,Y}(x, y) \, dy\) |
Conditional Probability | \(p_{X|Y}(x|y) = \frac{p_{X,Y}(x,y)}{p_Y(y)}\) | \(f_{X|Y}(x|y) = \frac{f_{X,Y}(x,y)}{f_Y(y)}\) |
Conditional Expectation | \(\mathbb{E}[X|Y=y] = \sum_x x \, p_{X|Y}(x|y)\) | \(\mathbb{E}[X|Y=y] = \int x \, f_{X|Y}(x|y) \, dx\) |
Independence | \(p_{X,Y}(x,y) = p_X(x) \cdot p_Y(y)\) | \(f_{X,Y}(x,y) = f_X(x) \cdot f_Y(y)\) |
Summary of the Distributions section.
Distribution | Type | Support | Parameters | PDF / PMF | Common Use Case |
---|---|---|---|---|---|
Bernoulli | Discrete | \(x \in \{0, 1\}\) | \(p\) (success probability) | \(p_X(x) = \begin{cases} p & \text{if } x = 1, \\ 1 - p & \text{if } x = 0 \end{cases}\) | Binary outcomes (e.g., success/failure, yes/no) |
Gaussian (Normal) | Continuous | \(x \in (-\infty, \infty)\) | \(\mu\) (mean), \(\sigma^2\) (variance) | \(f_X(x)=\frac{1}{\sqrt{2\pi\sigma^{2}}}e^{-\frac{(x-\mu)^2}{2\sigma^2}}\) | Modeling natural phenomena, basis of CLT |
Beta | Continuous | \(x \in (0, 1)\) | \(\alpha\), \(\beta\) | \(f_X(x) = \frac{\Gamma(\alpha + \beta)}{\Gamma(\alpha)\Gamma(\beta)}x^{\alpha - 1}(1 - x)^{\beta - 1}\) | Bayesian priors for probabilities, modeling proportions |