Probability
Table of Contents
1. Definitions
We need a Probability Space (aka Measure Space) to rigorously define probability.
Exclusive Events:
Iff, \(E_1 \cap E_2 = \phi\)
Independent Events:
Iff, \(P(E_1 \cap E_2) = P(E_1) \times P(E_2)\)
Independence is a quality we assume based on the problem. It doesn't derive from other things.
Mutual Independence:
Set of events \(E = \{E_1, E_2, ... E_k\}\) are mutually independent iff,
\[ \forall S \subset E, \ \ P \left( \bigcap_{E_i \in S} E_i \right) = \prod_{E_i \in S} P(E_i) \]
Mutual independence is stronger condition than pairwise independence. It is possible to have pairwise independent events that are not mutually independent.
Random Variable:
A random variable \(X\) is a function \(X: \Omega \to \mathbb{R}\). A random variable is neither random, nor a variable, hence a misnomer. It is a deterministic mapping from sample space to real line.
3. James Stein Estimator
Let \(y = \theta + \sigma \epsilon\) , where \(e \sim \mathcal{N}(0, \sigma^2\mathbf{I})\) is d-dim.
James-Stein estimator gives lower MSE than maximum likelihood estimator (for \(d \geq 3\))
\[ \hat{\theta}_{\text{JS}} = \left(1 - \frac{(d-2)\sigma^2}{\|Y\|^2}\right) Y \]