Kyle I S Harrington / kyle@eecs.tufts.edu
Some slides adapted from Roni Khardon, Andrew Moore, Mohand Said Allili, David Sontag, Frank Dellaert, Michael Jordan, Yair Weiss
Given some data $D$ drawn from an unknown distribution
Estimate the parameters $\theta$ of a GMM that fits the data
How do we find the Gaussian parameters?
2 coins, A and B, with bias to land on heads: $\theta_A$ and $\theta_B$
Repeat for $i$ = 1 to 5:
We can calculate the bias from these observations directly:
$\dot{\theta}_A = \frac{ \textit{# heads using A}}{ \textit{# flips using A}}$
$\dot{\theta}_A = \frac{8}{20}$
$\dot{\theta}_B = \frac{ \textit{# heads using B}}{ \textit{# flips using B}}$
$\dot{\theta}_B = \frac{19}{30}$
This information was recorded as # heads in $x$ and coin identifier in $z$
What if we didn't have $z$?
What if we didn't have $z$ (the coin flipped to produce a certain # of heads)?
$z$ is now a "hidden", "latent" variable
How do we find $\theta_A$ and $\theta_B$?
We know $x_1$, $x_2$, ..., $x_R$ which follow some $N( \mu, \sigma^2 )$
How do we find $\mu$ (assume we know $\sigma^2$)?
We know $x_1$, $x_2$, ..., $x_R$ which follow some $N( \mu, \sigma^2 )$
How do we find $\mu$ (assume we know $\sigma^2$)?
Maximum Likelihood Estimation: For which $\mu$ is $x_1, x_2, ... x_R$ most likely?
Maximum a posteriori: Which $\mu$ maximizes $p(\mu | x_1, x_2, ... x_R, \sigma^2 )$?
Expectation-Maximization (EM) algorithm is an approach to maximize likelihood
Iterate:
Initialization:
Termination:
Limitations
Guest Lecture on Aggregation