A Gaussian Mixture Model (GMM), which is a generative model for data
, is defined by the following set of parameters:
- K Number of mixture components
- A d-dimensional Gaussian
for every 
: Mixture weights- The parameters of a K-component GMM can be represented as:

- The likelihood of a point
in a GMM is given as![Rendered by QuickLaTeX.com \[\circ \quad p(\mathbf{x} \mid \theta)=\sum_{j=1}^{K} p_{j} \mathcal{N}\left(\mathbf{x}, \mu^{(j)}, \sigma_{j}^{2}\right)\]](https://ayoubb.com/wp-content/ql-cache/quicklatex.com-885f20237adbbc91380581bb8cdd9a48_l3.png)
Observed case
Let
and let
be three observed points in cluster 1 and
be three observed points in cluster 2 .
Unobserved Case: EM Algorithm
Estimates of Parameters of GMM: The Expectation Maximization (EM) Algorithm
We observe
data points
in
, we would like to maximize the GMM likelihood with respect to the parameter set:
![]()
Maximizing
The EM algorithm is an iterative algorithm that finds a locally optimal solution
E-Step
Assume that the initial means and variances of two clusters in a GMM are as follows:
![]()
Let
Let
Using the formula of E-Step:
![Rendered by QuickLaTeX.com \[p(j \mid i)=\frac{p_{j} \mathcal{N}\left(x^{(i)} ; \mu^{(j)}, \sigma_{j}^{2}\right)}{p\left(x^{(i)} \mid \theta\right)}\]](https://ayoubb.com/wp-content/ql-cache/quicklatex.com-452c82954a0e31177827accf1428abf9_l3.png)
![]()
![]()
![]()
![]()
![]()
M-Step
Using the formulae corresponding to the M-step,
![]()
![]()
![]()
A Gaussian mixture model can provide information about how likely it is that a given point belongs to each cluster.