01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

51<br />

k<br />

( σ<br />

j )<br />

2<br />

=<br />

k<br />

∑<br />

i<br />

α =<br />

r<br />

k<br />

( x ) 2<br />

,<br />

− µ<br />

k<br />

i i j j<br />

N⋅<br />

k<br />

i<br />

∑<br />

∑ r<br />

i<br />

∑∑<br />

i<br />

k<br />

i<br />

r<br />

r<br />

k<br />

i<br />

k<br />

i<br />

(3.12)<br />

(3.13)<br />

Soft K-means and the mixture of Gaussians are two examples of maximum likelihood algorithm.<br />

Figure 3-13: Examples of clustering with Mixtures of Gaussians using spherical Gaussians (left) and nonspherical<br />

Gaussians (i.e. with full covariance matrix) (right). Notice how the clusters become elongated<br />

along the direction of the clusters (the grey circles represent the first and second variances of the<br />

distributions).<br />

[DEMOS\CLUSTERING\GMM-ELONGATED.ML]<br />

A fatal flaw of maximum likelihood (KABOOM!):<br />

The major drawback of maximizing the likelihood (people tend to actually minimize the –log of the<br />

likelihood) is that there is no actual maximum. Indeed, the pdf is a positive but unbounded<br />

function. As a result, one may observe the so-called KABOOM effect.<br />

When the mean of one of the Gaussian is located on a particular data point, and, if the variance is<br />

2<br />

2<br />

already very smallσ<br />

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!