01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

22<br />

It is clear that PCA can be used as the basis of a clustering method, by grouping the points<br />

according to how they clump along each projection. Each cluster can then be treated separately.<br />

Figure 2-1 illustrates a situation in which PCA is useful in projecting the dataset onto a new frame<br />

of reference, from which the cluster of datapoints can be easily inferred. Figure 2-4 on the other<br />

hand shows an example in which PCA would fail to segment two clusters. By projecting the data<br />

along the principal direction, i.e. the vertical axis, PCA would merge the two clusters. ICA is one<br />

alternative method for separating such clusters.<br />

Figure 2-4: A classical illustration of problems occurring with variance-based methods such as PCA.<br />

2.1.4 Projection Pursuit<br />

Projection–Pursuit (PP) methods aim at finding structures in multivariate data by projecting them<br />

on a lower-dimensional subspace. Since there are infinitely many projections from a higher<br />

dimension to a lower dimension, PP aims at finding a technique to pursue a finite sequence of<br />

projections that can reveal the most “interesting” structures of the data. The interesting<br />

projections are often those that depart from that of normal/Gaussian distributions. For instance,<br />

these structures can take the form of trends, clusters, hypersurfaces, or anomalies. Figure 2-4<br />

illustrates the ``interestingness'' of non-gaussian projections. The data in this figure is clearly<br />

divided into two clusters. However, the principal component, i.e. the direction of maximum<br />

variance, would be vertical, providing no separation between the clusters. In contrast, the strongly<br />

non-gaussian projection pursuit direction would be horizontal, providing optimal separation of the<br />

clusters. Independent Component Analysis (see Section 2.3) is a method for discovering a linear<br />

decomposition into directions that maximize non-Gaussianity. In this respect, it provides a first<br />

step for PP decomposition.<br />

Typically, PP uses a projection index, a functional computed on a projected density (or data set),<br />

to measure the “interestingness” of the current projection and then uses a numerical optimizer to<br />

move the projection direction to a more interesting position. Interesting projections usually refer to<br />

non-linear structures in the data, i.e. structure than cannot be extracted by linear projections such<br />

as PCA and ICA. PP can be formalized as follows:<br />

Given a dataset<br />

P<br />

P<br />

P P<br />

X ∈° and a unit vector a ∈ ° one defines an index Ia<br />

: ° → ° that<br />

measures the interest of the associated projection P ( )<br />

maximize I a<br />

.<br />

a<br />

X . PP finds the parameters a that<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!