MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
22<br />
It is clear that PCA can be used as the basis of a clustering method, by grouping the points<br />
according to how they clump along each projection. Each cluster can then be treated separately.<br />
Figure 2-1 illustrates a situation in which PCA is useful in projecting the dataset onto a new frame<br />
of reference, from which the cluster of datapoints can be easily inferred. Figure 2-4 on the other<br />
hand shows an example in which PCA would fail to segment two clusters. By projecting the data<br />
along the principal direction, i.e. the vertical axis, PCA would merge the two clusters. ICA is one<br />
alternative method for separating such clusters.<br />
Figure 2-4: A classical illustration of problems occurring with variance-based methods such as PCA.<br />
2.1.4 Projection Pursuit<br />
Projection–Pursuit (PP) methods aim at finding structures in multivariate data by projecting them<br />
on a lower-dimensional subspace. Since there are infinitely many projections from a higher<br />
dimension to a lower dimension, PP aims at finding a technique to pursue a finite sequence of<br />
projections that can reveal the most “interesting” structures of the data. The interesting<br />
projections are often those that depart from that of normal/Gaussian distributions. For instance,<br />
these structures can take the form of trends, clusters, hypersurfaces, or anomalies. Figure 2-4<br />
illustrates the ``interestingness'' of non-gaussian projections. The data in this figure is clearly<br />
divided into two clusters. However, the principal component, i.e. the direction of maximum<br />
variance, would be vertical, providing no separation between the clusters. In contrast, the strongly<br />
non-gaussian projection pursuit direction would be horizontal, providing optimal separation of the<br />
clusters. Independent Component Analysis (see Section 2.3) is a method for discovering a linear<br />
decomposition into directions that maximize non-Gaussianity. In this respect, it provides a first<br />
step for PP decomposition.<br />
Typically, PP uses a projection index, a functional computed on a projected density (or data set),<br />
to measure the “interestingness” of the current projection and then uses a numerical optimizer to<br />
move the projection direction to a more interesting position. Interesting projections usually refer to<br />
non-linear structures in the data, i.e. structure than cannot be extracted by linear projections such<br />
as PCA and ICA. PP can be formalized as follows:<br />
Given a dataset<br />
P<br />
P<br />
P P<br />
X ∈° and a unit vector a ∈ ° one defines an index Ia<br />
: ° → ° that<br />
measures the interest of the associated projection P ( )<br />
maximize I a<br />
.<br />
a<br />
X . PP finds the parameters a that<br />
© A.G.Billard 2004 – Last Update March 2011