MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
137<br />
Figure 6-11: The weight vector moves toward the input vector<br />
The outstar rule, where the decay term is gated by the output term<br />
y<br />
i<br />
.<br />
dw<br />
dt<br />
ij<br />
{ }<br />
= α x − w y<br />
(6.28)<br />
j ij i<br />
In the discrete case, we have:<br />
( )<br />
Δ w = α ⋅x ⋅ y − w<br />
(6.29)<br />
ij i i ij<br />
( )<br />
y = 1 ⇒ Δ w = α ⋅ y −w<br />
i ij j ij<br />
( ) (1 α) ( 1)<br />
⇒ w t = − ⋅w t− + α⋅y<br />
ij ij j<br />
(6.30)<br />
In this case, the weight vector moves toward the output vector.<br />
6.6.3 Principal Components<br />
Let us consider again the activation function of the Perceptron and rewrite as follows:<br />
where<br />
i<br />
r r<br />
y = w x = w ⋅x<br />
∑<br />
i ij j<br />
i<br />
w r is the weight vector along i and x r the input vector. We have:<br />
whereθ is the angle across the two vector.<br />
r r r r<br />
w ⋅ x=<br />
w x cos θ<br />
i<br />
i<br />
This quantity is maximal when the angle θ is zero, i.e. when the two vectors are aligned. Thus, if<br />
the weight converge towards the principal components of the input space (i.e. w r is the 1 st<br />
1<br />
eigenvector, w r the second, etc), then the first output vector y<br />
2<br />
1<br />
will transmit maximally the input<br />
information along the direction with largest variance. In other words, if we define a learning rule,<br />
such that it projects the weights along the principal components of the input space, such a<br />
network would in effect produce a PCA analysis.<br />
i<br />
( )<br />
© A.G.Billard 2004 – Last Update March 2011