MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
149<br />
6. Decrease the size of the neighborhood and the learning rate<br />
( −r<br />
( )<br />
2<br />
ki , t )<br />
2<br />
2σ<br />
h t = e where s 2 is the<br />
Example of neighborhood function could be ( )<br />
width parameter that can gradually be decreased over time.<br />
7. Unless the net is adequately trained (for example η has become very small),<br />
go back to step 2.<br />
ki ,<br />
6.8.1.2 Normalized weight vectors<br />
The weights in Kohonen net are often normalized to unit length. This allows a somewhat cheaper<br />
computation when determining the winning unit. In effect, the network tries to adjust the direction<br />
of the weight vectors to line up with that of the input. The basic idea is that the distance between<br />
the two vectors is minimal, when the two vectors are aligned. So, minimizing the Euclidean<br />
distance is equivalent to maximizing the vector dot product<br />
r r i .<br />
x⋅<br />
w<br />
6.8.1.3 Caveats for the Kohonen Net<br />
The use of Kohonen nets is not as straightforward as it might appear. Although the net does<br />
much of the work in sorting out the organization implicit in the input space, problems can arise.<br />
The first difficulty is in choosing the initial weights. These are often set at random, but if the<br />
distribution of the weights actually selected does not match well with the organization of the input<br />
space, the net will converge very slowly. One way round to this is to set the initial weights<br />
deliberately to reflect the structure of the input space if that is known. For example, thee unit<br />
weight vectors could be smoothly and uniformly distributed across the input space.<br />
A second problem is in deciding how to vary the neighborhood and learning rate to achieve the<br />
best organization. The answer really depends on the application. Note, though, that the standard<br />
Kohonen training algorithm assumes that the training and performance are separate: first you<br />
train the network and, when that is complete, then you use it to classify inputs. For some<br />
applications (for instance, in robot control) it is more appropriate that the net continue learning<br />
and that it generates classifications even though partially trained.<br />
A third problem concerns the presentation of inputs. The training procedure above presupposes<br />
that inputs are presented fairly and at random – in other words – there is no systematic bias in<br />
favor of one part of the input space against others. Without this condition, the net may not<br />
produce a topographic organization, but rather a single clump of units all organized around the<br />
over represented part of the input space. This problem too arises in robotic applications, where<br />
the world is experienced sequentially and consecutive inputs are not randomly chosen from the<br />
input space. Batching up inputs and presenting them randomly from those batches can solve the<br />
problem.<br />
© A.G.Billard 2004 – Last Update March 2011