01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

149<br />

6. Decrease the size of the neighborhood and the learning rate<br />

( −r<br />

( )<br />

2<br />

ki , t )<br />

2<br />

2σ<br />

h t = e where s 2 is the<br />

Example of neighborhood function could be ( )<br />

width parameter that can gradually be decreased over time.<br />

7. Unless the net is adequately trained (for example η has become very small),<br />

go back to step 2.<br />

ki ,<br />

6.8.1.2 Normalized weight vectors<br />

The weights in Kohonen net are often normalized to unit length. This allows a somewhat cheaper<br />

computation when determining the winning unit. In effect, the network tries to adjust the direction<br />

of the weight vectors to line up with that of the input. The basic idea is that the distance between<br />

the two vectors is minimal, when the two vectors are aligned. So, minimizing the Euclidean<br />

distance is equivalent to maximizing the vector dot product<br />

r r i .<br />

x⋅<br />

w<br />

6.8.1.3 Caveats for the Kohonen Net<br />

The use of Kohonen nets is not as straightforward as it might appear. Although the net does<br />

much of the work in sorting out the organization implicit in the input space, problems can arise.<br />

The first difficulty is in choosing the initial weights. These are often set at random, but if the<br />

distribution of the weights actually selected does not match well with the organization of the input<br />

space, the net will converge very slowly. One way round to this is to set the initial weights<br />

deliberately to reflect the structure of the input space if that is known. For example, thee unit<br />

weight vectors could be smoothly and uniformly distributed across the input space.<br />

A second problem is in deciding how to vary the neighborhood and learning rate to achieve the<br />

best organization. The answer really depends on the application. Note, though, that the standard<br />

Kohonen training algorithm assumes that the training and performance are separate: first you<br />

train the network and, when that is complete, then you use it to classify inputs. For some<br />

applications (for instance, in robot control) it is more appropriate that the net continue learning<br />

and that it generates classifications even though partially trained.<br />

A third problem concerns the presentation of inputs. The training procedure above presupposes<br />

that inputs are presented fairly and at random – in other words – there is no systematic bias in<br />

favor of one part of the input space against others. Without this condition, the net may not<br />

produce a topographic organization, but rather a single clump of units all organized around the<br />

over represented part of the input space. This problem too arises in robotic applications, where<br />

the world is experienced sequentially and consecutive inputs are not randomly chosen from the<br />

input space. Batching up inputs and presenting them randomly from those batches can solve the<br />

problem.<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!