MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
154<br />
If<br />
( ) = ( − )<br />
⎛⎛ ⎞⎞<br />
xi ( t+ 1 ) = θ ( ai ( t)<br />
) ai ( t) =⎜⎜∑<br />
wjixj<br />
( t)<br />
⎟⎟<br />
⎝⎝ j ⎠⎠<br />
x t+ 1 = x t ∀ j ≠i<br />
j<br />
( ) ( )<br />
j<br />
x t x t 1 for all units j, then the network has reached a stable<br />
j<br />
j<br />
state, otherwise we keep changing the state of the units until it converges.<br />
Note that the properties of the Hopfield network may be sensitive to the choice of synchronous<br />
versus asynchronous activation.<br />
Figure 6-14: Hopfield Network trained on 4 patterns (top row). (middle row) increasing amounts of noise are<br />
added to one of the patterns. (bottom row) Hopfield network after convergence from the noisy state. Notice<br />
how the network is able to retrieve the original pattern even when most of the image is noise. The rightmost<br />
image is pure random noise, the network converges to a spurious state due to a local minimum in the<br />
network energy. [DEMOS\HOPFIELD\HOPFIELD.EXE]<br />
6.9.4 Capacity of the static Hopfield Network<br />
One can show that the maximal number of patterns N that can be stored in a network of size K is:<br />
Nmax 0.138⋅<br />
K<br />
; (6.70)<br />
The capacity of the Hopfield network decreases importantly in the face of correlated patterns.<br />
One way of suppressing the effect of correlated patterns is to modify the learning rule in order to<br />
decorrelate the patterns. The learning rule becomes:<br />
1 N N<br />
ji i j<br />
K<br />
µν<br />
µ = 1ν=<br />
1<br />
µ −1<br />
ν<br />
= ∑∑ ( )<br />
(6.71)<br />
w x C x<br />
© A.G.Billard 2004 – Last Update March 2011