01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

145<br />

∂<br />

∂<br />

J 1<br />

= fy −<br />

1 2 λ 1yf<br />

1 1<br />

w1<br />

f ( y 1y<br />

)<br />

1 2 λ 1<br />

J 1<br />

w y f x w f x y<br />

v1<br />

= −<br />

∂<br />

∂<br />

w<br />

2 2<br />

= ( 1− ) −λ<br />

( 1−<br />

)<br />

2<br />

( 1 f ) ( y y )<br />

1 x 2 λ 1<br />

1 2 1 1 1 1 1 1 1<br />

= − −<br />

1 1 1<br />

(6.51)<br />

where in this last equation, all the vector multiplications are to be understood as being on an<br />

element by element basis. Similarly with w 2 , v 2 , and λ 2 . This gives us a method for changing the<br />

weights and the Lagrange multipliers on an online basis. This leads to the joint learning rules:<br />

w<br />

v<br />

( )<br />

f y λ y<br />

Δ = η −<br />

1 1 2 1 1<br />

x w<br />

1i 1i 1i<br />

1<br />

2<br />

( y2 λ y1)( 1 f<br />

1)<br />

Δ = η − −<br />

(6.52)<br />

and similarly with the second set of weights.<br />

6.7.3 ICA Revisited<br />

An important aspect of anti-Hebbian learning is its ability to decorrelate inputs. As discussed in<br />

Chapter 2.1.5, decorrelating the dataset if often too soft a constraint, and, it is often more<br />

desirable to find independent components of the dataset, so as to reduce maximally its<br />

dimensionaly, and, in the case of ANN, so as to maximize the information transmitted by the<br />

network.<br />

Jutten and Herault proposed a neural network architecture, based on anti-Hebbian learning, that<br />

can perform Independent Component Analysis. The activation function is as follows:<br />

Which is equivalent to:<br />

i i ij j<br />

j=<br />

1<br />

n<br />

y x w y<br />

= −∑ (6.53)<br />

y= x− Wy<br />

(6.54)<br />

( ) 1<br />

−<br />

y= I+ W x<br />

(6.55)<br />

While this is similar to the Foldiak’s model, a crucial difference lies in the non-linearity of the<br />

learning rule, defined as follows:<br />

( ) ( ) for<br />

Δ w =−α<br />

f y g y i≠ j<br />

(6.56)<br />

ij i j<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!