MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA MACHINE LEARNING TECHNIQUES - LASA

01.11.2014 Views

124 6.3 Perceptron The earliest neural network models go back to 1940 with the McCulloch & Pitts perceptron. The perceptron consists of a simple threshold unit taking binary inputs X = { x x } r 1 ,..., n . Its output y is the product of its entries multiplied by a (1-dim) weight matrix W, modulated by a transfer function or an activity function f. n ⎛⎛ ⎞⎞ y= f ⎜⎜ w ⋅ x = f W⋅X ⎝⎝ i= 0 ⎠⎠ ∑ i i⎟⎟ ( ) (6.1) where wxis the bias and is generally negative. 0 0 Figure 6-1: Perceptron with constant bias Classical transfer functions are: The linear function f( x) The step function = x ⎧⎧1 if x ≥ 0 f( x) = ⎨⎨ ⎩⎩0 if x < 0 © A.G.Billard 2004 – Last Update March 2011

125 The sigmoid f x ( x) ( ) = tanh ∈[ − 1,1] Or its positive form: 1 f( x) = ∈[0,1] Dx 1+ e − ⋅ , where D defines the slope of the function 2 ( ) f( x) = e −ax Radial Basis Functions A radial basis function is simply a Gaussian. The sigmoid is probably the most popular activation function, as it is differentiable; this is very important once one attempts to prove the convergence of the system. Figure 6-2: The XOR problem is unsolvable using a perceptron and a sigmoid activation function (left). However, by chosing a Radial Basis Function (RBF) it is possible to obtain a working solution (right). Note that the parameter a of the RBF kernel needs to be suitably adapted to the size of the data. [DEMOS\CLASSIFICATION\MLP-XOR.ML] Exercise: Can you find suitable weights that would make the perceptron behave like an ANDgate, NOR-gate and NAND-gate? © A.G.Billard 2004 – Last Update March 2011

125<br />

The sigmoid f x ( x)<br />

( ) = tanh ∈[ − 1,1]<br />

Or its positive form:<br />

1<br />

f( x) = ∈[0,1]<br />

Dx<br />

1+<br />

e − ⋅<br />

, where D defines the slope of the function<br />

2<br />

( )<br />

f( x)<br />

= e −ax<br />

Radial Basis Functions<br />

A radial basis function is simply a Gaussian.<br />

The sigmoid is probably the most popular activation function, as it is differentiable; this is very<br />

important once one attempts to prove the convergence of the system.<br />

Figure 6-2: The XOR problem is unsolvable using a perceptron and a sigmoid activation function (left).<br />

However, by chosing a Radial Basis Function (RBF) it is possible to obtain a working solution (right). Note<br />

that the parameter a of the RBF kernel needs to be suitably adapted to the size of the data.<br />

[DEMOS\CLASSIFICATION\MLP-XOR.ML]<br />

Exercise: Can you find suitable weights that would make the perceptron behave like an ANDgate,<br />

NOR-gate and NAND-gate?<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!