17.01.2013 Views

Chapter 2. Prehension

Chapter 2. Prehension

Chapter 2. Prehension

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Appendix C - Computational Neural Modelling 391<br />

Another common rule which uses the presynaptic signal and<br />

postsynaptic error is called the delta rule4 (or Widrow-Hoff rule), and<br />

it uses a target value as follows (right side of Figure C.5):<br />

Instead of using the output of neuron i to determine by how much to<br />

change the weights, the actual output Oi is first subtracted from the<br />

desired output ti. To do this, supervised learning is used. A small<br />

training set is provided to the network for learning some part of the<br />

task space. The training set is a collection of input/output pairs (or<br />

patterns) that identifies to the network what the output should be for a<br />

given set of inputs. For linear units, this rule minimizes the squares of<br />

the differences between the actual and desired output values summed<br />

over the output neuons and all pairs of input/output vectors<br />

(Rumelhart, Hinton & Williams, 1986a).<br />

The delta rule can be derived from a technique called gradient<br />

descent. ‘Gradient descent’ means ‘go down hill’ in a mathematical<br />

sense. For a traveller in a hilly area, define z to be the person’s<br />

altitude, x to be longitude (i.e., how far east she is located) and y to be<br />

latitude (how far north she is). Then z is a function of x and y: the<br />

altitude of the ground varies with its x-y location. The variables x and<br />

y are independent ones and the traveler can shift these at will by<br />

walking. Since the traveler’s altitude changes as x and y do, z is the<br />

dependent variable. The gradient, or change, of z is the vector of<br />

partial derivatives of z with respect to its independent variables:<br />

(az/ax, az/ay). In this case the gradient represents a horizontal vector<br />

which points in the direction of travel for which z increases most<br />

rapidly. The gradient descent rule simply says “move in the direction<br />

opposite to which the gradient vector points”. Mathematically, it is<br />

written as:<br />

4An early model was a percepuon, which consisted of linear threshold units using<br />

the delta rule for synaptic weight modification mosenblatt, 1962).

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!