17.01.2013 Views

Chapter 2. Prehension

Chapter 2. Prehension

Chapter 2. Prehension

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Appendix C - Computational Neural Modelling 397<br />

O=WI (16)<br />

Now, if there are p input/output pairs, we can define two n x p<br />

matrices, 1. and 0, where the columns of 1. are the p input patterns and<br />

the columns of 0 are the corresponding p output patterns. Then we<br />

have<br />

for correct performance. But this equation also defines W and<br />

answers our first question. By inverting it we obtain:<br />

which gives an explicit formula for computing W (although it is<br />

necessary to perform the time consuming task of inverting I.) It also<br />

provides an upper limit on p (the memory’s capacity), answering the<br />

second question: I is only guaranteed to be invertible if pln. For the<br />

remainder of the discussion we will assume p=n, that is the matrices<br />

are square. The input patterns must then be linearly independent,<br />

since I must be invertible (its determinant must be non-zero).<br />

Computing W can be time consuming, as mentioned above, but a<br />

special case occurs when all the input patterns are orthonormal. In this<br />

case, L-l =LT, and the computation of W simply requires a single<br />

matrix multiplication. The orthonormal condition is often assumed<br />

true when the patterns are binary, long (large n), and sparse (mostly<br />

zero), because they tend to be orthogonal (their dot products are likely<br />

to be zero). If they all have about the same number of one’s, then they<br />

can all be scaled down by that number to be roughly normal (of unit<br />

magnitude). The resulting performance of the memory is close to<br />

correct. For non-orthonormal situations, there is the option of<br />

applying a gradient descent technique (e.g. the delta rule) to iteratively<br />

learn the weight matrix.<br />

C.4 Processing Example<br />

As was shown in the previous section, a heteroassociative memory<br />

allows two patterns to be associated with one another, such that one<br />

pattern (the input pattern) triggers the other pattern to appear at the<br />

output units. For this example, a network with three inputs and three<br />

outputs (all linear) will be used, as seen in Figure C.9. Assume your

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!