06.09.2021 Views

Mind, Body, World- Foundations of Cognitive Science, 2013a

Mind, Body, World- Foundations of Cognitive Science, 2013a

Mind, Body, World- Foundations of Cognitive Science, 2013a

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

idea is to give the network as many degrees <strong>of</strong> freedom as possible to discover representational<br />

regularities that have not been imposed or predicted by the researcher.<br />

These decisions all involve the architectural level <strong>of</strong> investigation.<br />

One issue, though, is that networks are greedy, in the sense that they will exploit<br />

whatever resources are available to them. As a result, fairly idiosyncratic and specialized<br />

detectors are likely to be found if too many hidden units are provided to the<br />

network, and the network’s performance may not transfer well when presented with<br />

novel stimuli. To deal with this, one must impose constraints by looking for the simplest<br />

network that will reliably learn the mapping <strong>of</strong> interest. The idea here is that<br />

such a network might be the one most likely to discover a representation general<br />

enough to transfer the network’s ability to new patterns.<br />

Importantly, sometimes when one makes architectural decisions to seek the<br />

simplest network capable <strong>of</strong> solving a problem, one discovers that the required network<br />

is merely a perceptron that does not employ any hidden units. In the remaining<br />

sections <strong>of</strong> this chapter I provide some examples <strong>of</strong> simple networks that are<br />

capable <strong>of</strong> performing interesting tasks. In section 4.15 the relevance <strong>of</strong> perceptrons<br />

to modern theories <strong>of</strong> associative learning is described. In section 4.16 I present a<br />

perceptron model <strong>of</strong> the reorientation task. In section 4.17 an interpretation is given<br />

for the structure <strong>of</strong> a perceptron that learns a seemingly complicated progression<br />

<strong>of</strong> musical chords.<br />

4.15 New Powers <strong>of</strong> Old Networks<br />

The history <strong>of</strong> artificial neural networks can be divided into two periods, Old<br />

Connectionism and New Connectionism (Medler, 1998). New Connectionism<br />

studies powerful networks consisting <strong>of</strong> multiple layers <strong>of</strong> units, and connections<br />

are trained to perform complex tasks. Old Connectionism studied networks that<br />

belonged to one <strong>of</strong> two classes. One was powerful multilayer networks that were<br />

hand wired, not trained (McCulloch & Pitts, 1943). The other was less powerful<br />

networks that did not have hidden units but were trained (Rosenblatt, 1958, 1962;<br />

Widrow, 1962; Widrow & H<strong>of</strong>f, 1960).<br />

Perceptrons (Rosenblatt, 1958, 1962) belong to Old Connectionism. A perceptron<br />

is a standard pattern associator whose output units employ a nonlinear activation<br />

function. Rosenblatt’s perceptrons used the Heaviside step function to convert<br />

net input into output unit activity. Modern perceptrons use continuous nonlinear<br />

activation functions, such as the logistic or the Gaussian (Dawson, 2004, 2005, 2008;<br />

Dawson et al., 2009; Dawson et al., 2010).<br />

Perceptrons are trained using an error-correcting variant <strong>of</strong> Hebb-style learning<br />

(Dawson, 2004). Perceptron training associates input activity with output unit<br />

error as follows. First, a pattern is presented to the input units, producing output<br />

Elements <strong>of</strong> Connectionist <strong>Cognitive</strong> <strong>Science</strong> 189

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!