06.09.2021 Views

Mind, Body, World- Foundations of Cognitive Science, 2013a

Mind, Body, World- Foundations of Cognitive Science, 2013a

Mind, Body, World- Foundations of Cognitive Science, 2013a

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

esponses to determine the trigger features that the hidden units detect. It is also<br />

shown that changing the activation function <strong>of</strong> a hidden unit can lead to interesting<br />

complexities in defining the notion <strong>of</strong> a trigger feature, because some kinds <strong>of</strong><br />

hidden units capture families <strong>of</strong> trigger features that require further analysis.<br />

In Section 4.12 we describe how interpreting the internal structure <strong>of</strong> a network<br />

begins to shed light on the relationship between algorithms and architectures. Also<br />

described is a network that, as a result <strong>of</strong> training, translates a classical model <strong>of</strong> a<br />

task into a connectionist one. This illustrates an intertheoretic reduction between<br />

classical and connectionist theories, raising the possibility that both types <strong>of</strong> theories<br />

can be described in the same architecture.<br />

4.9 Empiricism and Internal Representations<br />

The ability <strong>of</strong> hidden units to increase the computational power <strong>of</strong> artificial neural<br />

networks was well known to Old Connectionism (McCulloch & Pitts, 1943). Its problem<br />

was that while a learning rule could be used to train networks with no hidden<br />

units (Rosenblatt, 1958, 1962), no such rule existed for multilayered networks. The<br />

reason that a learning rule did not exist for multilayered networks was because<br />

learning was defined in terms <strong>of</strong> minimizing the error <strong>of</strong> unit responses. While it<br />

was straightforward to define output unit error, no parallel definition existed for<br />

hidden unit error. A hidden unit’s error could not be defined because it was not<br />

related to any directly observable outcome (e.g., external behaviour). If a hidden<br />

unit’s error could not be defined, then Old Connectionist rules could not be used to<br />

modify its connections.<br />

The need to define and compute hidden unit error is an example <strong>of</strong> the credit<br />

assignment problem:<br />

In playing a complex game such as chess or checkers, or in writing a computer<br />

program, one has a definite success criterion—the game is won or lost. But in the<br />

course <strong>of</strong> play, each ultimate success (or failure) is associated with a vast number <strong>of</strong><br />

internal decisions. If the run is successful, how can we assign credit for the success<br />

among the multitude <strong>of</strong> decisions? (Minsky, 1963, p. 432)<br />

The credit assignment problem that faced Old Connectionism was the inability<br />

to assign the appropriate credit—or more to the point, the appropriate blame—<br />

to each hidden unit for its contribution to output unit error. Failure to solve this<br />

problem prevented Old Connectionism from discovering methods to make their<br />

most powerful networks belong to the domain <strong>of</strong> empiricism and led to its demise<br />

(Papert, 1988).<br />

The rebirth <strong>of</strong> connectionist cognitive science in the 1980s (McClelland &<br />

Rumelhart, 1986; Rumelhart & McClelland, 1986c) was caused by the discovery<br />

Elements <strong>of</strong> Connectionist <strong>Cognitive</strong> <strong>Science</strong> 159

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!