01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

120<br />

6 Artificial Neural Networks<br />

Artificial neural network (ANN) has become one of those buzzwords that is either popular or<br />

unpopular, but leaves no one indifferent. Unfortunately, for most people, the word relates only to<br />

one type of neural network, namely feed-forward neural networks. ANNs are, however, far more<br />

than this. We will see, in this chapter and the next one, different ANN architectures that greatly<br />

differ from one another, both in terms of computation they can perform as well as in their<br />

algorithms. We will see how the architecture and the learning rule determine the type of<br />

computation an ANN can perform.<br />

6.1 Applications of ANN<br />

ANNs can perform diverse types of computation. Here is a non-exhaustive list:<br />

Type of Computation<br />

Pattern recognition<br />

Associative memory<br />

PCA<br />

ICA<br />

Dimensionality Reduction<br />

Learning Time series<br />

ANNs<br />

Feed-forward ANN and Perceptron<br />

Hebbian Network, Willshaw Net, Hopfield Net<br />

Hebbian Network<br />

Anti-hebbian learning<br />

SOM, Kohonen<br />

TDNN, RNN<br />

ANNs are powerful tools that can deal with large, multidimensional, non-linear, datasets. They<br />

have found application in numerous domains, as statistical tools for data mining in finance and<br />

particle physics or as optimization tools in industrial control systems and robotics.<br />

There are two modes of functioning for an ANN:<br />

1. Activation transfer mode when activation is transmitted throughout the network.<br />

2. Learning mode when the network organizes itself, usually on the basis of the most<br />

recent activation transfer.<br />

6.2 Biological motivation<br />

ANNs were developed with the goal to mimic the highly parallel and distributed computation<br />

performed in the human brain.<br />

6.2.1 The Brain as an Information Processing System<br />

The human brain contains about 10 billion nerve cells, or neurons. On average, each neuron is<br />

connected to other neurons through about 10 000 synapses. (The actual figures vary greatly,<br />

depending on the local neuroanatomy.) The brain's network of neurons forms a massively parallel<br />

information processing system. This contrasts with conventional computers, in which a single<br />

processor executes a single series of instructions.<br />

Against this, consider the time taken for each elementary operation: neurons typically operate at<br />

a maximum rate of about 100 Hz, while a conventional CPU carries out several hundred million<br />

machine level operations per second. Despite of being built with very slow hardware, the brain<br />

has quite remarkable capabilities:<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!