22.12.2012 Views

3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures

3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures

3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Chem. Listy, 102, s265–s1311 (2008) Food Chemistry & Biotechnology<br />

L08 ARTIFICIAL NEuRAL NETwORKS IN <strong>FOOD</strong><br />

ANALySIS<br />

Ján MOCáK a,b , VIERA MRáZOVá a , DáŠA<br />

KRUŽLICOVá b and FILIP KRAIC a<br />

a Department of Chemistry, University of Ss. Cyril & Methodius,<br />

Nám. J. Herdu 2, 917 01 Trnava, Slovakia,<br />

b Institute of Analytical Chemistry, Slovak University of Technology,<br />

Radlinskeho 9, 812 37 Bratislava, Slovakia, jan.<br />

mocak@ucm.sk<br />

Introduction<br />

During the last twenty years the chemists have get<br />

accustomed to the use of computers and consequently to the<br />

exploitation of various complex mathematical and statistical<br />

methods, by which they have been trying to explore multivariate<br />

correlations between the output and input variables<br />

more and more in detail. Input variables of different kind are<br />

often continuous and represent usually the results of instrumental<br />

measurements but also other observations, sometime<br />

discrete or categorical, are important for characterizing the<br />

investigated objects. Such kinds of input variables are for<br />

example the results of sensorial assessment of foodstuffs or<br />

simply some qualitative attributes like odour or colour (agreeable/disagreeable<br />

or clear/yellowish/yellow). The output<br />

variables are the targets of the corresponding study, which<br />

again can be represented by some continous variable or a categorical<br />

one with two or more levels. With the increasing complexity<br />

of analytical measurements, and the analysed sample<br />

itself, it becomes clear that all effects that are of interest cannot<br />

be described by a simple univariate relation but they are<br />

multivariate and often they are not linear. A set of methods,<br />

which allow study of multivariate and non-linear correlations<br />

that have recently found very intensive use among chemists<br />

are the artificial neural networks (Anns for short) 1 .<br />

The Anns are difficult to describe using a simple definition.<br />

Perhaps the closest description would be a comparison<br />

with a black box having multiple input and multiple output<br />

which operates using a large number of mostly parallel connected<br />

simple arithmetic units. The most important thing to<br />

characterize about all Ann methods is that they work best<br />

if they are dealing with non-linear dependence between<br />

the inputs and outputs. They have been applied for various<br />

purposes, e.g. optimisation 2,3 quantification of unresolved<br />

peaks 4,5 , estimation of peak parameters, estimation of model<br />

parameters in the equilibria studies, etc. Pattern recognition<br />

and sample classification is also an important application area<br />

for the Ann 6–8 , which is important and fully applicable in<br />

food chemistry.<br />

The most widespread application areas of implementing<br />

the Anns for solution of foodstuff problems are: (i) wine<br />

characterization and authentification, (ii) edible oil characterization<br />

and authentification, (iii) classification of dairy products<br />

and cheese, (iv) classification of soft drinks and fruit<br />

products, (v) classification of strong drinks. In this paper two<br />

examples are given, which exemplify the application of arti-<br />

s556<br />

ficial neural networks for authentification of varietal wines<br />

and olive oils.<br />

Theory<br />

The theory of the Ann is well described in monographs<br />

9–13 and scientific literature. Therefore only a short<br />

description of the principles needed for understanding the<br />

Ann application will be given here. The use of the Ann for<br />

data and knowledge processing can be characterized by analogy<br />

with biological neurons. The artificial neural network<br />

itself consists of neurons connected into networks. The neurons<br />

are sorted in an input layer, one or more hidden layer(s)<br />

and an output layer. The input neurons accept the input data<br />

characteristic for each observation, the output neurons provide<br />

predicted value or pattern of the studied objects, and<br />

the hidden neurons are interconnected with the neurons of<br />

two adjacent layers but neither receive inputs directly nor<br />

provide the output values directly. In most cases, the Ann<br />

architecture consists of the input layer and two active layers<br />

– one hidden and one output layer. The neurons of any two<br />

adjacent layers are mutually connected and the importance of<br />

each connection is expressed by weights.<br />

The role of the Ann is to transform the input information<br />

into the output one. During the training process the weights<br />

are gradually corrected so as to produce the output values as<br />

close as possible to the desired (or target) values, which are<br />

known for all objects included into training set. The training<br />

procedure requires a pair of vectors, x and d, which together<br />

create a training set. The vector x is the actual input into the<br />

network, and the corresponding target - the desired pre-specified<br />

answer, is the vector d. The propagation of the signal through<br />

the network is determined by the connections between<br />

the neurons and by their associated weights, so these weights<br />

represent the synaptic strengths of the biological neurons. The<br />

goal of the training step is to correct the weights w ij so that<br />

they will give a correct output vector y for the vector x from<br />

the training set. In other words, the output vector y should be<br />

as close as possible to the vector d. After the training process<br />

has been completed successfully, it is hoped that the network,<br />

functioning as a black box, will give correct predictions for<br />

any new object x n , which is not included in the training set.<br />

The hidden x i and the output y i neuron activities are defined<br />

by the relations:<br />

where j = 1, …, p concern neurons x j in the previous<br />

layer which precede the given neuron i. ξ i is the net signal<br />

– the sum of the weighted inputs from the previous layer, υ i is<br />

the bias (offset), w ij is the weight and, finally, t(ξ i ) is transfer<br />

function, expressed in various ways, usually as the threshold<br />

(1)<br />

(2)<br />

(3)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!