3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures
3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures
3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Chem. Listy, 102, s265–s1311 (2008) Food Chemistry & Biotechnology<br />
L08 ARTIFICIAL NEuRAL NETwORKS IN <strong>FOOD</strong><br />
ANALySIS<br />
Ján MOCáK a,b , VIERA MRáZOVá a , DáŠA<br />
KRUŽLICOVá b and FILIP KRAIC a<br />
a Department of Chemistry, University of Ss. Cyril & Methodius,<br />
Nám. J. Herdu 2, 917 01 Trnava, Slovakia,<br />
b Institute of Analytical Chemistry, Slovak University of Technology,<br />
Radlinskeho 9, 812 37 Bratislava, Slovakia, jan.<br />
mocak@ucm.sk<br />
Introduction<br />
During the last twenty years the chemists have get<br />
accustomed to the use of computers and consequently to the<br />
exploitation of various complex mathematical and statistical<br />
methods, by which they have been trying to explore multivariate<br />
correlations between the output and input variables<br />
more and more in detail. Input variables of different kind are<br />
often continuous and represent usually the results of instrumental<br />
measurements but also other observations, sometime<br />
discrete or categorical, are important for characterizing the<br />
investigated objects. Such kinds of input variables are for<br />
example the results of sensorial assessment of foodstuffs or<br />
simply some qualitative attributes like odour or colour (agreeable/disagreeable<br />
or clear/yellowish/yellow). The output<br />
variables are the targets of the corresponding study, which<br />
again can be represented by some continous variable or a categorical<br />
one with two or more levels. With the increasing complexity<br />
of analytical measurements, and the analysed sample<br />
itself, it becomes clear that all effects that are of interest cannot<br />
be described by a simple univariate relation but they are<br />
multivariate and often they are not linear. A set of methods,<br />
which allow study of multivariate and non-linear correlations<br />
that have recently found very intensive use among chemists<br />
are the artificial neural networks (Anns for short) 1 .<br />
The Anns are difficult to describe using a simple definition.<br />
Perhaps the closest description would be a comparison<br />
with a black box having multiple input and multiple output<br />
which operates using a large number of mostly parallel connected<br />
simple arithmetic units. The most important thing to<br />
characterize about all Ann methods is that they work best<br />
if they are dealing with non-linear dependence between<br />
the inputs and outputs. They have been applied for various<br />
purposes, e.g. optimisation 2,3 quantification of unresolved<br />
peaks 4,5 , estimation of peak parameters, estimation of model<br />
parameters in the equilibria studies, etc. Pattern recognition<br />
and sample classification is also an important application area<br />
for the Ann 6–8 , which is important and fully applicable in<br />
food chemistry.<br />
The most widespread application areas of implementing<br />
the Anns for solution of foodstuff problems are: (i) wine<br />
characterization and authentification, (ii) edible oil characterization<br />
and authentification, (iii) classification of dairy products<br />
and cheese, (iv) classification of soft drinks and fruit<br />
products, (v) classification of strong drinks. In this paper two<br />
examples are given, which exemplify the application of arti-<br />
s556<br />
ficial neural networks for authentification of varietal wines<br />
and olive oils.<br />
Theory<br />
The theory of the Ann is well described in monographs<br />
9–13 and scientific literature. Therefore only a short<br />
description of the principles needed for understanding the<br />
Ann application will be given here. The use of the Ann for<br />
data and knowledge processing can be characterized by analogy<br />
with biological neurons. The artificial neural network<br />
itself consists of neurons connected into networks. The neurons<br />
are sorted in an input layer, one or more hidden layer(s)<br />
and an output layer. The input neurons accept the input data<br />
characteristic for each observation, the output neurons provide<br />
predicted value or pattern of the studied objects, and<br />
the hidden neurons are interconnected with the neurons of<br />
two adjacent layers but neither receive inputs directly nor<br />
provide the output values directly. In most cases, the Ann<br />
architecture consists of the input layer and two active layers<br />
– one hidden and one output layer. The neurons of any two<br />
adjacent layers are mutually connected and the importance of<br />
each connection is expressed by weights.<br />
The role of the Ann is to transform the input information<br />
into the output one. During the training process the weights<br />
are gradually corrected so as to produce the output values as<br />
close as possible to the desired (or target) values, which are<br />
known for all objects included into training set. The training<br />
procedure requires a pair of vectors, x and d, which together<br />
create a training set. The vector x is the actual input into the<br />
network, and the corresponding target - the desired pre-specified<br />
answer, is the vector d. The propagation of the signal through<br />
the network is determined by the connections between<br />
the neurons and by their associated weights, so these weights<br />
represent the synaptic strengths of the biological neurons. The<br />
goal of the training step is to correct the weights w ij so that<br />
they will give a correct output vector y for the vector x from<br />
the training set. In other words, the output vector y should be<br />
as close as possible to the vector d. After the training process<br />
has been completed successfully, it is hoped that the network,<br />
functioning as a black box, will give correct predictions for<br />
any new object x n , which is not included in the training set.<br />
The hidden x i and the output y i neuron activities are defined<br />
by the relations:<br />
where j = 1, …, p concern neurons x j in the previous<br />
layer which precede the given neuron i. ξ i is the net signal<br />
– the sum of the weighted inputs from the previous layer, υ i is<br />
the bias (offset), w ij is the weight and, finally, t(ξ i ) is transfer<br />
function, expressed in various ways, usually as the threshold<br />
(1)<br />
(2)<br />
(3)