22.12.2012 Views

3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures

3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures

3. FOOD ChEMISTRy & bIOTEChNOLOGy 3.1. Lectures

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Chem. Listy, 102, s265–s1311 (2008) Food Chemistry & Biotechnology<br />

L08 ARTIFICIAL NEuRAL NETwORKS IN <strong>FOOD</strong><br />

ANALySIS<br />

Ján MOCáK a,b , VIERA MRáZOVá a , DáŠA<br />

KRUŽLICOVá b and FILIP KRAIC a<br />

a Department of Chemistry, University of Ss. Cyril & Methodius,<br />

Nám. J. Herdu 2, 917 01 Trnava, Slovakia,<br />

b Institute of Analytical Chemistry, Slovak University of Technology,<br />

Radlinskeho 9, 812 37 Bratislava, Slovakia, jan.<br />

mocak@ucm.sk<br />

Introduction<br />

During the last twenty years the chemists have get<br />

accustomed to the use of computers and consequently to the<br />

exploitation of various complex mathematical and statistical<br />

methods, by which they have been trying to explore multivariate<br />

correlations between the output and input variables<br />

more and more in detail. Input variables of different kind are<br />

often continuous and represent usually the results of instrumental<br />

measurements but also other observations, sometime<br />

discrete or categorical, are important for characterizing the<br />

investigated objects. Such kinds of input variables are for<br />

example the results of sensorial assessment of foodstuffs or<br />

simply some qualitative attributes like odour or colour (agreeable/disagreeable<br />

or clear/yellowish/yellow). The output<br />

variables are the targets of the corresponding study, which<br />

again can be represented by some continous variable or a categorical<br />

one with two or more levels. With the increasing complexity<br />

of analytical measurements, and the analysed sample<br />

itself, it becomes clear that all effects that are of interest cannot<br />

be described by a simple univariate relation but they are<br />

multivariate and often they are not linear. A set of methods,<br />

which allow study of multivariate and non-linear correlations<br />

that have recently found very intensive use among chemists<br />

are the artificial neural networks (Anns for short) 1 .<br />

The Anns are difficult to describe using a simple definition.<br />

Perhaps the closest description would be a comparison<br />

with a black box having multiple input and multiple output<br />

which operates using a large number of mostly parallel connected<br />

simple arithmetic units. The most important thing to<br />

characterize about all Ann methods is that they work best<br />

if they are dealing with non-linear dependence between<br />

the inputs and outputs. They have been applied for various<br />

purposes, e.g. optimisation 2,3 quantification of unresolved<br />

peaks 4,5 , estimation of peak parameters, estimation of model<br />

parameters in the equilibria studies, etc. Pattern recognition<br />

and sample classification is also an important application area<br />

for the Ann 6–8 , which is important and fully applicable in<br />

food chemistry.<br />

The most widespread application areas of implementing<br />

the Anns for solution of foodstuff problems are: (i) wine<br />

characterization and authentification, (ii) edible oil characterization<br />

and authentification, (iii) classification of dairy products<br />

and cheese, (iv) classification of soft drinks and fruit<br />

products, (v) classification of strong drinks. In this paper two<br />

examples are given, which exemplify the application of arti-<br />

s556<br />

ficial neural networks for authentification of varietal wines<br />

and olive oils.<br />

Theory<br />

The theory of the Ann is well described in monographs<br />

9–13 and scientific literature. Therefore only a short<br />

description of the principles needed for understanding the<br />

Ann application will be given here. The use of the Ann for<br />

data and knowledge processing can be characterized by analogy<br />

with biological neurons. The artificial neural network<br />

itself consists of neurons connected into networks. The neurons<br />

are sorted in an input layer, one or more hidden layer(s)<br />

and an output layer. The input neurons accept the input data<br />

characteristic for each observation, the output neurons provide<br />

predicted value or pattern of the studied objects, and<br />

the hidden neurons are interconnected with the neurons of<br />

two adjacent layers but neither receive inputs directly nor<br />

provide the output values directly. In most cases, the Ann<br />

architecture consists of the input layer and two active layers<br />

– one hidden and one output layer. The neurons of any two<br />

adjacent layers are mutually connected and the importance of<br />

each connection is expressed by weights.<br />

The role of the Ann is to transform the input information<br />

into the output one. During the training process the weights<br />

are gradually corrected so as to produce the output values as<br />

close as possible to the desired (or target) values, which are<br />

known for all objects included into training set. The training<br />

procedure requires a pair of vectors, x and d, which together<br />

create a training set. The vector x is the actual input into the<br />

network, and the corresponding target - the desired pre-specified<br />

answer, is the vector d. The propagation of the signal through<br />

the network is determined by the connections between<br />

the neurons and by their associated weights, so these weights<br />

represent the synaptic strengths of the biological neurons. The<br />

goal of the training step is to correct the weights w ij so that<br />

they will give a correct output vector y for the vector x from<br />

the training set. In other words, the output vector y should be<br />

as close as possible to the vector d. After the training process<br />

has been completed successfully, it is hoped that the network,<br />

functioning as a black box, will give correct predictions for<br />

any new object x n , which is not included in the training set.<br />

The hidden x i and the output y i neuron activities are defined<br />

by the relations:<br />

where j = 1, …, p concern neurons x j in the previous<br />

layer which precede the given neuron i. ξ i is the net signal<br />

– the sum of the weighted inputs from the previous layer, υ i is<br />

the bias (offset), w ij is the weight and, finally, t(ξ i ) is transfer<br />

function, expressed in various ways, usually as the threshold<br />

(1)<br />

(2)<br />

(3)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!