06.03.2013 Views

Artificial Intelligence and Soft Computing: Behavioral ... - Arteimi.info

Artificial Intelligence and Soft Computing: Behavioral ... - Arteimi.info

Artificial Intelligence and Soft Computing: Behavioral ... - Arteimi.info

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

estimated through a forward pass [21] on the network. After the forward pass<br />

is over, the error vector at the output layer is estimated by taking the<br />

component-wise difference of the target pattern <strong>and</strong> the generated output<br />

vector. A function of errors of the output layered nodes is then propagated<br />

back through the network to each layer for adjustment of weights in that layer.<br />

The weight adaptation policy in back-propagation algorithm is derived<br />

following the principle of steepest descent approach [16] of finding minima<br />

a multi-valued function [5]. A derivation of the algorithm is given in<br />

B.<br />

The most significant issue of a back-propagation algorithm is the<br />

propagation of error through non-linear inhibiting function in backward<br />

direction. Before this was invented, training with a multi-layered feed-forward<br />

neural network was just beyond imagination. In this section, the process of<br />

propagation of error from one layer to its previous layer will be discussed<br />

shortly. Further, how these propagated errors are used for weight adaptation<br />

will also be presented schematically.<br />

Typical neurons employed in back-propagation learning contain two<br />

modules (vide fig. 14.6(a)). The circle containing ∑ wi xi denotes a weighted<br />

sum of the inputs xi for i= 1 to n. The rectangular box in fig. 14.6(a) represents<br />

the sigmoid type non-linearity. It may be added here that the sigmoid has been<br />

chosen here because of the continuity of the function over a wide range. The<br />

continuity of the nonlinear function is required in back-propagation, as we<br />

have to differentiate the function to realize the steepest descent criteria of<br />

learning. Fig. 14.6(b) is a symbolic representation of the neurons used in fig.<br />

14.6(c ).<br />

In fig. 14.6(c ), two layers of neurons have been shown. The left side<br />

layer is the penultimate (k –1)-th layer, whereas the single neuron in the next<br />

k-th layer represents one of the output layered neurons. We denote the top two<br />

neurons at the (k-1)-th <strong>and</strong> k-th layer by neuron p <strong>and</strong> q respectively. The<br />

connecting weight between them is denoted by wp,q,k. For computing<br />

Wp,q,k(n+1), from its value at iteration n, we use the formula presented in<br />

expression (14.2-14.4).<br />

We already mentioned that a function of error is propagated from the<br />

nodes in the output layer to other layers for the adjustment of weight. This<br />

functional form of the back-propagated error is presented in expression (14.4)<br />

<strong>and</strong> illustrated in fig. 14.7. It is seen from expression (14.4) that the<br />

contribution of the errors of each node at the output layer is taken into account<br />

in an exhaustively connected neural net.<br />

For training a network by this algorithm, one has to execute the<br />

following 4 steps in order for all patterns one by one.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!