17.01.2013 Views

Chapter 2. Prehension

Chapter 2. Prehension

Chapter 2. Prehension

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Appendix C - Computational Neural Modelling 391<br />

Another common rule which uses the presynaptic signal and<br />

postsynaptic error is called the delta rule4 (or Widrow-Hoff rule), and<br />

it uses a target value as follows (right side of Figure C.5):<br />

Instead of using the output of neuron i to determine by how much to<br />

change the weights, the actual output Oi is first subtracted from the<br />

desired output ti. To do this, supervised learning is used. A small<br />

training set is provided to the network for learning some part of the<br />

task space. The training set is a collection of input/output pairs (or<br />

patterns) that identifies to the network what the output should be for a<br />

given set of inputs. For linear units, this rule minimizes the squares of<br />

the differences between the actual and desired output values summed<br />

over the output neuons and all pairs of input/output vectors<br />

(Rumelhart, Hinton & Williams, 1986a).<br />

The delta rule can be derived from a technique called gradient<br />

descent. ‘Gradient descent’ means ‘go down hill’ in a mathematical<br />

sense. For a traveller in a hilly area, define z to be the person’s<br />

altitude, x to be longitude (i.e., how far east she is located) and y to be<br />

latitude (how far north she is). Then z is a function of x and y: the<br />

altitude of the ground varies with its x-y location. The variables x and<br />

y are independent ones and the traveler can shift these at will by<br />

walking. Since the traveler’s altitude changes as x and y do, z is the<br />

dependent variable. The gradient, or change, of z is the vector of<br />

partial derivatives of z with respect to its independent variables:<br />

(az/ax, az/ay). In this case the gradient represents a horizontal vector<br />

which points in the direction of travel for which z increases most<br />

rapidly. The gradient descent rule simply says “move in the direction<br />

opposite to which the gradient vector points”. Mathematically, it is<br />

written as:<br />

4An early model was a percepuon, which consisted of linear threshold units using<br />

the delta rule for synaptic weight modification mosenblatt, 1962).

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!