17.01.2013 Views

Chapter 2. Prehension

Chapter 2. Prehension

Chapter 2. Prehension

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Appendix C - Computational Neural Modelling 397<br />

O=WI (16)<br />

Now, if there are p input/output pairs, we can define two n x p<br />

matrices, 1. and 0, where the columns of 1. are the p input patterns and<br />

the columns of 0 are the corresponding p output patterns. Then we<br />

have<br />

for correct performance. But this equation also defines W and<br />

answers our first question. By inverting it we obtain:<br />

which gives an explicit formula for computing W (although it is<br />

necessary to perform the time consuming task of inverting I.) It also<br />

provides an upper limit on p (the memory’s capacity), answering the<br />

second question: I is only guaranteed to be invertible if pln. For the<br />

remainder of the discussion we will assume p=n, that is the matrices<br />

are square. The input patterns must then be linearly independent,<br />

since I must be invertible (its determinant must be non-zero).<br />

Computing W can be time consuming, as mentioned above, but a<br />

special case occurs when all the input patterns are orthonormal. In this<br />

case, L-l =LT, and the computation of W simply requires a single<br />

matrix multiplication. The orthonormal condition is often assumed<br />

true when the patterns are binary, long (large n), and sparse (mostly<br />

zero), because they tend to be orthogonal (their dot products are likely<br />

to be zero). If they all have about the same number of one’s, then they<br />

can all be scaled down by that number to be roughly normal (of unit<br />

magnitude). The resulting performance of the memory is close to<br />

correct. For non-orthonormal situations, there is the option of<br />

applying a gradient descent technique (e.g. the delta rule) to iteratively<br />

learn the weight matrix.<br />

C.4 Processing Example<br />

As was shown in the previous section, a heteroassociative memory<br />

allows two patterns to be associated with one another, such that one<br />

pattern (the input pattern) triggers the other pattern to appear at the<br />

output units. For this example, a network with three inputs and three<br />

outputs (all linear) will be used, as seen in Figure C.9. Assume your

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!