08.02.2013 Views

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

4.3. LEARNING<br />

when all afferent SI nodes have a flat A distribution, as in blank regions<strong>of</strong> the image, the parent<br />

CI node will also show a flal distribution.<br />

In summary, the CI layer becomes an iniermediale step ihai converts combinations <strong>of</strong> SI fea­<br />

tures and spatial arrangements into the stales <strong>of</strong> a single CI node. The max operation only oc­<br />

curs during the generation <strong>of</strong> the output A messages lo the S2 layers, which groups these states<br />

via the learned weight matrices. This methixl also provides a way to feed back information from<br />

complex to simple layers, where each complex feature corresponds to a specific arrangement <strong>of</strong><br />

simple features. The method is equivalent to that employed by Hierarchical Temporal Networks<br />

(George and Hawkins 2009). where features in each node are combined into temporal groups<br />

or Markov chains. The meihixl used here, however, preserves the Bayesian network simciure<br />

by implementing the grouping olT'eatures in the weights <strong>of</strong> the CPTs.<br />

4.3.3 C1-S2CPTS<br />

To learn the selectivity weights between layers CI and S2. the minimum distance algorithm is<br />

employed. This algorithm was also used to extract the most common spatial patterns (equiva­<br />

lent to selectivity weights) in the Hierarchical Temporal Memory mixlel (George and Hawkins<br />

2009). In the HMAX model, the selectivity weights, or the prototypes which serve as centres for<br />

the Radial Basis Functions, were extracted at random from the CI maps generated by the train­<br />

ing images. However, in our mtxlel. the minimum distance algorithm provides better results, as<br />

it ensures the extracted prototypes maximize the Euclidean distance between each other. The<br />

algorithm works as follows:<br />

I. All features, potential S2 prototypes Ppoieniiiih ^re extracted by sampling from all the<br />

locations and bands <strong>of</strong> the A (CI) response generated for each <strong>of</strong> the training images, i.e.<br />

A(Cli,_jy - () V h.x.y. The number <strong>of</strong> elements for each prototype is ANsi x AA'M X<br />

(Kci/Kc]group)< i.c- the S2 RF size times the number <strong>of</strong> CI states divided by the states<br />

per group. To learn the S2 prototypes, it is more eflicient to obtain a single value for<br />

each C1 group by summing over all the features belonging to that group, [n other words,<br />

although each CI node is composed <strong>of</strong> 40 stales, only 4 values, corresponding to the sum<br />

<strong>of</strong> each group, are used lo compute the S2 prototypes.<br />

163

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!