08.02.2013 Views

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

4.3. LEARNING<br />

4.3 Learning<br />

This section describes how to learn the conditional probability tables (CPTs) <strong>of</strong> each <strong>of</strong> the<br />

nodes in the Bayesian network in order to approximate the selectivity and invariance operations<br />

<strong>of</strong> the HMAX model. For this learning stage an important assumption is made in order to<br />

simplify the process. The network is assumed have a single parent per node (tree stracture<br />

with no loops) so that the feedforward A messages are not affected by the top-down feedback n<br />

messages. The bottom-up A message from a node with a single parent does not include evidence<br />

from that parent (see Section 3.3.3 for details).<br />

The reason for making this assumption is that the CPTs in the network are learned in an unsu­<br />

pervised manner, starling from the bottom layer (following HMAX learning methcxls), based<br />

on the response obtained at each layer. In order to calculate the response <strong>of</strong> nodes with multiple<br />

parents, the messages from all parents need lo be combined using the CPTs thai relate the node<br />

lo its parents. However, these CPTs would still be unknown. This implie-s that, theoretically.<br />

in this type <strong>of</strong> network, all the CPTs would need to be learned at the same time. By assuming<br />

nodes with a single parent, the A messages, based solely on bottom-up evidence, can be used<br />

as a reference to learn the appropriate weights layer by layer. Similar assumptions are made in<br />

other related models (Hpshtein et al. 2008, George and Hawkins 2009. Hinton ct al. 2006). The<br />

learning process is now described one layer at a lime.<br />

4.3.1 Image-Si weights<br />

The input image is pre-processed with a battery <strong>of</strong> Gabor filters described by Equation (4.1)<br />

with the parameter range described in Table 4.1, i.e. at 4 different orientations and 8 sizes. Each<br />

<strong>of</strong> the lilters is applied at every location <strong>of</strong> the image. The filtered responses, normalized over<br />

the four orientations at each location and scale, are used as the output A messages <strong>of</strong> a set <strong>of</strong><br />

dummy nodes that feed onto the SI nodes. As explained in Section 3.3.3. dummy nodes do not<br />

encode a variable or have a belief, but just generate k messages for the parent niRies. For this<br />

reason there is no need lo define the Cl^s between the dummy nodes and the SI nodes.<br />

157

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!