08.02.2013 Views

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

6.1. ANALYSIS OF RESULTS<br />

different objects (false negative), leading to reduced invariance.<br />

The graph in Figure 5.16 indicates that the S2 RF size also affects recognition performance but<br />

has different effects for each <strong>of</strong> the distorted test sets. The occluded test set works best with the<br />

smallest S2 RF size, probably because it better captures ihe non-occluded parts <strong>of</strong> the object,<br />

whereas the bigger RF sizes tend to include more occluded sections. Bigger RF sizes show a<br />

slight advantage when recognizing scaled objects, as the difference in size is less accentuated<br />

within large RFs, while it may lead to radically different smaller-sized features. Overall, it is<br />

clear that the best results are obtained by averaging over the four different S2 RF sizes, as one<br />

size's shortcomings are compensated for by another one's strengths.<br />

Finally, a comparison between different models is shown in Figure 5.17, demonstrating that<br />

the proposed Bayesian network can achieve similar feedforward categorization results to the<br />

original HMAX model. Note that the comparison is only rigorously valid between the HMAX<br />

model and the Bayesian Belief Propagation (BBP) 3-level model, as these have equivalcnl num­<br />

bers <strong>of</strong> layers, nodes and features per layer. The alternative BBP 3-level Yamane version was<br />

specitically modilied. by reducing the pooling region and increasing the number <strong>of</strong> nodes ol" the<br />

lop layers, to improve the categorization <strong>of</strong> the translated lest set. Implementing the same mod­<br />

ifications in the original HMAX miKlel would, presumably, yield belter results itian the BBP<br />

version, in the same way that the original 3-level HMAX version produces better results than<br />

the 3-level BBP model.<br />

The superior results <strong>of</strong> HMAX are, however, not surprising as it was specitically designed<br />

to perform feedforward categorization and employs more exact and sophisticated operations,<br />

namely the max and the Radial Basis function, than the BBP model. In fact, it is remarkable that<br />

the BBP miKlel can achieve comparable categorization results using the local belitfpropagation<br />

operations, namely a weighted product operation for selectivity and a weighted sum operation<br />

for invariance. Crucially, using the same algorithm and structure, the BBP model also achieves<br />

recursive feedback modulation, which has been pinpointed as the major limitation <strong>of</strong> HMAX<br />

(Serre et al. 200.*ia).<br />

With respect to the HTMIike model, as was previously noted, the Numenta Vision Toolkit was<br />

237

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!