08.02.2013 Views

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

Bernal S D_2010.pdf - University of Plymouth

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

2.1. OBJECT RECnCNmON<br />

unit will be equivalent to the response <strong>of</strong> the afferent simple unit with the highest value. If any<br />

<strong>of</strong> the simple units within the complex unit's spatial pmtling range is activated, then the complex<br />

unit will also emit an equivalent response. This means complex units achieve a certain degree<br />

<strong>of</strong> invariance to spatial translation and scale.<br />

Selectivity is generated by a template-matching operation over a set <strong>of</strong> afferents tuned to differ­<br />

ent features, implemented as a Radial Basis function network (Bishop 1995). First, a dictionary<br />

<strong>of</strong> features or prototypes is learned. Each prototype represents a specific response configura­<br />

tion <strong>of</strong> the afferent complex units from the level below, feeding into the simple unit in the level<br />

above. Each simple unit is then luned to a specific feature <strong>of</strong> the dictionary, eliciting the max­<br />

imum response when the input stimuli in the spatial region covered by the unit matches the<br />

learned feature. The response is determined by a Gaussian tuning function which provides a<br />

similarity measure between the input and the prototype. The malhematical formulation for both<br />

the selectivity and invariance operations is described in Section 4.4.<br />

With respect to the implementation <strong>of</strong> the tup level, in the first proposed model (Riesenhubcr and<br />

Poggio 1999) this was described as a set <strong>of</strong> view-tuned units connected to the output <strong>of</strong> the C2<br />

layer The weights were set so that the center <strong>of</strong> the Ciaussian ass(K;iated with each view-tuned<br />

unit corresponded to a specific view <strong>of</strong> an input image. More recent versions have employed<br />

C2 features as the input to a linear support vector machine (Serre el al. 2005b, 2007c). or have<br />

implemented an additional unsupervised S,1/C3 level analogous to the intermediate level (Serre<br />

et al. 2005a), In one particular implementation the model was extended to include an additional<br />

supervised S4 level trained for a categorization task, possibly corresponding to categorization<br />

finiis in prefrontal cortex (Serre el al. 2007b,a), A further extension, consisting <strong>of</strong> two extra<br />

sublevels S2b and C2b, has enabled some <strong>of</strong> the models to account for bypass routes, such as<br />

direct projections from V2 to IT which bypass V4 (Serre et al. 2005a. 2007b.a).<br />

Learning in the model lakes place at the top level in a supervised way, while at the intermediate<br />

levels the feature prototypes are learned in an unsupervised manner. The model implements<br />

developmenlal-like learning, such that units store the synaptic weights <strong>of</strong> the current pattern <strong>of</strong><br />

activity from its afferent inputs, in response to the part <strong>of</strong> image that falls within its receptive<br />

IH

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!