27.12.2013 Views

The influence of the place-value structure of the Arabic number ...

The influence of the place-value structure of the Arabic number ...

The influence of the place-value structure of the Arabic number ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

In <strong>the</strong> holistic model, two input arrays, each reflecting a holistic representation <strong>of</strong> <strong>the</strong><br />

<strong>number</strong>s ranging from 11 to 99, are implemented. <strong>The</strong>se input arrays are connected to two<br />

output nodes, where<strong>of</strong> <strong>the</strong> left one represents <strong>the</strong> “first <strong>number</strong> larger“ decision and <strong>the</strong> right<br />

one <strong>the</strong> “second <strong>number</strong> larger” decision.<br />

Training <strong>of</strong> connection weights<br />

<strong>The</strong> initial connection weights were pseudo-random <strong>value</strong>s ranging from 0 to 1 taken<br />

from a uniform distribution. Basically, <strong>the</strong> same learning algorithm was used for all three<br />

models. Never<strong>the</strong>less, as <strong>the</strong> fully decomposed and <strong>the</strong> hybrid model involved a hidden layer<br />

learning followed a back-propagation approach in <strong>the</strong>se two models. For <strong>the</strong> holistic model<br />

which did not incorporate a hidden layer <strong>the</strong> learning algorithm used reflected delta rule<br />

guided learning, respectively. Please note that back-propagation is a generalization <strong>of</strong> <strong>the</strong><br />

delta rule approach to networks involving at least one hidden layer (cf. Rumelhart, Hinton, &<br />

Williams, 1986). Based on this and <strong>the</strong> fact that <strong>the</strong> same ma<strong>the</strong>matical algorithm was used<br />

for <strong>the</strong> training <strong>of</strong> connection weights in all three models we are confident that differences in<br />

<strong>the</strong> learning algorithm did not determine <strong>the</strong> current results. Finally, <strong>the</strong> learning constant η<br />

was arbitrarily set to 0.1.<br />

<strong>The</strong> training phase comprised 50.000 for all models. In each training circle two<br />

randomly chosen <strong>number</strong>s between 11 and 99 (excluding multiples <strong>of</strong> 10) were presented to<br />

<strong>the</strong> network. <strong>The</strong> frequency <strong>of</strong> occurrence <strong>of</strong> each <strong>number</strong> during <strong>the</strong> training cycles was<br />

determined by its frequency <strong>of</strong> occurrence in daily life as assessed by taking <strong>the</strong> entries<br />

referenced in Google for <strong>the</strong> respective <strong>number</strong>s as an estimate <strong>of</strong> <strong>the</strong>ir every-day occurrence<br />

(see Appendix A; Verguts & Fias, 2008 for a similar approach).<br />

<strong>The</strong> procedure <strong>of</strong> <strong>the</strong> training phase will be described for <strong>the</strong> case <strong>of</strong> <strong>the</strong> decomposed<br />

model: At first, two <strong>number</strong>s between 11 and 99 were randomly chosen with <strong>the</strong> exception <strong>of</strong><br />

tie <strong>number</strong>s (e.g., 44) and multiples <strong>of</strong> 10 as empirical studies showed anomalous processing<br />

248

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!