Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
4.5 Conclusion<br />
<strong>of</strong> the VP-predict<strong>in</strong>g NP <strong>in</strong> the representational cycle <strong>of</strong> the SRN decays faster. This <strong>in</strong><br />
turn is compensated by <strong>in</strong>creased tra<strong>in</strong><strong>in</strong>g on center-embedd<strong>in</strong>g compared to the English<br />
simulation, result<strong>in</strong>g <strong>in</strong> comparable error values when tra<strong>in</strong>ed without commas. This<br />
is, <strong>of</strong> course, an ad-hoc hypothesis and needs further <strong>in</strong>vestigation, which is beyond the<br />
scope <strong>of</strong> this thesis.<br />
4.5 Conclusion<br />
This thesis <strong>in</strong>vestigated the explanatory power <strong>of</strong> a certa<strong>in</strong> implementation <strong>of</strong> the experience<br />
account. The well approved SRN model<strong>in</strong>g approach <strong>of</strong> MacDonald and Christiansen<br />
(2002) was adopted to test its predictions on two phenomena currently discussed<br />
<strong>in</strong> literature. The RC extraction type preference <strong>in</strong> Mandar<strong>in</strong> and the forgett<strong>in</strong>g effect<br />
<strong>in</strong> complex center-embedd<strong>in</strong>g was discussed and then modeled. At first, the two problems<br />
were approached theoretically, review<strong>in</strong>g results <strong>of</strong> empirical studies and discuss<strong>in</strong>g<br />
potential predictions <strong>of</strong> available theories. Concern<strong>in</strong>g the Mandar<strong>in</strong> relative clauses,<br />
the studies showed exceptionally mixed results. However, an observed object advantage<br />
appeared always on the RC region, whereas a subject advantage was found only on the<br />
relativizer/head noun region. That fact and the experiment by Qiao and Forster (2008)<br />
are suggest<strong>in</strong>g that Ch<strong>in</strong>ese Mandar<strong>in</strong> might have to be counted as an exception to a<br />
universal subject preference. On the other hand, the results for the forgett<strong>in</strong>g effect were<br />
very clear and best expla<strong>in</strong>ed by language-specific experience.<br />
In chapter 3 the simple recurrent network was <strong>in</strong>troduced and its properties were<br />
discussed. An SRN is a very simple and doma<strong>in</strong>-unspecific model, but it accounts for<br />
the three necessities <strong>in</strong>troduced <strong>in</strong> the beg<strong>in</strong>n<strong>in</strong>g <strong>of</strong> this thesis: a) biological factors<br />
(architectural limits), b) cont<strong>in</strong>uous gradedness <strong>in</strong> performance, and c) experience.<br />
In chapter 4 the experience theory predictions regard<strong>in</strong>g the two sample problems were<br />
practically assessed. Just like <strong>in</strong> the discussion and critique <strong>of</strong> MC02, the simulation<br />
results presented here looked promis<strong>in</strong>g on first sight; but sub-experiments and detailed<br />
data analysis revealed considerable <strong>in</strong>consistencies with respect to human data. The<br />
Mandar<strong>in</strong> RC simulation predicted an object preference, but the location <strong>of</strong> the effect<br />
was not consistent with human data. In addition, the second simulation demonstrated<br />
that the regularity effect was very weak. It becomes clear that the tra<strong>in</strong><strong>in</strong>g material used<br />
must be carefully chosen <strong>in</strong> order to guarantee comparability with other simulations and<br />
empirical studies. The forgett<strong>in</strong>g effect was predicted to be present <strong>in</strong> English but not <strong>in</strong><br />
German, consistent with human data. However, further simulations revealed the comma<br />
<strong>in</strong>sertion as the most important factor.<br />
Of course, it has to be clear that the simple network tra<strong>in</strong>ed on a simple grammar<br />
would not learn the same constra<strong>in</strong>ts as humans do. These simulations are rather approximations<br />
po<strong>in</strong>t<strong>in</strong>g to a certa<strong>in</strong> direction. A noticeable problem <strong>of</strong> the SRN predictions<br />
is their dependency on local coherence, which can also be described as a low memory<br />
span. This is, however, ma<strong>in</strong>ly dependent upon the specific properties <strong>of</strong> the learn<strong>in</strong>g<br />
81