29.08.2013 Views

Connectionist Modeling of Experience-based Effects in Sentence ...

Connectionist Modeling of Experience-based Effects in Sentence ...

Connectionist Modeling of Experience-based Effects in Sentence ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

4.5 Conclusion<br />

<strong>of</strong> the VP-predict<strong>in</strong>g NP <strong>in</strong> the representational cycle <strong>of</strong> the SRN decays faster. This <strong>in</strong><br />

turn is compensated by <strong>in</strong>creased tra<strong>in</strong><strong>in</strong>g on center-embedd<strong>in</strong>g compared to the English<br />

simulation, result<strong>in</strong>g <strong>in</strong> comparable error values when tra<strong>in</strong>ed without commas. This<br />

is, <strong>of</strong> course, an ad-hoc hypothesis and needs further <strong>in</strong>vestigation, which is beyond the<br />

scope <strong>of</strong> this thesis.<br />

4.5 Conclusion<br />

This thesis <strong>in</strong>vestigated the explanatory power <strong>of</strong> a certa<strong>in</strong> implementation <strong>of</strong> the experience<br />

account. The well approved SRN model<strong>in</strong>g approach <strong>of</strong> MacDonald and Christiansen<br />

(2002) was adopted to test its predictions on two phenomena currently discussed<br />

<strong>in</strong> literature. The RC extraction type preference <strong>in</strong> Mandar<strong>in</strong> and the forgett<strong>in</strong>g effect<br />

<strong>in</strong> complex center-embedd<strong>in</strong>g was discussed and then modeled. At first, the two problems<br />

were approached theoretically, review<strong>in</strong>g results <strong>of</strong> empirical studies and discuss<strong>in</strong>g<br />

potential predictions <strong>of</strong> available theories. Concern<strong>in</strong>g the Mandar<strong>in</strong> relative clauses,<br />

the studies showed exceptionally mixed results. However, an observed object advantage<br />

appeared always on the RC region, whereas a subject advantage was found only on the<br />

relativizer/head noun region. That fact and the experiment by Qiao and Forster (2008)<br />

are suggest<strong>in</strong>g that Ch<strong>in</strong>ese Mandar<strong>in</strong> might have to be counted as an exception to a<br />

universal subject preference. On the other hand, the results for the forgett<strong>in</strong>g effect were<br />

very clear and best expla<strong>in</strong>ed by language-specific experience.<br />

In chapter 3 the simple recurrent network was <strong>in</strong>troduced and its properties were<br />

discussed. An SRN is a very simple and doma<strong>in</strong>-unspecific model, but it accounts for<br />

the three necessities <strong>in</strong>troduced <strong>in</strong> the beg<strong>in</strong>n<strong>in</strong>g <strong>of</strong> this thesis: a) biological factors<br />

(architectural limits), b) cont<strong>in</strong>uous gradedness <strong>in</strong> performance, and c) experience.<br />

In chapter 4 the experience theory predictions regard<strong>in</strong>g the two sample problems were<br />

practically assessed. Just like <strong>in</strong> the discussion and critique <strong>of</strong> MC02, the simulation<br />

results presented here looked promis<strong>in</strong>g on first sight; but sub-experiments and detailed<br />

data analysis revealed considerable <strong>in</strong>consistencies with respect to human data. The<br />

Mandar<strong>in</strong> RC simulation predicted an object preference, but the location <strong>of</strong> the effect<br />

was not consistent with human data. In addition, the second simulation demonstrated<br />

that the regularity effect was very weak. It becomes clear that the tra<strong>in</strong><strong>in</strong>g material used<br />

must be carefully chosen <strong>in</strong> order to guarantee comparability with other simulations and<br />

empirical studies. The forgett<strong>in</strong>g effect was predicted to be present <strong>in</strong> English but not <strong>in</strong><br />

German, consistent with human data. However, further simulations revealed the comma<br />

<strong>in</strong>sertion as the most important factor.<br />

Of course, it has to be clear that the simple network tra<strong>in</strong>ed on a simple grammar<br />

would not learn the same constra<strong>in</strong>ts as humans do. These simulations are rather approximations<br />

po<strong>in</strong>t<strong>in</strong>g to a certa<strong>in</strong> direction. A noticeable problem <strong>of</strong> the SRN predictions<br />

is their dependency on local coherence, which can also be described as a low memory<br />

span. This is, however, ma<strong>in</strong>ly dependent upon the specific properties <strong>of</strong> the learn<strong>in</strong>g<br />

81

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!