Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
GPE<br />
0.0 0.2 0.4 0.6 0.8 1.0<br />
English SRC<br />
rep. that attacked the senator praised the judge<br />
Region<br />
epoch 1<br />
epoch 2<br />
epoch 3<br />
Chapter 4 Two SRN Prediction Studies<br />
GPE<br />
0.0 0.2 0.4 0.6 0.8 1.0<br />
English ORC<br />
rep. that the sen. attacked praised the judge<br />
Region<br />
Figure 4.1: Replication <strong>of</strong> MacDonald and Christiansen (2002)<br />
4.3 RC Extraction <strong>in</strong> Mandar<strong>in</strong><br />
4.3.1 Simulation 1: Regularity<br />
Model Parameters<br />
The first simulation should assess the degree <strong>of</strong> the regularity advantage the ORC receives<br />
due to its canonical word order. A regularity effect is assessable only when the frequencies<br />
<strong>of</strong> both RC types <strong>in</strong> the corpora are identical. Therefore the SRC and the ORC received<br />
the same probability <strong>in</strong> the generation grammar. Although the replications were done<br />
with an RC probability <strong>of</strong> 0.1, I used the orig<strong>in</strong>al value <strong>of</strong> 0.05, reported <strong>in</strong> MacDonald<br />
and Christiansen (2002), for the Mandar<strong>in</strong> regularity simulation. Compared to English<br />
the Ch<strong>in</strong>ese grammar used here is very simple. Sett<strong>in</strong>g the RC probability too high<br />
could speed up <strong>in</strong> the learn<strong>in</strong>g process <strong>in</strong> a way that conceals tra<strong>in</strong><strong>in</strong>g effects.<br />
The Grammar The grammar used to generate the corpora covered simple regular<br />
Mandar<strong>in</strong> SVO sentences as well as SR and OR clauses. Relative clause attachment<br />
could happen at every noun with a probability <strong>of</strong> 0.05. The embedd<strong>in</strong>g depth was<br />
theoretically unlimited but with the small attachment probability <strong>of</strong> 0.05 the longest<br />
sentence <strong>in</strong> the corpora had a length <strong>of</strong> 16 words. The 17-word lexicon consisted <strong>of</strong> 9<br />
plural and s<strong>in</strong>gular nouns, three transitive and four <strong>in</strong>transitive verbs, <strong>of</strong> which one (lijie<br />
“understand”) belongs to both categories, the relativizer de and the EOS. Note however,<br />
that there is no number agreement between nouns and verbs <strong>in</strong> Mandar<strong>in</strong>. The full<br />
lexicon is given <strong>in</strong> the Appendix. Note further that <strong>in</strong> normal Mandar<strong>in</strong> <strong>in</strong>transitive<br />
66<br />
epoch 1<br />
epoch 2<br />
epoch 3