29.08.2013 Views

Connectionist Modeling of Experience-based Effects in Sentence ...

Connectionist Modeling of Experience-based Effects in Sentence ...

Connectionist Modeling of Experience-based Effects in Sentence ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

GPE<br />

0.0 0.2 0.4 0.6 0.8 1.0<br />

English SRC<br />

rep. that attacked the senator praised the judge<br />

Region<br />

epoch 1<br />

epoch 2<br />

epoch 3<br />

Chapter 4 Two SRN Prediction Studies<br />

GPE<br />

0.0 0.2 0.4 0.6 0.8 1.0<br />

English ORC<br />

rep. that the sen. attacked praised the judge<br />

Region<br />

Figure 4.1: Replication <strong>of</strong> MacDonald and Christiansen (2002)<br />

4.3 RC Extraction <strong>in</strong> Mandar<strong>in</strong><br />

4.3.1 Simulation 1: Regularity<br />

Model Parameters<br />

The first simulation should assess the degree <strong>of</strong> the regularity advantage the ORC receives<br />

due to its canonical word order. A regularity effect is assessable only when the frequencies<br />

<strong>of</strong> both RC types <strong>in</strong> the corpora are identical. Therefore the SRC and the ORC received<br />

the same probability <strong>in</strong> the generation grammar. Although the replications were done<br />

with an RC probability <strong>of</strong> 0.1, I used the orig<strong>in</strong>al value <strong>of</strong> 0.05, reported <strong>in</strong> MacDonald<br />

and Christiansen (2002), for the Mandar<strong>in</strong> regularity simulation. Compared to English<br />

the Ch<strong>in</strong>ese grammar used here is very simple. Sett<strong>in</strong>g the RC probability too high<br />

could speed up <strong>in</strong> the learn<strong>in</strong>g process <strong>in</strong> a way that conceals tra<strong>in</strong><strong>in</strong>g effects.<br />

The Grammar The grammar used to generate the corpora covered simple regular<br />

Mandar<strong>in</strong> SVO sentences as well as SR and OR clauses. Relative clause attachment<br />

could happen at every noun with a probability <strong>of</strong> 0.05. The embedd<strong>in</strong>g depth was<br />

theoretically unlimited but with the small attachment probability <strong>of</strong> 0.05 the longest<br />

sentence <strong>in</strong> the corpora had a length <strong>of</strong> 16 words. The 17-word lexicon consisted <strong>of</strong> 9<br />

plural and s<strong>in</strong>gular nouns, three transitive and four <strong>in</strong>transitive verbs, <strong>of</strong> which one (lijie<br />

“understand”) belongs to both categories, the relativizer de and the EOS. Note however,<br />

that there is no number agreement between nouns and verbs <strong>in</strong> Mandar<strong>in</strong>. The full<br />

lexicon is given <strong>in</strong> the Appendix. Note further that <strong>in</strong> normal Mandar<strong>in</strong> <strong>in</strong>transitive<br />

66<br />

epoch 1<br />

epoch 2<br />

epoch 3

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!