Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
Chapter 1 Prelim<strong>in</strong>aries<br />
rial, <strong>in</strong> Levy’s approach the structural probabilities <strong>of</strong> the grammar are perfectly known<br />
to the parser. Consequently, the parallel probabilistic resource allocation theory by Levy<br />
constitutes as sort <strong>of</strong> competence model with surprisal or relative entropy as a “bottleneck”<br />
to comprehension, thus yield<strong>in</strong>g performance-related predictions. Predictions <strong>of</strong><br />
frequency-<strong>based</strong> approaches like the tun<strong>in</strong>g hypothesis (Mitchell, Cuetos, Corley, and<br />
Brysbaert, 1995) are quite similar to surprisal most <strong>of</strong> the time but differ fundamentally<br />
<strong>in</strong> head-f<strong>in</strong>al structures. Similarly DLT and surprisal make comparable predictions only<br />
<strong>in</strong> structures that are not head-f<strong>in</strong>al. In head-f<strong>in</strong>al constructions the preced<strong>in</strong>g dependents<br />
provide statistical <strong>in</strong>formation about the nature <strong>of</strong> the head, thus narrow<strong>in</strong>g the<br />
prediction. Follow<strong>in</strong>g the theory, a better prediction (or lower surprisal) facilitates <strong>in</strong>tegration<br />
on the head. Thus an expectation-<strong>based</strong> theory predicts language-<strong>in</strong>dependent<br />
anti-locality effects <strong>in</strong> head-f<strong>in</strong>al structures.<br />
1.3.3 Canonicity<br />
In literature the term canonicity with respect to word order is <strong>of</strong>ten used as synonymous<br />
to regularity and structural frequency. Here these terms shall be dist<strong>in</strong>guished <strong>in</strong> order<br />
to clearly formulate respective theories.<br />
A theory <strong>of</strong> canonicity has to answer two questions:<br />
1. What categorial doma<strong>in</strong> is the focus <strong>of</strong> the canonicity?<br />
2. What makes specific structures canonical?<br />
The categorial focus <strong>of</strong> canonicity can be grammatical functions, thematic roles, letter<br />
sequences, prosody and the like. The specific structures count<strong>in</strong>g as canonical <strong>in</strong> these<br />
doma<strong>in</strong>s can be chosen by structural regularity, complexity, or simply by convention.<br />
The most common canonicity account goes back to Greenberg (1963) and relates to<br />
the basic grammatical functions subject, object, and predicate and is justified by structural<br />
regularities. Greenberg classified languages <strong>in</strong> terms <strong>of</strong> their canonical word order.<br />
He and subsequent literature count English as a subject-verb-object (SVO) language because<br />
simple sentences and most subord<strong>in</strong>ate constructions follow that order. Therewith<br />
it belongs to the second biggest class (41.79%) preceded by the SOV order attributed<br />
to 44.78% (accord<strong>in</strong>g to Toml<strong>in</strong>, 1986) <strong>of</strong> the languages <strong>of</strong> the world. However, the<br />
classification is not as clear for all languages. German is arguably an SOV language<br />
although the simplest sentence structure is built with an SVO order like English. For<br />
example, Erdmann (1990) concludes that German does not fulfill all requirements for<br />
an SOV language and should therefore be categorized as SVO.<br />
As mentioned above, structural regularity <strong>based</strong> on corpus occurences is not the only<br />
possibility to ground a canonicity account on. A generative grammar-<strong>based</strong> account<br />
that relates word order canonicity to language process<strong>in</strong>g assumes the language-specific<br />
canonical structure as an <strong>in</strong>ternal representation underly<strong>in</strong>g the surface structure (L<strong>in</strong><br />
et al., 2005). Thus, <strong>in</strong> order to comprehend a non-canonically structured sentence the<br />
12