Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
Chapter 1 Prelim<strong>in</strong>aries<br />
empirically consistent, as comparisons with studies like K<strong>in</strong>g and Just (1991) show. All<br />
predicted process<strong>in</strong>g difficulties are expla<strong>in</strong>ed by demand <strong>of</strong> activation. This also covers<br />
<strong>in</strong>dividual differences by ascrib<strong>in</strong>g them to different limits <strong>of</strong> the total activation amount.<br />
The effect <strong>of</strong> decay is only <strong>in</strong>directly accounted for, depend<strong>in</strong>g on the number <strong>of</strong> newly<br />
activated elements, which is very similar to the DLT Integration Cost. The capacity<br />
limit causes newly needed activation to be preferably drawn from older elements. This<br />
results <strong>in</strong> a cont<strong>in</strong>uously graded decay, which, however, is not temporally dependent but<br />
rather depends on storage/process<strong>in</strong>g demands just like <strong>in</strong> the DLT.<br />
ACT-R <strong>Sentence</strong> Process<strong>in</strong>g Model<br />
Another computationally implemented sentence process<strong>in</strong>g model (Lewis and Vasishth,<br />
2005; Lewis et al., 2006; Vasishth and Lewis, 2006b) is built on the cognitive architecture<br />
ACT-R (Anderson and Lebiere, 1998; Anderson et al., 2004). Similar to the CAPS<br />
model ACT-R works on a sub-symbolic activation propagation basis. Rule application,<br />
however, happens on a more symbolic-fashioned condition-action relation. Process<strong>in</strong>g<br />
difficulties are predom<strong>in</strong>antly retrieval-<strong>based</strong>. Elements (memory chunks) like lexical<br />
entries <strong>in</strong>volved <strong>in</strong> a production need to be retrieved from declarative memory. The<br />
success <strong>of</strong> the retrieval process depends on the chunk’s current activation level and its<br />
match<strong>in</strong>g <strong>of</strong> the retrieval cues specified <strong>in</strong> the production condition. Retrieval cues<br />
are feature value pairs that <strong>in</strong>crease the activation <strong>of</strong> chunks depend<strong>in</strong>g on the number <strong>of</strong><br />
matched features (associative activation). The total activation <strong>of</strong> a memory chunk<br />
calculated from the activation level and cue-<strong>based</strong> activation determ<strong>in</strong>es the probability<br />
<strong>of</strong> be<strong>in</strong>g retrieved and its retrieval latency. The possibility <strong>of</strong> several chunks match<strong>in</strong>g<br />
the retrieval cues partially enables the model to simulate associative retrieval<br />
<strong>in</strong>terference. Retrieval <strong>in</strong>terference causes distribution <strong>of</strong> associative activation onto<br />
several lexical entries, caus<strong>in</strong>g latencies and potentially the retrieval <strong>of</strong> the wrong chunk.<br />
How severely <strong>in</strong>terference affects retrieval depends on the beforementioned activation<br />
level. Activation is a fluctuat<strong>in</strong>g value which is a function <strong>of</strong> usage and decay over time.<br />
Cue-<strong>based</strong> activation and retrieval <strong>of</strong> a particular element cause its reactivation that<br />
slows down the decay process. The pars<strong>in</strong>g process <strong>of</strong> the ACT-R sentence process<strong>in</strong>g<br />
model (Lewis and Vasishth, 2005) is a comb<strong>in</strong>ation <strong>of</strong> a left corner <strong>in</strong>cremental structure<br />
build<strong>in</strong>g mechanism and a top-down goal-guided syntactic expectation that specifies the<br />
phrasal category <strong>of</strong> the structure to be constructed. A very unconventional assumption<br />
<strong>of</strong> the model regard<strong>in</strong>g the pars<strong>in</strong>g process is that, <strong>in</strong> spite <strong>of</strong> an <strong>in</strong>cremental pars<strong>in</strong>g<br />
process, the memory representation does not conta<strong>in</strong> serial order <strong>in</strong>formation that could<br />
guide retrieval and attachment preferences. Recency is only implicitly accounted for<br />
by the decay function that affects pars<strong>in</strong>g decisions <strong>in</strong> addition to cue-match<strong>in</strong>g. What<br />
differentiates this model from CC-READER and DLT is the account for <strong>in</strong>terference<br />
effects and a temporal decay function. Furthermore, process<strong>in</strong>g difficulty is not represented<br />
by process<strong>in</strong>g cycles but directly by estimated process<strong>in</strong>g time. Sett<strong>in</strong>g retrieval<br />
cues, structural attachment, and shift<strong>in</strong>g attention to the next word have fixed time<br />
10