Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
Connectionist Modeling of Experience-based Effects in Sentence ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
OUTPUT<br />
PUT PLAN<br />
Chapter 3 <strong>Connectionist</strong> Modell<strong>in</strong>g <strong>of</strong> Language Comprehension<br />
vides much contextual <strong>in</strong>formation help<strong>in</strong>g to <strong>in</strong>terpret current <strong>in</strong>put. The context <strong>of</strong><br />
an utterance has a great <strong>in</strong>fluence on ambiguity resolution and predictions <strong>of</strong> content.<br />
There have been someThere accounts are many <strong>of</strong> provid<strong>in</strong>g ways <strong>in</strong> connectionist which this can networks be with temporal<br />
representation which were accomplished, <strong>of</strong> explicit and nature. a number This posed <strong>of</strong> <strong>in</strong>terest<strong>in</strong>g limits to the number and richness<br />
<strong>of</strong> representations. proposals Elman (1990) have appeared describes<strong>in</strong> athe simple literature way to (e.g. provide a connectionist<br />
network with memory, called Jordan, a simple 1986; recurrent Tank & network Hopfield, (SRN, 1987; figure 3.1). The hidden<br />
representations <strong>in</strong> the network Stornetta, are Hogg, copied & <strong>in</strong>to Huberman, a so-called1987; context layer, which <strong>in</strong>fluences<br />
the hidden representations Watrous <strong>in</strong> the next & Shastri, step through 1987; weighted Waibel, activation feed<strong>in</strong>g. This<br />
memory loop goes without Hanazawa, any explicit H<strong>in</strong>ton, representation Shikano, & <strong>of</strong> Lang, time1987; or relations between <strong>in</strong>put<br />
chunks. It is the iterative P<strong>in</strong>eda, procedure 1988; Williams <strong>of</strong> copy<strong>in</strong>g & Zipser, and back-feed<strong>in</strong>g 1988). One itself that produces<br />
temporal relations on an<strong>of</strong> implicit the most level. promis<strong>in</strong>g Becausewas every suggested copy <strong>of</strong> the by activation pattern has<br />
been <strong>in</strong>fluenced by all earlier Jordan copies, (1986). the Jordan contextual described memory a network reaches <strong>in</strong>to the “past” <strong>in</strong> a<br />
cont<strong>in</strong>uously graded way (shown over several <strong>in</strong> Figure <strong>in</strong>put1) steps. conta<strong>in</strong><strong>in</strong>g The <strong>in</strong>formation recurrent <strong>of</strong> earlier <strong>in</strong>put representations<br />
is still <strong>in</strong> the connections representation which aswere a trace, used but to associate newer <strong>in</strong>put a has more <strong>in</strong>fluential<br />
power. Elman (1990) writes: static pattern (a “Plan”) with a serially<br />
ordered output pattern (a sequence <strong>of</strong><br />
“In this account, “Actions”). memory isThe neither recurrent passive connections nor a separate allowsubsystem.<br />
One<br />
cannot properly speak the network’s <strong>of</strong> a memory hidden forunits sequences; to see that its own memory is <strong>in</strong>extricably<br />
bound up withprevious the rest output, <strong>of</strong> the process<strong>in</strong>g so that the mechanism.” subsequent<br />
behavior can be shaped by previous<br />
This very simple wayresponses. <strong>of</strong> memory These supply recurrent yields connections architecturally are determ<strong>in</strong>ed plausible<br />
properties that can abstractly what give be described the network asmemory. storage limitations, memory span or decay<br />
<strong>of</strong> memorized representations over time. These are properties explicitly accounted for <strong>in</strong><br />
symbolic models like ACT-R or CC-READER.<br />
rchitecture used by Jordan (1986).<br />
from output to state units are one-forxed<br />
weight <strong>of</strong> 1.0. Not all connections<br />
ach can be modified <strong>in</strong><br />
<strong>in</strong>g way. Suppose a<br />
own <strong>in</strong> Figure 2) is<br />
at the <strong>in</strong>put level by<br />
nits; call these Context<br />
units are also “hidden”<br />
se that they <strong>in</strong>teract<br />
with other nodes<br />
he network, and not the<br />
ld.<br />
that there is a<br />
nput to be processed,<br />
clock which regulates<br />
<strong>of</strong> the <strong>in</strong>put to the<br />
cess<strong>in</strong>g would then<br />
e follow<strong>in</strong>g sequence <strong>of</strong><br />
time t, the <strong>in</strong>put units<br />
first <strong>in</strong>put <strong>in</strong> the sequence. Each <strong>in</strong>put might be a s<strong>in</strong>gle scalar value or a vector,<br />
n the nature <strong>of</strong> the problem. The context units are <strong>in</strong>itially set to 0.5. 2 OUTPUT UNITS<br />
HIDDEN UNITS<br />
INPUT UNITS CONTEXT UNITS<br />
Figure 2. A simple recurrent network <strong>in</strong> which activations are<br />
Figure 3.1: copied Architecture from hidden <strong>of</strong> layer a simple to context recurrent layer network on a one-for-one (SRN, Elman, 1990). The<br />
solid l<strong>in</strong>ebasis, represents with fixed fixed weight one-to-one <strong>of</strong> 1.0. Dotted connections l<strong>in</strong>es represent to thetra<strong>in</strong>able context layer. Dashed l<strong>in</strong>es<br />
connections.<br />
represent tra<strong>in</strong>able connections.<br />
Both the <strong>in</strong>put<br />
ntext units activate the hidden units; and then the hidden units feed forward to<br />
48<br />
tion function used here bounds values between 0.0 and 1.0.<br />
Page 4