29.08.2013 Views

Connectionist Modeling of Experience-based Effects in Sentence ...

Connectionist Modeling of Experience-based Effects in Sentence ...

Connectionist Modeling of Experience-based Effects in Sentence ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

OUTPUT<br />

PUT PLAN<br />

Chapter 3 <strong>Connectionist</strong> Modell<strong>in</strong>g <strong>of</strong> Language Comprehension<br />

vides much contextual <strong>in</strong>formation help<strong>in</strong>g to <strong>in</strong>terpret current <strong>in</strong>put. The context <strong>of</strong><br />

an utterance has a great <strong>in</strong>fluence on ambiguity resolution and predictions <strong>of</strong> content.<br />

There have been someThere accounts are many <strong>of</strong> provid<strong>in</strong>g ways <strong>in</strong> connectionist which this can networks be with temporal<br />

representation which were accomplished, <strong>of</strong> explicit and nature. a number This posed <strong>of</strong> <strong>in</strong>terest<strong>in</strong>g limits to the number and richness<br />

<strong>of</strong> representations. proposals Elman (1990) have appeared describes<strong>in</strong> athe simple literature way to (e.g. provide a connectionist<br />

network with memory, called Jordan, a simple 1986; recurrent Tank & network Hopfield, (SRN, 1987; figure 3.1). The hidden<br />

representations <strong>in</strong> the network Stornetta, are Hogg, copied & <strong>in</strong>to Huberman, a so-called1987; context layer, which <strong>in</strong>fluences<br />

the hidden representations Watrous <strong>in</strong> the next & Shastri, step through 1987; weighted Waibel, activation feed<strong>in</strong>g. This<br />

memory loop goes without Hanazawa, any explicit H<strong>in</strong>ton, representation Shikano, & <strong>of</strong> Lang, time1987; or relations between <strong>in</strong>put<br />

chunks. It is the iterative P<strong>in</strong>eda, procedure 1988; Williams <strong>of</strong> copy<strong>in</strong>g & Zipser, and back-feed<strong>in</strong>g 1988). One itself that produces<br />

temporal relations on an<strong>of</strong> implicit the most level. promis<strong>in</strong>g Becausewas every suggested copy <strong>of</strong> the by activation pattern has<br />

been <strong>in</strong>fluenced by all earlier Jordan copies, (1986). the Jordan contextual described memory a network reaches <strong>in</strong>to the “past” <strong>in</strong> a<br />

cont<strong>in</strong>uously graded way (shown over several <strong>in</strong> Figure <strong>in</strong>put1) steps. conta<strong>in</strong><strong>in</strong>g The <strong>in</strong>formation recurrent <strong>of</strong> earlier <strong>in</strong>put representations<br />

is still <strong>in</strong> the connections representation which aswere a trace, used but to associate newer <strong>in</strong>put a has more <strong>in</strong>fluential<br />

power. Elman (1990) writes: static pattern (a “Plan”) with a serially<br />

ordered output pattern (a sequence <strong>of</strong><br />

“In this account, “Actions”). memory isThe neither recurrent passive connections nor a separate allowsubsystem.<br />

One<br />

cannot properly speak the network’s <strong>of</strong> a memory hidden forunits sequences; to see that its own memory is <strong>in</strong>extricably<br />

bound up withprevious the rest output, <strong>of</strong> the process<strong>in</strong>g so that the mechanism.” subsequent<br />

behavior can be shaped by previous<br />

This very simple wayresponses. <strong>of</strong> memory These supply recurrent yields connections architecturally are determ<strong>in</strong>ed plausible<br />

properties that can abstractly what give be described the network asmemory. storage limitations, memory span or decay<br />

<strong>of</strong> memorized representations over time. These are properties explicitly accounted for <strong>in</strong><br />

symbolic models like ACT-R or CC-READER.<br />

rchitecture used by Jordan (1986).<br />

from output to state units are one-forxed<br />

weight <strong>of</strong> 1.0. Not all connections<br />

ach can be modified <strong>in</strong><br />

<strong>in</strong>g way. Suppose a<br />

own <strong>in</strong> Figure 2) is<br />

at the <strong>in</strong>put level by<br />

nits; call these Context<br />

units are also “hidden”<br />

se that they <strong>in</strong>teract<br />

with other nodes<br />

he network, and not the<br />

ld.<br />

that there is a<br />

nput to be processed,<br />

clock which regulates<br />

<strong>of</strong> the <strong>in</strong>put to the<br />

cess<strong>in</strong>g would then<br />

e follow<strong>in</strong>g sequence <strong>of</strong><br />

time t, the <strong>in</strong>put units<br />

first <strong>in</strong>put <strong>in</strong> the sequence. Each <strong>in</strong>put might be a s<strong>in</strong>gle scalar value or a vector,<br />

n the nature <strong>of</strong> the problem. The context units are <strong>in</strong>itially set to 0.5. 2 OUTPUT UNITS<br />

HIDDEN UNITS<br />

INPUT UNITS CONTEXT UNITS<br />

Figure 2. A simple recurrent network <strong>in</strong> which activations are<br />

Figure 3.1: copied Architecture from hidden <strong>of</strong> layer a simple to context recurrent layer network on a one-for-one (SRN, Elman, 1990). The<br />

solid l<strong>in</strong>ebasis, represents with fixed fixed weight one-to-one <strong>of</strong> 1.0. Dotted connections l<strong>in</strong>es represent to thetra<strong>in</strong>able context layer. Dashed l<strong>in</strong>es<br />

connections.<br />

represent tra<strong>in</strong>able connections.<br />

Both the <strong>in</strong>put<br />

ntext units activate the hidden units; and then the hidden units feed forward to<br />

48<br />

tion function used here bounds values between 0.0 and 1.0.<br />

Page 4

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!