31.07.2015 Views

Steven Pinker -- How the Mind Works - Hampshire High Italian ...

Steven Pinker -- How the Mind Works - Hampshire High Italian ...

Steven Pinker -- How the Mind Works - Hampshire High Italian ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

110 I HOW THE MIND WORKSup its inputs and get <strong>the</strong> right answer. Here is how it can be done forexclusive-or:Cy) AxorB.6/ NO 2(AorB) (5), Y5) (AandB).6 >v -4The two hidden units between <strong>the</strong> input and <strong>the</strong> output calculateuseful intermediate products. The one on <strong>the</strong> left computes <strong>the</strong> simplecase of "A or B," which in turn simply excites <strong>the</strong> output node. The oneon <strong>the</strong> right computes <strong>the</strong> vexing case of "A and B," and it inhibits <strong>the</strong>output node. The output node can simply compute "(A or B) and not (Aand B)," which is well within its feeble powers. Note that even at <strong>the</strong>microscopic level of building <strong>the</strong> simplest demons out of toy neurons,internal representations are indispensable; stimulus-response connectionsare not enough.Even better, a hidden-layer network can be trained to set its ownweights, using a fancier version of <strong>the</strong> perceptron learning procedure. Asbefore, a teacher gives <strong>the</strong> network <strong>the</strong> correct output for every input, and<strong>the</strong> network adjusts <strong>the</strong> connection weights up or down to try to reduce<strong>the</strong> difference. But that poses a problem <strong>the</strong> perceptron did not have toworry about: how to adjust <strong>the</strong> connections from <strong>the</strong> input units to <strong>the</strong>hidden units. It is problematic because <strong>the</strong> teacher, unless it is 1 a mindreader, has no way of knowing <strong>the</strong> "correct" states for <strong>the</strong> hidden units,which are sealed inside <strong>the</strong> network. The psychologists David Rumelhart,Geoffrey Hinton, and Ronald Williams hit on a clever solution. The outputunits propagate back to each hidden unit a signal that represents <strong>the</strong>sum of <strong>the</strong> hidden unit's errors across all <strong>the</strong> output units it connects to("you're sending too much activation," or "you're sending too little activation,"and by what amount). That signal can serve as a surrogate teachingsignal which may be used to adjust <strong>the</strong> hidden layer's inputs. The connectionsfrom <strong>the</strong> input layer to each hidden unit can be nudged up or downto reduce <strong>the</strong> hidden unit's tendency to overshoot or undershoot, given<strong>the</strong> current input pattern. This procedure, called "error back-propagation"or simply "backprop," can be iterated backwards to any number of layers.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!