10.07.2015 Views

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.38Introduction to Neural NetworksIn the field of neural networks, we study the properties of networks of idealized‘neurons’.Three motivations underlie work in this broad <strong>and</strong> interdisciplinary field.Biology. The task of underst<strong>and</strong>ing how the brain works is one of the outst<strong>and</strong>ingunsolved problems in science. Some neural network models areintended to shed light on the way in which computation <strong>and</strong> memoryare performed by brains.Engineering. Many researchers would like to create machines that can ‘learn’,perform ‘pattern recognition’ or ‘discover patterns in data’.Complex systems. A third motivation for being interested in neural networksis that they are complex adaptive systems whose properties areinteresting in their own right.I should emphasize several points at the outset.• This book gives only a taste of this field. There are many interestingneural network models which we will not have time to touch on.• The models that we discuss are not intended to be faithful models ofbiological systems. If they are at all relevant to biology, their relevanceis on an abstract level.• I will describe some neural network methods that are widely used innonlinear data modelling, but I will not be able to give a full descriptionof the state of the art. If you wish to solve real problems with neuralnetworks, please read the relevant papers.38.1 MemoriesIn the next few chapters we will meet several neural network models whichcome with simple learning algorithms which make them function as memories.Perhaps we should dwell for a moment on the conventional idea of memoryin digital computation. A memory (a string of 5000 bits describing the nameof a person <strong>and</strong> an image of their face, say) is stored in a digital computerat an address. To retrieve the memory you need to know the address. Theaddress has nothing to do with the memory itself. Notice the properties thatthis scheme does not have:1. Address-based memory is not associative. Imagine you know half of amemory, say someone’s face, <strong>and</strong> you would like to recall the rest of the468

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!