10.07.2015 Views

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.25Exact Marginalization in TrellisesIn this chapter we will discuss a few exact methods that are used in probabilisticmodelling. As an example we will discuss the task of decoding a linearerror-correcting code. We will see that inferences can be conducted most efficientlyby message-passing algorithms, which take advantage of the graphicalstructure of the problem to avoid unnecessary duplication of computations(see Chapter 16).25.1 Decoding problemsA codeword t is selected from a linear (N, K) code C, <strong>and</strong> it is transmittedover a noisy channel; the received signal is y. In this chapter we will assumethat the channel is a memoryless channel such as a Gaussian channel. Givenan assumed channel model P (y | t), there are two decoding problems.The codeword decoding problem is the task of inferring which codewordt was transmitted given the received signal.The bitwise decoding problem is the task of inferring for each transmittedbit t n how likely it is that that bit was a one rather than a zero.As a concrete example, take the (7, 4) Hamming code. In Chapter 1, wediscussed the codeword decoding problem for that code, assuming a binarysymmetric channel. We didn’t discuss the bitwise decoding problem <strong>and</strong> wedidn’t discuss how to h<strong>and</strong>le more general channel models such as a Gaussianchannel.Solving the codeword decoding problemBy Bayes’ theorem, the posterior probability of the codeword t isP (t | y) =P (y | t)P (t). (25.1)P (y)Likelihood function. The first factor in the numerator, P (y | t), is the likelihoodof the codeword, which, for any memoryless channel, is a separablefunction,N∏P (y | t) = P (y n | t n ). (25.2)n=1For example, if the channel is a Gaussian channel with transmissions ±x<strong>and</strong> additive noise of st<strong>and</strong>ard deviation σ, then the probability density324

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!