10.07.2015 Views

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.34Independent Component Analysis <strong>and</strong>Latent Variable Modelling34.1 Latent variable modelsMany statistical models are generative models (that is, models that specifya full probability density over all variables in the situation) that make use oflatent variables to describe a probability distribution over observables.Examples of latent variable models include Chapter 22’s mixture models,which model the observables as coming from a superposed mixture of simpleprobability distributions (the latent variables are the unknown class labelsof the examples); hidden Markov models (Rabiner <strong>and</strong> Juang, 1986; Durbinet al., 1998); <strong>and</strong> factor analysis.The decoding problem for error-correcting codes can also be viewed interms of a latent variable model – figure 34.1. In that case, the encodingmatrix G is normally known in advance. In latent variable modelling, theparameters equivalent to G are usually not known, <strong>and</strong> must be inferred fromthe data along with the latent variables s.Usually, the latent variables have a simple distribution, often a separabledistribution. Thus when we fit a latent variable model, we are finding a descriptionof the data in terms of ‘independent components’. The ‘independentcomponent analysis’ algorithm corresponds to perhaps the simplest possiblelatent variable model with continuous latent variables.y 1s 1Gs KFigure 34.1. Error-correctingcodes as latent variable models.The K latent variables are theindependent source bitss 1 , . . . , s K ; these give rise to theobservables via the generatormatrix G.y N34.2 The generative model for independent component analysisA set of N observations D = {x (n) } N n=1 are assumed to be generated as follows.Each J-dimensional vector x is a linear mixture of I underlying source signals,s:x = Gs, (34.1)where the matrix of mixing coefficients G is not known.The simplest algorithm results if we assume that the number of sourcesis equal to the number of observations, i.e., I = J. Our aim is to recoverthe source variables s (within some multiplicative factors, <strong>and</strong> possibly permuted).To put it another way, we aim to create the inverse of G (within apost-multiplicative factor) given only a set of examples {x}. We assume thatthe latent variables are independently distributed, with marginal distributionsP (s i | H) ≡ p i (s i ). Here H denotes the assumed form of this model <strong>and</strong> theassumed probability distributions p i of the latent variables.The probability of the observables <strong>and</strong> the hidden variables, given G <strong>and</strong>437

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!