10.07.2015 Views

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

Information Theory, Inference, and Learning ... - Inference Group

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.360 29 — Monte Carlo Methodswhere Z = ∫ dx dy P ∗ (x) is the volume of the lake. You are provided with aboat, a satellite navigation system, <strong>and</strong> a plumbline. Using the navigator, youcan take your boat to any desired location x on the map; using the plumblineyou can measure P ∗ (x) at that point. You can also measure the planktonconcentration there.Problem 1 is to draw 1 cm 3 water samples at r<strong>and</strong>om from the lake, insuch a way that each sample is equally likely to come from any point withinthe lake. Problem 2 is to find the average plankton concentration.These are difficult problems to solve because at the outset we know nothingabout the depth P ∗ (x). Perhaps much of the volume of the lake is containedin narrow, deep underwater canyons (figure 29.3), in which case, to correctlysample from the lake <strong>and</strong> correctly estimate Φ our method must implicitlydiscover the canyons <strong>and</strong> find their volume relative to the rest of the lake.Difficult problems, yes; nevertheless, we’ll see that clever Monte Carlo methodscan solve them.Figure 29.3. A slice through a lakethat includes some canyons.Uniform samplingHaving accepted that we cannot exhaustively visit every location x in thestate space, we might consider trying to solve the second problem (estimatingthe expectation of a function φ(x)) by drawing r<strong>and</strong>om samples {x (r) } R r=1uniformly from the state space <strong>and</strong> evaluating P ∗ (x) at those points. Thenwe could introduce a normalizing constant Z R , defined byZ R =R∑P ∗ (x (r) ), (29.16)r=1<strong>and</strong> estimate Φ = ∫ d N x φ(x)P (x) byˆΦ =R∑r=1φ(x (r) ) P ∗ (x (r) )Z R. (29.17)Is anything wrong with this strategy? Well, it depends on the functions φ(x)<strong>and</strong> P ∗ (x). Let us assume that φ(x) is a benign, smoothly varying function<strong>and</strong> concentrate on the nature of P ∗ (x). As we learnt in Chapter 4, a highdimensionaldistribution is often concentrated in a small region of the statespace known as its typical set T , whose volume is given by |T | ≃ 2 H(X) , whereH(X) is the entropy of the probability distribution P (x). If almost all theprobability mass is located in the typical set <strong>and</strong> φ(x) is a benign function,the value of Φ = ∫ d N x φ(x)P (x) will be principally determined by the valuesthat φ(x) takes on in the typical set. So uniform sampling will only st<strong>and</strong>a chance of giving a good estimate of Φ if we make the number of samplesR sufficiently large that we are likely to hit the typical set at least once ortwice. So, how many samples are required? Let us take the case of the Isingmodel again. (Strictly, the Ising model may not be a good example, since itdoesn’t necessarily have a typical set, as defined in Chapter 4; the definitionof a typical set was that all states had log probability close to the entropy,which for an Ising model would mean that the energy is very close to themean energy; but in the vicinity of phase transitions, the variance of energy,also known as the heat capacity, may diverge, which means that the energyof a r<strong>and</strong>om state is not necessarily expected to be very close to the meanenergy.) The total size of the state space is 2 N states, <strong>and</strong> the typical set hassize 2 H . So each sample has a chance of 2 H /2 N of falling in the typical set.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!