01.08.2013 Views

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981<br />

You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.<br />

42 2 — Probability, Entropy, <strong>and</strong> <strong>Inference</strong><br />

so<br />

f(a) ≡<br />

<br />

1 1 1 − e−a = + 1<br />

1 + exp(−a) 2 1 + e−a = 1<br />

<br />

<br />

2<br />

= 1<br />

(tanh(a/2) + 1).<br />

2<br />

(2.67)<br />

ea/2 − e−a/2 ea/2 + 1<br />

+ e−a/2 In the case b = log2 p/q, we can repeat steps (2.63–2.65), replacing e by 2,<br />

to obtain<br />

1<br />

p = . (2.68)<br />

1 + 2−b Solution to exercise 2.18 (p.36).<br />

⇒<br />

⇒ log<br />

P (x | y) =<br />

P (y | x)P (x)<br />

P (y)<br />

(2.69)<br />

P (x = 1 | y)<br />

P (x = 0 | y)<br />

=<br />

P (y | x = 1) P (x = 1)<br />

P (y | x = 0) P (x = 0)<br />

(2.70)<br />

P (x = 1 | y)<br />

P (x = 0 | y)<br />

=<br />

P (y | x = 1) P (x = 1)<br />

log + log .<br />

P (y | x = 0) P (x = 0)<br />

(2.71)<br />

Solution to exercise 2.19 (p.36). The conditional independence of d1 <strong>and</strong> d2<br />

given x means<br />

P (x, d1, d2) = P (x)P (d1 | x)P (d2 | x). (2.72)<br />

This gives a separation of the posterior probability ratio into a series of factors,<br />

one for each data point, times the prior probability ratio.<br />

P (x = 1 | {di})<br />

P (x = 0 | {di}) = P ({di} | x = 1) P (x = 1)<br />

P ({di} | x = 0) P (x = 0)<br />

(2.73)<br />

= P (d1 | x = 1) P (d2 | x = 1) P (x = 1)<br />

.<br />

P (d1 | x = 0) P (d2 | x = 0) P (x = 0)<br />

(2.74)<br />

Life in high-dimensional spaces<br />

Solution to exercise 2.20 (p.37). The volume of a hypersphere of radius r in<br />

N dimensions is in fact<br />

V (r, N) = πN/2<br />

(N/2)! rN , (2.75)<br />

but you don’t need to know this. For this question all that we need is the<br />

r-dependence, V (r, N) ∝ r N . So the fractional volume in (r − ɛ, r) is<br />

r N − (r − ɛ) N<br />

r N<br />

<br />

= 1 − 1 − ɛ<br />

r<br />

The fractional volumes in the shells for the required cases are:<br />

N<br />

N 2 10 1000<br />

ɛ/r = 0.01 0.02 0.096 0.99996<br />

ɛ/r = 0.5 0.75 0.999 1 − 2 −1000<br />

. (2.76)<br />

Notice that no matter how small ɛ is, for large enough N essentially all the<br />

probability mass is in the surface shell of thickness ɛ.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!