01.08.2013 Views

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981<br />

You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.<br />

60 3 — More about <strong>Inference</strong><br />

8<br />

6<br />

4<br />

2<br />

0<br />

-2<br />

-4<br />

(b) P (pa | s = bbb, F = 3) ∝ (1 − pa) 3 . The most probable value of pa (i.e.,<br />

the value that maximizes the posterior probability density) is 0. The<br />

mean value of pa is 1/5.<br />

See figure 3.7b.<br />

H0 is true H1 is true<br />

pa = 1/6<br />

0 50 100 150<br />

1000/1<br />

100/1<br />

10/1<br />

1/1<br />

1/10<br />

1/100<br />

200<br />

pa = 0.25<br />

pa = 0.5<br />

8<br />

6<br />

4<br />

2<br />

1000/1<br />

100/1<br />

10/1<br />

8<br />

6<br />

4<br />

2<br />

1000/1<br />

100/1<br />

10/1<br />

0<br />

1/1 0<br />

1/1<br />

-2<br />

-4<br />

1/10<br />

1/100<br />

-2<br />

-4<br />

1/10<br />

1/100<br />

0 50 100 150 200 0 50 100 150 200<br />

Solution to exercise 3.7 (p.54). The curves in figure 3.8 were found by finding<br />

the mean <strong>and</strong> st<strong>and</strong>ard deviation of Fa, then setting Fa to the mean ± two<br />

st<strong>and</strong>ard deviations to get a 95% plausible range for Fa, <strong>and</strong> computing the<br />

three corresponding values of the log evidence ratio.<br />

Solution to exercise 3.8 (p.57). Let Hi denote the hypothesis that the prize is<br />

behind door i. We make the following assumptions: the three hypotheses H1,<br />

H2 <strong>and</strong> H3 are equiprobable a priori, i.e.,<br />

P (H1) = P (H2) = P (H3) = 1<br />

. (3.36)<br />

3<br />

The datum we receive, after choosing door 1, is one of D = 3 <strong>and</strong> D = 2 (meaning<br />

door 3 or 2 is opened, respectively). We assume that these two possible<br />

outcomes have the following probabilities. If the prize is behind door 1 then<br />

the host has a free choice; in this case we assume that the host selects at<br />

r<strong>and</strong>om between D = 2 <strong>and</strong> D = 3. Otherwise the choice of the host is forced<br />

<strong>and</strong> the probabilities are 0 <strong>and</strong> 1.<br />

P (D = 2 | H1) = 1/2 P (D = 2 | H2) = 0 P (D = 2 | H3) = 1<br />

P (D = 3 | H1) = 1/2 P (D = 3 | H2) = 1 P (D = 3 | H3) = 0<br />

(3.37)<br />

Now, using Bayes’ theorem, we evaluate the posterior probabilities of the<br />

hypotheses:<br />

P (Hi | D = 3) = P (D = 3 | Hi)P (Hi)<br />

(3.38)<br />

P (D = 3)<br />

P (H1 | D = 3) = (1/2)(1/3)<br />

P (D=3)<br />

P (H2 | D = 3) = (1)(1/3)<br />

P (D=3) P (H3 | D = 3) = (0)(1/3)<br />

P (D=3)<br />

(3.39)<br />

The denominator P (D = 3) is (1/2) because it is the normalizing constant for<br />

this posterior distribution. So<br />

P (H1 | D = 3) = 1/3 P (H2 | D = 3) = 2/3 P (H3 | D = 3) = 0.<br />

(3.40)<br />

So the contestant should switch to door 2 in order to have the biggest chance<br />

of getting the prize.<br />

Many people find this outcome surprising. There are two ways to make it<br />

more intuitive. One is to play the game thirty times with a friend <strong>and</strong> keep<br />

track of the frequency with which switching gets the prize. Alternatively, you<br />

can perform a thought experiment in which the game is played with a million<br />

doors. The rules are now that the contestant chooses one door, then the game<br />

Figure 3.8. Range of plausible<br />

values of the log evidence in<br />

favour of H1 as a function of F .<br />

The vertical axis on the left is<br />

log ; the right-h<strong>and</strong><br />

P (s | F,H1)<br />

P (s | F,H0)<br />

vertical axis shows the values of<br />

P (s | F,H1)<br />

P (s | F,H0) .<br />

The solid line shows the log<br />

evidence if the r<strong>and</strong>om variable<br />

Fa takes on its mean value,<br />

Fa = paF . The dotted lines show<br />

(approximately) the log evidence<br />

if Fa is at its 2.5th or 97.5th<br />

percentile.<br />

(See also figure 3.6, p.54.)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!