08.10.2016 Views

Foundations of Data Science

2dLYwbK

2dLYwbK

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

C 1<br />

S 2<br />

C 2<br />

causes<br />

D 1<br />

D 2<br />

diseases<br />

S 1<br />

symptoms<br />

Figure 9.1: A Bayesian network<br />

contact with farm animals, whether he had eaten certain foods, or whether the patient has<br />

an hereditary predisposition to any diseases. Using the above Bayesian network where the<br />

variables are true or false, the doctor may wish to determine one <strong>of</strong> two things. What is<br />

the marginal probability <strong>of</strong> a given disease or what is the most likely set <strong>of</strong> diseases. In determining<br />

the most likely set <strong>of</strong> diseases, we are given a T or F assignment to the causes<br />

and symptoms and ask what assignment <strong>of</strong> T or F to the diseases maximizes the joint<br />

probability. This latter problem is called the maximum a posteriori probability (MAP).<br />

Given the conditional probabilities and the probabilities p (C 1 ) and p (C 2 ) in Figure<br />

9.1, the joint probability p (C 1 , C 2 , D 1 , . . .) can be computed easily for any combination<br />

<strong>of</strong> values <strong>of</strong> C 1 , C 2 , D 1 , . . .. However, we might wish to find that value <strong>of</strong> the variables<br />

<strong>of</strong> highest probability (MAP) or we might want one <strong>of</strong> the marginal probabilities p (D 1 )<br />

or p (D 2 ). The obvious algorithms for these two problems require evaluating the probability<br />

p (C 1 , C 2 , D 1 , . . .) over exponentially many input values or summing the probability<br />

p (C 1 , C 2 , D 1 , . . .) over exponentially many values <strong>of</strong> the variables other than those for<br />

which we want the marginal probability. In certain situations, when the joint probability<br />

distribution can be expressed as a product <strong>of</strong> factors, a belief propagation algorithm can<br />

solve the maximum a posteriori problem or compute all marginal probabilities quickly.<br />

9.5 Markov Random Fields<br />

The Markov random field model arose first in statistical mechanics where it was called<br />

the Ising model. It is instructive to start with a description <strong>of</strong> it. The simplest version<br />

<strong>of</strong> the Ising model consists <strong>of</strong> n particles arranged in a rectangular √ n × √ n grid. Each<br />

particle can have a spin that is denoted ±1. The energy <strong>of</strong> the whole system depends<br />

on interactions between pairs <strong>of</strong> neighboring particles. Let x i be the spin, ±1, <strong>of</strong> the i th<br />

particle. Denote by i ∼ j the relation that i and j are adjacent in the grid. In the Ising<br />

309

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!