08.10.2016 Views

Foundations of Data Science

2dLYwbK

2dLYwbK

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

model, the energy <strong>of</strong> the system is given by<br />

(<br />

f(x 1 , x 2 , . . . , x n ) = exp c ∑ i∼j<br />

|x i − x j |<br />

)<br />

.<br />

The constant c can be positive or negative. If c < 0, then energy is lower if many adjacent<br />

pairs have opposite spins and if c > 0 the reverse holds. The model was first<br />

used to model probabilities <strong>of</strong> spin configurations. The hypothesis was that for each<br />

{x 1 , x 2 , . . . , x n } in {−1, +1} n , the energy <strong>of</strong> the configuration with these spins is proportional<br />

to f(x 1 , x 2 , . . . , x n ).<br />

In most computer science settings, such functions are mainly used as objective functions<br />

that are to be optimized subject to some constraints. The problem is to find the<br />

minimum energy set <strong>of</strong> spins under some constraints on the spins. Usually the constraints<br />

just specify the spins <strong>of</strong> some particles. Note that when c > 0, this is the problem <strong>of</strong><br />

minimizing ∑ |x i − x j | subject to the constraints. The objective function is convex and<br />

i∼j<br />

so this can be done efficiently. If c < 0, however, we need to minimize a concave function<br />

for which there is no known efficient algorithm. The minimization <strong>of</strong> a concave function<br />

in general is NP-hard.<br />

A second important motivation comes from the area <strong>of</strong> vision. It has to to do with<br />

reconstructing images. Suppose we are given observations <strong>of</strong> the intensity <strong>of</strong> light at<br />

individual pixels, x 1 , x 2 , . . . , x n , and wish to compute the true values, the true intensities,<br />

<strong>of</strong> these variables y 1 , y 2 , . . . , y n . There may be two sets <strong>of</strong> constraints, the first stipulating<br />

that the y i must be close to the corresponding x i and the second, a term correcting possible<br />

observation errors, stipulating that y i must be close to the values <strong>of</strong> y j for j ∼ i. This<br />

can be formulated as ( ∑<br />

min |x i − y i | + ∑ )<br />

|y i − y j | ,<br />

y<br />

i<br />

i∼j<br />

where the values <strong>of</strong> x i are constrained to be the observed values. The objective function<br />

is convex and polynomial time minimization algorithms exist. Other objective functions<br />

using say sum <strong>of</strong> squares instead <strong>of</strong> sum <strong>of</strong> absolute values can be used and there are<br />

polynomial time algorithms as long as the function to be minimized is convex.<br />

More generally, the correction term may depend on all grid points within distance two<br />

<strong>of</strong> each point rather than just immediate neighbors. Even more generally, we may have n<br />

variables y 1 , y 2 , . . . y n with the value <strong>of</strong> some already specified and subsets S 1 , S 2 , . . . S m <strong>of</strong><br />

these variables constrained in some way. The constraints are accumulated into one objective<br />

function which is a product <strong>of</strong> functions f 1 , f 2 , . . . , f m , where function f i is evaluated<br />

on the variables in subset S i . The problem is to minimize ∏ m<br />

i=1 f i(y j , j ∈ S i ) subject to<br />

constrained values. Note that the vision example had a sum instead <strong>of</strong> a product, but by<br />

taking exponentials we can turn the sum into a product as in the Ising model.<br />

310

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!