01.08.2013 Views

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981<br />

You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.<br />

272 19 — Why have Sex? <strong>Information</strong> Acquisition <strong>and</strong> Evolution<br />

<strong>and</strong> variance<br />

m(1 − m)<br />

G<br />

m<br />

. (19.4)<br />

G<br />

If the population of parents has mean δf(t) <strong>and</strong> variance σ2 (t) ≡ βm/G, then<br />

the child population, before selection, will have mean (1 − 2m)δf(t) <strong>and</strong> variance<br />

(1+β)m/G. Natural selection chooses the upper half of this distribution,<br />

so the mean fitness <strong>and</strong> variance of fitness at the next generation are given by<br />

δf(t+1) = (1 − 2m)δf(t) + α <br />

m<br />

(1 + β) , (19.5)<br />

G<br />

σ 2 (t+1) = γ(1 + β) m<br />

, (19.6)<br />

G<br />

where α is the mean deviation from the mean, measured in st<strong>and</strong>ard deviations,<br />

<strong>and</strong> γ is the factor by which the child distribution’s variance is reduced<br />

by selection. The numbers α <strong>and</strong> γ are of order 1. For the case of a Gaussian<br />

distribution, α = 2/π 0.8 <strong>and</strong> γ = (1 − 2/π) 0.36. If we assume that<br />

the variance is in dynamic equilibrium, i.e., σ2 (t+1) σ2 (t), then<br />

γ(1 + β) = β, so (1 + β) = 1<br />

, (19.7)<br />

1 − γ<br />

<strong>and</strong> the factor α (1 + β) in equation (19.5) is equal to 1, if we take the results<br />

for the Gaussian distribution, an approximation that becomes poorest when<br />

the discreteness of fitness becomes important, i.e., for small m. The rate of<br />

increase of normalized fitness is thus:<br />

df<br />

dt<br />

−2m δf +<br />

which, assuming G(δf) 2 ≫ 1, is maximized for<br />

mopt =<br />

<br />

m<br />

, (19.8)<br />

G<br />

1<br />

, (19.9)<br />

16G(δf) 2<br />

at which point, <br />

df 1<br />

= . (19.10)<br />

dt opt 8G(δf)<br />

So the rate of increase of fitness F = fG is at most<br />

dF<br />

dt<br />

= 1<br />

8(δf)<br />

per generation. (19.11)<br />

For a population with low fitness (δf < 0.125), the rate of increase of fitness<br />

may exceed 1 unit per generation. Indeed, if δf 1/ √ G, the rate of increase, if<br />

m = 1/2, is of order √ G; this initial spurt can last only of order √ G generations.<br />

For δf > 0.125, the rate of increase of fitness is smaller than one per generation.<br />

As the fitness approaches G, the optimal mutation rate tends to m = 1/(4G), so<br />

that an average of 1/4 bits are flipped per genotype, <strong>and</strong> the rate of increase of<br />

fitness is also equal to 1/4; information is gained at a rate of about 0.5 bits per<br />

generation. It takes about 2G generations for the genotypes of all individuals<br />

in the population to attain perfection.<br />

For fixed m, the fitness is given by<br />

δf(t) =<br />

1<br />

2 √ mG (1 − c e−2mt ), (19.12)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!