01.08.2013 Views

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

Information Theory, Inference, and Learning ... - MAELabs UCSD

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/0521642981<br />

You can buy this book for 30 pounds or $50. See http://www.inference.phy.cam.ac.uk/mackay/itila/ for links.<br />

258 17 — Communication over Constrained Noiseless Channels<br />

Solution to exercise 17.10 (p.255). Let the invariant distribution be<br />

P (s) = αe (L)<br />

s e(R)<br />

s , (17.22)<br />

where α is a normalization constant. The entropy of St given St−1, assuming Here, as in Chapter 4, St denotes<br />

St−1 comes from the invariant distribution, is<br />

H(St|St−1) = −<br />

the ensemble whose r<strong>and</strong>om<br />

variable is the state st.<br />

<br />

= − <br />

s,s ′<br />

α e (R)<br />

s<br />

s,s ′<br />

= − <br />

αe (L)<br />

s,s ′<br />

e (L)<br />

s ′ As ′ <br />

s<br />

log e<br />

λ<br />

(L)<br />

s ′<br />

P (s)P (s ′ |s) log P (s ′ |s) (17.23)<br />

s e (R)<br />

s<br />

e (L)<br />

s ′ As ′ s<br />

λe (L)<br />

s<br />

log e(L)<br />

s ′ As ′ s<br />

λe (L)<br />

s<br />

(17.24)<br />

+ log As ′ s − log λ − log e (L)<br />

<br />

s . (17.25)<br />

Now, As ′ s is either 0 or 1, so the contributions from the terms proportional to<br />

As ′ s log As ′ s are all zero. So<br />

H(St|St−1) = log λ + − α <br />

λ<br />

s ′<br />

<br />

<br />

s<br />

<br />

α <br />

e<br />

λ<br />

(L)<br />

s ′ As ′ <br />

s<br />

= log λ − α<br />

λ<br />

<br />

s ′<br />

s<br />

s ′<br />

λe (R)<br />

s ′ e (L)<br />

s ′ log e (L)<br />

s ′ + α<br />

λ<br />

As ′ se (R)<br />

s<br />

<br />

e (R)<br />

s log e (L)<br />

s<br />

<br />

s<br />

e (L)<br />

s ′ log e (L)<br />

s ′ +<br />

(17.26)<br />

λe (L)<br />

s e(R) s log e (L)<br />

s (17.27)<br />

= log λ. (17.28)<br />

Solution to exercise 17.11 (p.255). The principal eigenvalues of the connection<br />

matrices of the two channels are 1.839 <strong>and</strong> 1.928. The capacities (log λ) are<br />

0.879 <strong>and</strong> 0.947 bits.<br />

Solution to exercise 17.12 (p.256). The channel is similar to the unconstrained<br />

binary channel; runs of length greater than L are rare if L is large, so we only<br />

expect weak differences from this channel; these differences will show up in<br />

contexts where the run length is close to L. The capacity of the channel is<br />

very close to one bit.<br />

A lower bound on the capacity is obtained by considering the simple<br />

variable-length code for this channel which replaces occurrences of the maximum<br />

runlength string 111. . .1 by 111. . .10, <strong>and</strong> otherwise leaves the source file<br />

unchanged. The average rate of this code is 1/(1 + 2 −L ) because the invariant<br />

distribution will hit the ‘add an extra zero’ state a fraction 2 −L of the time.<br />

We can reuse the solution for the variable-length channel in exercise 6.18<br />

(p.125). The capacity is the value of β such that the equation<br />

L+1 <br />

Z(β) = 2 −βl = 1 (17.29)<br />

l=1<br />

is satisfied. The L+1 terms in the sum correspond to the L+1 possible strings<br />

that can be emitted, 0, 10, 110, . . . , 11. . .10. The sum is exactly given by:<br />

Z(β) = 2 −β<br />

<br />

2−β L+1 − 1<br />

2−β . (17.30)<br />

− 1

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!