28.07.2014 Views

STA 36-786: Bayesian Theoretical Statistics I

STA 36-786: Bayesian Theoretical Statistics I

STA 36-786: Bayesian Theoretical Statistics I

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>STA</strong> <strong>36</strong>-<strong>786</strong>: <strong>Bayesian</strong> <strong>Theoretical</strong> <strong>Statistics</strong> I<br />

Assignment 1: Spring 2013<br />

Due: 1/31/2013 at 10:30 a.m.<br />

Show all your work to obtain full/partial credit. You are not to consult outside sources<br />

other than your class notes/slides for this assignment (except for the instructor or TA). No<br />

late assignments will be accepted. Please follow the instructions for writing up solutions<br />

that were given out in class on 1.12.13.<br />

1. (a) Find the Bayes estimator when the loss function is L(θ, δ(x)) = c (θ − δ(x)) 2 ,<br />

where c is a constant.<br />

(b) Derive the Bayes estimator when L(θ, δ(x)) = w(θ)(g(θ) − δ(x)) 2 . Do so without<br />

writing any integrals. Note that you can write ρ(π, δ(x)) = E[L(θ, δ(x))|X].<br />

Solution: When L(θ, δ) = c(θ − δ) 2 , the Bayes estimator is simply the posterior mean.<br />

When L(θ, δ) = w(θ)(g(θ) − δ(x)) 2 , we want to minimize<br />

ρ = E[w(θ){g(θ) − δ(x)} 2 | X] = E[w(θ){g(θ) 2 + δ(x) 2 − 2δ(x)g(θ)} | X].<br />

Then<br />

∂ρ<br />

E[w(θ)g(θ) | X]<br />

= 2E[w(θ)δ(x) | X] − 2E[w(θ)g(θ) | X] = 0 =⇒ δ(x) =<br />

∂δ(x) E[w(θ) | X]<br />

Then verify that this is indeed a minimum.<br />

2. Consider the decision problem in which Θ = {θ 1 , θ 2 }, A = {a 1 , a 2 , a 3 , a 4 , a 5 }, and the<br />

loss function is given as follows:<br />

L(θ 1 , a 1 ) = 0, L(θ 1 , a 2 ) = 3, L(θ 1 , a 3 ) = 1, L(θ 1 , a 4 ) = 3, L(θ 1 , a 5 ) = 4;<br />

L(θ 2 , a 1 ) = 4, L(θ 2 , a 2 ) = 6, L(θ 2 , a 3 ) = 0, L(θ 2 , a 4 ) = 0, L(θ 2 , a 5 ) = 1.<br />

Consider the prior π under which π(θ 1 ) = 4/5 and π(θ 2 ) = 1/5. Find the Bayes<br />

action(s) under this prior.<br />

Solution: Note that this is a no data decision problem in that we don’t observe X|θ.<br />

Thus, simply find the δ = a i such that E θ [L(θ, a i )] is minimized.<br />

Recall that the posterior risk ρ(π, a i ) = ∑ i L(θ, a i)π(θ i ). Then<br />

ρ(π, a 1 ) = 4/5,<br />

Thus, a 1 and a 3 are Bayes actions.<br />

ρ(π, a 2 ) = 4/5 × 3 + 1/5 × 6 = 18/5,<br />

ρ(π, a 3 ) = 4/5 × 1 = 4/5,<br />

ρ(π, a 4 ) = 4/5 × 3 = 12/5,<br />

ρ(π, a 5 ) = 4/5 × 4 + 1/5 × 1 = 17/5.


3. Suppose that both the parameter space and the action space are the [0, 1] interval and<br />

the loss L(θ, a) = 100(a − θ) 2 . Consider the prior π(θ) = 2θ, 0 ≤ θ ≤ 1.<br />

(a) Show that the Bayes action under the prior π is given by a = 2/3 and that the<br />

corresponding Bayes risk is 50/9.<br />

(b) Next, suppose that one changes the prior to π(θ) = 3θ 2 , 0 ≤ θ ≤ 1, but all else<br />

remains the same. What do you know about the Bayes action under squared<br />

error loss for this prior? What is the Bayes action now?<br />

Solution: Note that this is a no data problem so R(θ, δ) = L(θ, δ). Then<br />

∫<br />

ρ(π, a) = R(θ, δ) π(θ) dθ<br />

= 100<br />

= 200<br />

= 200<br />

∫ 1<br />

0<br />

∫ 1<br />

0<br />

∫ 1<br />

0<br />

(a − θ) 2 2θ dθ<br />

(a − θ) 2 θ dθ<br />

(a 2 − 2θ + θ 2 ) θ dθ<br />

= 200(a 2 /2 − 2a/3 + 1/4).<br />

Then ∂r<br />

∂a = 200(a − 2/3) and ∂2 r<br />

= 200 > 0. Thus, r(π, a) is minimized at a = 2/3.<br />

∂a2 The corresponding Bayes risk is<br />

r(π, 2/3) = 200(2/9 − 4/9 + 1/4) = 200/<strong>36</strong> = 50/9.<br />

Since the weight is constant, the Bayes rule is the same as under squared error loss.<br />

Thus, the Bayes action is E(θ) = ∫ 1<br />

0 2θ2 dθ = 2/3.<br />

Under the other prior, the Bayes action is simply ∫ 1<br />

0 3θ3 dθ = 3/4.<br />

4. Suppose that the parameter space Θ = (θ 1 , θ 2 ) and that the action space A = (a 1 , a 2 ).<br />

Also suppose that<br />

L(θ 1 , a 1 ) = L(θ 2 , a 2 ) = 0; L(θ 1 , a 2 ) = 5, L(θ 2 , a 1 ) = 10.<br />

The statistician observes a random variable X such that conditional on θ = θ 1 ,<br />

X ∼ N(0, 1), and conditional on θ = θ 2 , X ∼ N(1, 1). Now consider the prior<br />

under which P (θ = θ 1 ) = η. Show that the Bayes rule under this prior is to take<br />

action a 1 if and only if<br />

x ≤ 1 2 + log η<br />

2(1 − η) .<br />

Solution: Consider<br />

p(θ = θ 1 |x) =<br />

p(x|θ 1 )p(θ = θ 1 )<br />

p(x|θ 1 )p(θ = θ 1 ) + p(x|θ 2 )p(θ = θ 2 )<br />

= 1 − p(θ = θ 2 |x).<br />

2


Now the posterior risk under action a 1 is<br />

L(θ 1 , a 1 ) p(θ = θ 1 |x) + L(θ 2 , a 1 ) p(θ = θ 2 |x) = 10 p(θ = θ 2 |x)<br />

while the posterior risk under action a 2 is<br />

L(θ 1 , a 2 ) p(θ = θ 1 |x) + L(θ 2 , a 2 ) p(θ = θ 2 |x) = 5 p(θ = θ 1 |x).<br />

Hence, a 1 is preferred to a 2 if and only if<br />

This implies<br />

10 p(θ = θ 2 |x) ≤ 5 p(θ = θ 1 |x).<br />

p(x|θ 2 ) p(θ = θ 2 ) ≤ (1/2) p(x|θ 1 ) p(θ = θ 1 ) =⇒<br />

1<br />

(1 − η) √ e −1/2(x−1)2 ≤ η 1<br />

√ e −1/2(x)2 =⇒<br />

2π 2 2π<br />

e x−1/2 η<br />

≤<br />

2(1 − η) =⇒<br />

x ≤ 1/2 + log<br />

η<br />

2(1 − η) .<br />

3

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!