Student Notes To Accompany MS4214: STATISTICAL INFERENCE
Student Notes To Accompany MS4214: STATISTICAL INFERENCE
Student Notes To Accompany MS4214: STATISTICAL INFERENCE
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
Also<br />
implies that<br />
S2(µ, v) = ∂ℓ<br />
∂v<br />
ˆv = 1<br />
n<br />
n�<br />
i=1<br />
n 1<br />
= − +<br />
2v 2v2 (xi − ˆµ) 2 = 1<br />
n<br />
n�<br />
(xi − µ) 2 = 0<br />
i=1<br />
n�<br />
(xi − ¯x) 2 . (2.5)<br />
Calculating second derivatives and multiplying by −1 gives that the information matrix<br />
I(µ, v) equals<br />
I(µ, v) =<br />
⎛<br />
⎜<br />
⎝<br />
1<br />
v 2<br />
i=1<br />
i=1<br />
n<br />
1<br />
v<br />
v2 n�<br />
(xi − µ)<br />
i=1<br />
n�<br />
(xi − µ) − n<br />
2v2 + 1<br />
v3 n�<br />
(xi − µ) 2<br />
⎞<br />
⎟<br />
⎠<br />
Hence I(ˆµ, ˆv) is given by : �<br />
n<br />
ˆv 0<br />
0 n<br />
2v2 �<br />
Clearly both diagonal terms are positive and the determinant is positive and so (ˆµ, ˆv)<br />
are, indeed, the MLEs of (µ, v).<br />
Go back to equation (2.4), and ¯ X ∼ N (µ, v/n). Clearly E( ¯ X) = µ (unbiased) and<br />
Var( ¯ X) = v/n, so ¯ X achieved the CRLB. Go back to equation (2.5). Then from lemma<br />
2.9 we have<br />
so that<br />
nˆv<br />
v ∼ χ2 n−1<br />
i=1<br />
� �<br />
nˆv<br />
E = n − 1<br />
v<br />
� �<br />
n − 1<br />
⇒ E(ˆv) = v<br />
n<br />
Instead, propose the (unbiased) estimator<br />
Observe that<br />
E(˜v) =<br />
˜v =<br />
n<br />
ˆv =<br />
n − 1<br />
� �<br />
n<br />
E(ˆv) =<br />
n − 1<br />
1<br />
n − 1<br />
n�<br />
(xi − ¯x) 2<br />
i=1<br />
� n<br />
n − 1<br />
� � n − 1<br />
and ˜v is unbiased as suggested. We can easily show that<br />
Hence<br />
Var(˜v) =<br />
2v 2<br />
(n − 1)<br />
n<br />
�<br />
v = v<br />
(2.6)<br />
eff(˜v) = 2v2 2v2 1<br />
÷ = 1 −<br />
n (n − 1) n<br />
Clearly ˜v is not efficient, but is asymptotically efficient. �<br />
29