PDF of Lecture Notes - School of Mathematical Sciences
PDF of Lecture Notes - School of Mathematical Sciences
PDF of Lecture Notes - School of Mathematical Sciences
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
2. STATISTICAL INFERENCE<br />
I(λ) =<br />
( ) ∂ 2 l<br />
−E<br />
∂λ 2<br />
=⇒ I(λ) =<br />
( ) n ¯X<br />
E<br />
λ 2<br />
= n E( ¯X)<br />
λ2 = nλ<br />
λ 2<br />
= n λ .<br />
(3) Finally, observe that Var( ¯X) = λ n = 1<br />
I(λ) .<br />
By Theorem 2.2.2, any unbiased estimator T for λ must have<br />
Var(T ) ≥ 1<br />
I(λ) = λ n<br />
=⇒ ¯X is a MVUE.<br />
Theorem. 2.2.3<br />
The unbiased estimator T (x) can achieve the Cramer-Rao Lower Bound only if the<br />
joint <strong>PDF</strong>/probability function has the form:<br />
f(x; θ) = exp{A(θ)T (x) + B(θ) + h(x)},<br />
where A, B are functions such that θ = −B′ (θ)<br />
, and h is some function <strong>of</strong> x.<br />
A ′ (θ)<br />
Pro<strong>of</strong>. Recall from the pro<strong>of</strong> <strong>of</strong> Theorem 2.2.2 that the bound arises from the inequality<br />
where U(θ; x) = ∂l<br />
∂θ .<br />
Cor 2 (T (X), U(θ; X)) ≤ 1,<br />
Moreover, it is easy to see that the Cramer-Rao Lower Bound (CRLB) can be achieved<br />
only when<br />
Cor 2 {T (X), U(θ; X)} = 1<br />
87