06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

312 3 Basic Statistical <strong>Theory</strong><br />

an unknown parameter. In that case, we can set a confidence interval based<br />

on the approximate distribution <strong>of</strong> g(Xn) as N(g(c), v/a 2 n).<br />

To speak <strong>of</strong> the asymptotic distribution <strong>of</strong> an(g(Xn) − g(c)) is clear; but<br />

to refer to “the” asymptotic distribution <strong>of</strong> g(Xn) is somewhat less so.<br />

Because it is the useful approximate distribution resulting from asymptotic<br />

expectations, we <strong>of</strong>ten say that “the asymptotic distribution” <strong>of</strong> g(Xn)<br />

is N(g(c), v/a 2 n). You should recognize that “the” in this statement is somewhat<br />

arbitrary. It might be better to call it “the asymptotically approximate<br />

distribution that I’m going to use in this application”.<br />

Again, we should distinguish “asymptotic” from “limiting”.<br />

In the example <strong>of</strong> the delta method above, it is likely that<br />

g(Xn) d → g(c);<br />

that is, g(Xn) converges in distribution to the constant g(c); or the limiting<br />

distribution <strong>of</strong> g(Xn) is degenerate at g(c). “The” asymptotic variance is 0.<br />

**** discuss expansion <strong>of</strong> statistical functionals *** refer to Serfling<br />

This would not be very useful in asymptotic inference. We therefore seek<br />

“an” asymptotic variance that is more useful. In asymptotic estimation using<br />

g(Xn), we begin with an expression <strong>of</strong> the form an(g(Xn) − g(c)) that has<br />

a limiting distribution <strong>of</strong> the desired form (usually that means such that the<br />

variance does not involve any unknown parameters and it does not involve n).<br />

If this distribution is in a location-scale family, then we make the appropriate<br />

linear transformation (which probably results in a variance that does involve<br />

n).<br />

We then <strong>of</strong>ten refer to this as the asymptotic distribution <strong>of</strong> g(Xn). Sometimes,<br />

as mentioned above, however, the limiting distribution <strong>of</strong> g(Xn) is<br />

degenerate.<br />

This is not to imply that asymptotic expectations are entirely arbitrary.<br />

Proposition 2.3 in MS2 shows that there is a certain uniqueness in the asymptotic<br />

expectation. This proposition involves three cases regarding whether the<br />

expectation <strong>of</strong> g(Xn) (without the an sequence) is 0. In the example above,<br />

we have a degenerate distribution, and hence the asymptotic expectation that<br />

defines the asymptotic variance is 0.<br />

3.8.4 Properties <strong>of</strong> Estimators <strong>of</strong> a Variance Matrix<br />

If the statistic is a vector, we need an estimator <strong>of</strong> the variance-covariance matrix.<br />

Because a variance-covariance matrix is positive definite, it is reasonable<br />

to consider only estimators that are positive definite a.s.<br />

We have defined what it means for such an estimator to be consistent<br />

(Definition 3.18 on page 305).<br />

Theorem 3.16<br />

conditions for the consistency <strong>of</strong> substitution estimators.<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!