08.03.2014 Views

Variance Estimation for the General Regression Estimator

Variance Estimation for the General Regression Estimator

Variance Estimation for the General Regression Estimator

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

6<br />

⎧<br />

⎨<br />

⎪<br />

⎢<br />

⎩⎣<br />

⎪<br />

design variance,<br />

ˆ<br />

ˆ<br />

( ) ( ) 2<br />

E E ⎡ Y Y E E Y Y ⎤<br />

− − −<br />

M π G M π G<br />

useful <strong>for</strong> both var<br />

ˆ<br />

M ( Y G − Y)<br />

and var<br />

ˆ<br />

π ( Y G)<br />

⎥⎦<br />

⎫⎪<br />

⎬. Ra<strong>the</strong>r we seek estimators that are<br />

⎪⎭<br />

. The arguments given here are largely heuristic<br />

ones used to motivate <strong>the</strong> <strong>for</strong>ms of <strong>the</strong> variance estimators. Additional, <strong>for</strong>mal conditions such<br />

as those found in Royall and Cumberland (1978) or Yung and Rao (2000) are needed <strong>for</strong> modelbased<br />

and design-based consistency and approximate unbiasedness.<br />

First, consider estimation of <strong>the</strong> approximate model-variance given in (1.5). In <strong>the</strong><br />

following development, we assume that, as N and n become large,<br />

(i) Nmax( ) O( n)<br />

i<br />

π = and<br />

i<br />

(ii) A πs N converges to a matrix of constants, A o.<br />

A residual associated with sample unit i is<br />

r = Y − Yˆ<br />

where Y ˆ =xB. ′ ˆ The vector of<br />

i i i<br />

i<br />

i<br />

predicted values <strong>for</strong> <strong>the</strong> sample units can be written as<br />

Yˆs = HY s s<br />

(2.1)<br />

where<br />

1 1 1<br />

s = s π − s ′ s s − s<br />

−<br />

=∑ ∈<br />

H XA XV Π . The predicted value <strong>for</strong> an individual unit is Y ˆi hY ij j<br />

j s<br />

1<br />

where hij ′<br />

−<br />

i πs j ( vjπ<br />

j)<br />

=xA x is <strong>the</strong> (ij) th element of H s. The matrix<br />

H s is <strong>the</strong> analog to <strong>the</strong><br />

usual hat matrix (Belsley, Kuh, and Welsch 1980) from standard regression analysis. The<br />

diagonal elements of <strong>the</strong> hat matrix are known as leverages and are a measure of <strong>the</strong> effect that a<br />

unit has on its own predicted value. Notice that <strong>the</strong> inverses of <strong>the</strong> selection probabilities are<br />

involved in (2.1), although <strong>the</strong>se would have no role in purely model-based analysis.<br />

The following lemma, which is a variation of some results in Lemma 5.3.1 of (Valliant,<br />

Dorfman, and Royall 2000), gives some properties of <strong>the</strong> leverages and <strong>the</strong> hat matrix.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!