15.11.2014 Views

Takashi Yamano

Takashi Yamano

Takashi Yamano

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Lecture Notes on Advanced Econometrics<br />

Lecture 6: OLS Asymptotic Properties<br />

<strong>Takashi</strong> <strong>Yamano</strong><br />

Fall Semester 2004<br />

Reviewing: Finite-Sample Properties of Least Squares<br />

In the previous section, we have studied finite-sample properties of OLS estimators.<br />

Under the finite-sample properties, the OLS estimators are unbiased and the error terms<br />

are normally distributed even when sample sizes are small. Here is a quick review:<br />

Assumptions:<br />

E 1 (Linear in parameters): y = X % + u<br />

E 2 (Zero conditional mean): E(u|X) = 0<br />

E 3 (No perfect collinearity): X has rank k.<br />

Under these assumptions, the OLS estimators are unbiased because:<br />

Thus,<br />

Βˆ = (XtX) -1 Xty<br />

(because of E3 (XtX) -1 exists)<br />

Βˆ = (XtX) -1 Xt( X % + u) (from E1)<br />

= % + (XtX) -1 Xtu<br />

E( Β ˆ ) = % + (XtX)<br />

XtE(u|X) (from E2)<br />

E(Β ˆ ) = %<br />

(unbiased)<br />

To find the variance of the OLS estimators, we added one more assumptioin.<br />

E 4 (Homoskedasticity): Var(u i |X)=F 2 and Cov(u i ,u j )=0, thus Var(u|X)= F 2 I<br />

Therefore,<br />

Var( Β ˆ ) = E [(<br />

Βˆ<br />

−Β)(<br />

Βˆ<br />

− Β)<br />

′ | X ]<br />

− 1<br />

−1<br />

= ( X ′X ) X ′ E[<br />

u u′<br />

| X ] X ( X ′ X )<br />

− 1 2<br />

−1<br />

= ( X ′X ) X ′(<br />

σ I)<br />

X ( X ′ X ) (from E4)<br />

= F 2 (XtX) -1<br />

Finally, for us to be able to conduct statistical tests, we need to know the distribution of<br />

the OLS estimators, then we added an assumption about the distribution of the error term:<br />

E 5 (Normality of Errors): u n x 1 ~ N (0 n x 1, F 2 I n x n )<br />

1


Under the assumptions E1-5, each of following estimators has an exact distribution.<br />

ˆ 1<br />

2 −<br />

Β ~ N[<br />

Β,<br />

σ ( X ′ X ) ]<br />

(Normal distribution)<br />

This holds for<br />

ˆ β − β<br />

k k<br />

t<br />

k<br />

=<br />

(t distribution)<br />

s 2 S<br />

kk<br />

F<br />

( SSR<br />

≡<br />

SSR<br />

r<br />

ur<br />

− SSR<br />

ur<br />

) / q<br />

/ ( n − k −1)<br />

(F distribution)<br />

However, these are strong assumptions and can be relaxed easily by using asymptotic<br />

theory. Therefore, in this lecture, we study the asymptotic properties or large sample<br />

properties of the OLS estimators. Under the asymptotic properties, the properties of the<br />

OLS estimators depend on the sample size. In short, we can show that the OLS<br />

estimators could be biased with a small sample size but consistent with a sufficiently<br />

large sample size.<br />

Consistency (instead of unbiasedness)<br />

First, we need to define consistency. Suppose W n is an estimator of θ on a sample of Y 1 ,<br />

Y 2 , …, Y n of size n. Then, W n is a consistent estimator of θ if for every e > 0,<br />

P(|W n - θ| > e) 0 as n <br />

This says that the probability that the absolute difference between W n and θ being larger<br />

than e goes to zero as n gets bigger. Which means that this probability could be non-zero<br />

while n is not large. For instance, let’s say that we are interested in finding the average<br />

income of American people and take small samples randomly. Let’s assume that the<br />

small samples include Bill Gates by chance. The sample mean income is way over the<br />

population average. Thus, when sample sizes are small, the probability that the<br />

difference between the sample and population averages is larger than e, which is any<br />

positive number, can be non-zero. However, the difference between the sample and<br />

population averages would be smaller as the sample size gets bigger (as long as the<br />

sampling is properly done). As a result, as the sample size goes to infinity, the<br />

probability that the difference between the two averages is bigger than e (no matter how<br />

small e is) becomes zero.<br />

In other words, we say that θ is the probability limit of W n :<br />

plim (W n ) = θ.<br />

Under the finite-sample properties, we say that W n is unbiased, E(W n ) =θ. Under the<br />

asymptotic properties, we say that W n is consistent because W n converges to θ as n gets<br />

larger.<br />

2


The OLS estimators<br />

From previous lectures, we know the OLS estimators can be written as<br />

Βˆ = (XtX) -1 Xty<br />

Βˆ = B + (XtX) -1 Xtu<br />

In the matrix form, we can examine the probability limit of OLS<br />

Here, we assume that<br />

−1<br />

⎛ 1 ⎞<br />

p limΒˆ<br />

= Β ˆ + ⎜ X ′ X ⎟<br />

⎝ n ⎠<br />

⎛ 1 ⎞<br />

p lim⎜<br />

X ′ u⎟<br />

⎝ n ⎠<br />

1<br />

p lim X ′ X =<br />

n<br />

Q<br />

and assume that Q -1 exists. From E2, we have<br />

⎛ 1 ⎞<br />

p lim ⎜ X ′u⎟<br />

= 0 .<br />

⎝ n ⎠<br />

Thus,<br />

lim ˆ<br />

−1<br />

p Β = Β + Q 0<br />

Thus, we have shown that the OLS estimator is consistent.<br />

Nest, we focus on the asymmetric inference of the OLS estimator. The asymptotic<br />

distribution of the OLS estimator is derived by writing<br />

−1<br />

⎛ ⎞<br />

Βˆ<br />

= Β ˆ 1<br />

+ ⎜ X ′ X ⎟<br />

⎝ n ⎠<br />

⎛ 1<br />

n ( Βˆ<br />

−Β)<br />

= ⎜ X ′ X<br />

⎝ n<br />

⎛ 1 ⎞<br />

⎜ X ′ u⎟<br />

⎝ n ⎠<br />

⎞<br />

⎟<br />

⎠<br />

−1<br />

⎛<br />

⎜<br />

⎝<br />

1 ⎞<br />

X ′ u ⎟<br />

n ⎠<br />

The probability limit of n ( Βˆ<br />

− Β)<br />

goes to zero because the consistency of Βˆ . The<br />

variance of n ( Βˆ<br />

−Β)<br />

is<br />

−1<br />

′<br />

−1<br />

ˆ ˆ ⎛ 1 ⎞ ⎛ 1 ⎞⎛<br />

1 ⎞ ⎛ 1 ⎞<br />

n ( Β − Β)<br />

⋅ n(<br />

Β −Β)<br />

′= ⎜ X ′ X ⎟ ⎜ X ′ u ⎟⎜<br />

X ′ u ⎟ ⎜ X ′ X ⎟<br />

⎝ n ⎠ ⎝ n ⎠⎝<br />

n ⎠ ⎝ n ⎠<br />

−1<br />

⎛ 1 ⎞ ⎛ 1 ⎞⎛<br />

1 ⎞<br />

= ⎜ X ′ X ⎟ ⎜ X ′ u u′<br />

X ⎟⎜<br />

X ′ X ⎟<br />

⎝ n ⎠ ⎝ n ⎠⎝<br />

n ⎠<br />

−1<br />

From E4, the probability limit of u u′ goes to F 2 I. And also we assumed plim of<br />

is Q. Thus,<br />

1<br />

X ′ X<br />

n<br />

3


⎛ ⎞<br />

= Q ⎜<br />

σ X ′ X<br />

⎟Q<br />

⎝ n ⎠<br />

2 −1<br />

−1<br />

= σ Q QQ<br />

2<br />

−1 −1<br />

2 −1<br />

= σ Q<br />

Therefore, the asymptotic distribution of the OLS estimator is<br />

n ( Βˆ<br />

− Β)<br />

~<br />

a<br />

2 −1<br />

N[0,<br />

σ Q ].<br />

From this, we can treat the OLS estimator, Βˆ , as if it is approximately normally<br />

distributed with mean Β and variance-covariance matrix σ<br />

2 Q −1 / n .<br />

Example 6-1: Consistency of OLS Estimators in Bivariate Linear Estimation<br />

A bivariate model:<br />

y = $ 0 + $ 1 x + u and<br />

ˆβ =<br />

1<br />

n<br />

∑<br />

i=<br />

1<br />

n<br />

∑<br />

i=<br />

1<br />

( x<br />

i<br />

( x<br />

− x)<br />

y<br />

i<br />

− x)<br />

To examine the biasedness of the OLS estimator, we take the expectation<br />

n<br />

⎛<br />

⎞<br />

⎜ ∑(<br />

xi<br />

− x)<br />

ui<br />

⎟<br />

ˆ = + ⎜ i=<br />

1<br />

E(<br />

β<br />

⎟<br />

1)<br />

β1<br />

E<br />

⎜ n ⎟<br />

2<br />

⎜ ∑(<br />

xi<br />

− x)<br />

⎟<br />

⎝ i=<br />

1 ⎠<br />

Under the assumption of zero conditional mean (SLR 3: E(u|x) = 0), we can separate the<br />

expectation of x and u:<br />

n<br />

⎛<br />

⎞<br />

⎜ ∑(<br />

xi<br />

− x)<br />

E(<br />

ui<br />

) ⎟<br />

ˆ = + ⎜ i=<br />

1<br />

E(<br />

β ⎟<br />

1)<br />

β1<br />

.<br />

⎜ n<br />

⎟<br />

2<br />

⎜ ∑(<br />

xi<br />

− x)<br />

⎟<br />

⎝ i=<br />

1<br />

⎠<br />

Thus we need the SLR 3 to show the OLS estimator is unbiased.<br />

Now, suppose we have a violation of SLR 3 and can not show the unbiasedness of the<br />

OLS estimator. We consider a consistency of the OLS estimator.<br />

i<br />

2<br />

p lim ˆ β = p lim β<br />

1<br />

1<br />

+<br />

p lim<br />

⎛<br />

⎜<br />

⎜<br />

⎜<br />

⎜<br />

⎝<br />

n<br />

∑<br />

i=<br />

1<br />

n<br />

∑<br />

i=<br />

1<br />

( x<br />

i<br />

( x<br />

− x)<br />

u<br />

i<br />

−<br />

x)<br />

2<br />

i<br />

⎞<br />

⎟<br />

⎟<br />

⎟<br />

⎟<br />

⎠<br />

4


n<br />

⎡1<br />

⎤<br />

p lim⎢<br />

∑(<br />

xi<br />

− x)<br />

ui<br />

⎥<br />

ˆ = +<br />

⎣n<br />

i=<br />

1<br />

p lim β<br />

⎦<br />

1<br />

β1<br />

n<br />

⎡1<br />

2 ⎤<br />

p lim ⎢ ∑(<br />

xi<br />

− x)<br />

⎥<br />

⎣n<br />

i=<br />

1 ⎦<br />

ˆ cov( x,<br />

u)<br />

p lim β<br />

1<br />

= β1<br />

+<br />

var( x)<br />

p lim ˆ β1 = β1<br />

if cov( x,<br />

u)<br />

= 0<br />

Thus, as long as the covariance between x and u is zero, the OLS estimator of a bivariate<br />

model consistent.<br />

End of Example 6-1<br />

5

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!