06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

3.2 Statistical Inference: Approaches and Methods 247<br />

Example 3.14 least squares in a linear model<br />

Consider the linear model (3.7)<br />

Y = Xβ + E,<br />

where Y is the random variable that is observable, in the least squares setup<br />

<strong>of</strong> equations (3.66) and (3.67) we have<br />

Sn(β ; y, X) = y − Xβ2. (3.69)<br />

In the case <strong>of</strong> the linear model, we have the estimating equation<br />

sn(β ; y, X) = X T y − X T Xβ = 0. (3.70)<br />

This system <strong>of</strong> equations is called the “normal equations”.<br />

For the estimand g(β) = l T β for some fixed l, the least squares estimator<br />

is l T (X T X) − X T Y , where M − denotes a generalized inverse <strong>of</strong> a matrix M.<br />

See page 420 and the following pages for a more thorough discussion <strong>of</strong> the<br />

linear model in this example.<br />

Example 3.14 illustrates a very simple and common application <strong>of</strong> estimation<br />

by fitting expected values; the expected values are those <strong>of</strong> the observable<br />

random variable. The next example is a somewhat less common situation <strong>of</strong><br />

defining which expected values to focus on.<br />

Example 3.15 estimation in a stable family; the empirical CF<br />

Most members <strong>of</strong> the stable family <strong>of</strong> distributions are quite complicated.<br />

In general, there is no closed form for the CDF <strong>of</strong> the PDF, and none <strong>of</strong><br />

the moments exist or else are infinite. The family <strong>of</strong> distributions is usually<br />

specified by means <strong>of</strong> the characteristic function (see equation (2.30)),<br />

ϕ(t) = exp (iµt − |σt| α (1 − iβ sign(t)ω(α, t))).<br />

Because the characteristic function is an expected value, its sample analogue,<br />

that is, the empirical characteristic function can be formed easily from a sample,<br />

x1, . . ., xn:<br />

ϕn(t) = 1<br />

n<br />

e<br />

n<br />

itxi . (3.71)<br />

The empirical characteristic function can be computed at any point t.<br />

The expected values would be fit by minimizing<br />

i=1<br />

Sn(x, r, µ, σ, α,β) = ϕn(t) − ϕ(t)r<br />

(3.72)<br />

for given r at various values <strong>of</strong> t.<br />

While this is a well-defined problem (for some given values <strong>of</strong> t) and the<br />

resulting estimators are strongly consistent (for the same reason that estimators<br />

based on the ECDF are strongly consistent), there are many practical<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!