06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

3.3 The Decision <strong>Theory</strong> Approach to Statistical Inference 255<br />

uct <strong>of</strong> the vector with itself). If the expectation exists, least squares<br />

yields unbiasedness.<br />

Another example <strong>of</strong> this approach is least absolute values (LAV) estimation,<br />

in which the L1 norm <strong>of</strong> the vector <strong>of</strong> residuals is minimized.<br />

This yields median-unbiasedness.<br />

• fit an empirical probability distribution<br />

This approach is somewhat similar to fitting an ECDF, but in the case<br />

<strong>of</strong> PDFs, the criterion <strong>of</strong> closeness <strong>of</strong> the fit must be based on regions <strong>of</strong><br />

nonzero probability; that is, it can be based on divergence measures.<br />

• define and use a loss function<br />

(This is an approach based on “decision theory”, which we introduce formally<br />

in Section 3.3. The specific types <strong>of</strong> estimators that result from this<br />

approach are the subjects <strong>of</strong> several later chapters.)<br />

The loss function increases the more the estimator differs from the estimand,<br />

and then estimate g(θ) so as to minimize the expected value <strong>of</strong> the<br />

loss function (that is, the “risk”) at points <strong>of</strong> interest in the parameter<br />

space.<br />

– require unbiasedness and minimize the variance at all points in the<br />

parameter space (this is UMVU estimation, which we discuss more<br />

fully in Section 5.1)<br />

– require equivariance and minimize the risk at all points in the parameter<br />

space (this is MRE or MRI estimation, which we discuss more fully<br />

in Section 3.4)<br />

– minimize the maximum risk over the full parameter space<br />

– define an a priori averaging function for the parameter, use the observed<br />

data to update the averaging function and minimize the risk defined<br />

by the updated averaging function.<br />

3.3 The Decision <strong>Theory</strong> Approach to Statistical<br />

Inference<br />

3.3.1 Decisions, Losses, Risks, and Optimal Actions<br />

In the decision-theoretic approach to statistical inference, we call the inference<br />

a decision or an action, and we identify a cost or loss that depends on the<br />

decision and the true (but unknown) state <strong>of</strong> nature modeled by P ∈ P.<br />

(Instead <strong>of</strong> loss, we could use its opposite, which is called utility.)<br />

Our objective is to choose an action that minimizes the expected loss, or<br />

conversely maximizes the expected utility.<br />

We call the set <strong>of</strong> allowable actions or decisions the action space or decision<br />

space, and we denote it as A. We base the inference on the random variable<br />

X; hence, the decision is a mapping from X, the range <strong>of</strong> X, to A.<br />

In estimation problems, the action space may be a set <strong>of</strong> real numbers<br />

corresponding to a parameter space. In tests <strong>of</strong> statistical hypotheses, we may<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!