06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

8.5 Nonparametric Estimation <strong>of</strong> PDFs 575<br />

In this chapter, we consider several nonparametric estimators <strong>of</strong> a density;<br />

that is, estimators <strong>of</strong> a general nonnegative function that integrates to 1<br />

and for which we make no assumptions about a functional form other than,<br />

perhaps, smoothness.<br />

It seems reasonable that we require the density estimate to have the characteristic<br />

properties <strong>of</strong> a density:<br />

• p(y) ≥ 0 for all y;<br />

• <br />

IR d p(y)dy = 1.<br />

A probability density estimator that is nonnegative and integrates to 1 is<br />

called a bona fide estimator.<br />

Rosenblatt has shown that no unbiased bona fide estimator can exist for<br />

all continuous p. Rather than requiring an unbiased estimator that cannot be<br />

a bona fide estimator, we generally seek a bona fide estimator with small mean<br />

squared error or a sequence <strong>of</strong> bona fide estimators pn that are asymptotically<br />

unbiased; that is,<br />

The Likelihood Function<br />

Ep(pn(y)) → p(y) for all y ∈ IR d as n → ∞.<br />

Suppose that we have a random sample, y1, . . ., yn, from a population with<br />

density p. Treating the density p as a variable, we write the likelihood functional<br />

as<br />

n<br />

L(p; y1, . . ., yn) = p(yi).<br />

The maximum likelihood method <strong>of</strong> estimation obviously cannot be used directly<br />

because this functional is unbounded in p. We may, however, seek an<br />

estimator that maximizes some modification <strong>of</strong> the likelihood. There are two<br />

reasonable ways to approach this problem. One is to restrict the domain <strong>of</strong><br />

the optimization problem. This is called restricted maximum likelihood. The<br />

other is to regularize the estimator by adding a penalty term to the functional<br />

to be optimized. This is called penalized maximum likelihood.<br />

We may seek to maximize the likelihood functional subject to the constraint<br />

that p be a bona fide density. If we put no further restrictions on<br />

the function p, however, infinite Dirac spikes at each observation give an unbounded<br />

likelihood, so a maximum likelihood estimator cannot exist, subject<br />

only to the restriction to the bona fide class. An additional restriction that<br />

p be Lebesgue-integrable over some domain D (that is, p ∈ L 1 (D)) does not<br />

resolve the problem because we can construct sequences <strong>of</strong> finite spikes at<br />

each observation that grow without bound.<br />

We therefore must restrict the class further. Consider a finite dimensional<br />

class, such as the class <strong>of</strong> step functions that are bona fide density estimators.<br />

We assume that the sizes <strong>of</strong> the regions over which the step function is constant<br />

are greater than 0.<br />

i=1<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!