06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Some Properties <strong>of</strong> the Histogram Estimator<br />

8.5 Nonparametric Estimation <strong>of</strong> PDFs 579<br />

The histogram estimator, being a step function, is discontinuous at cell boundaries,<br />

and it is zero outside <strong>of</strong> a finite range. It is sensitive both to the bin size<br />

and to the choice <strong>of</strong> the origin.<br />

An important advantage <strong>of</strong> the histogram estimator is its simplicity, both<br />

for computations and for analysis. In addition to its simplicity, as we have<br />

seen, it has two other desirable global properties:<br />

• It is a bona fide density estimator.<br />

• It is the unique maximum likelihood estimator confined to the subspace<br />

<strong>of</strong> functions <strong>of</strong> the form<br />

and where g(t) ≥ 0 and <br />

g(t) = ck, for t ∈ Tk,<br />

∪kTk<br />

Pointwise and Binwise Properties<br />

= 0, otherwise,<br />

g(t)dt = 1.<br />

Properties <strong>of</strong> the histogram vary from bin to bin. From equation (8.33), the<br />

expectation <strong>of</strong> the histogram estimator at the point y in bin Tk is<br />

where<br />

E(pH(y)) = pk<br />

, (8.35)<br />

<br />

pk =<br />

Tk<br />

vk<br />

p(t)dt (8.36)<br />

is the probability content <strong>of</strong> the k th bin.<br />

Some pointwise properties <strong>of</strong> the histogram estimator are the following:<br />

• The bias <strong>of</strong> the histogram at the point y within the k th bin is<br />

pk<br />

vk<br />

− p(y). (8.37)<br />

Note that the bias is different from bin to bin, even if the bins are <strong>of</strong><br />

constant size. The bias tends to decrease as the bin size decreases. We can<br />

bound the bias if we assume a regularity condition on p. If there exists γ<br />

such that for any y1 = y2 in an interval<br />

|p(y1) − p(y2)| < γy1 − y2,<br />

we say that p is Lipschitz-continuous on the interval, and for such a density,<br />

for any ξk in the k th bin, we have<br />

|Bias(pH(y))| = |p(ξk) − p(y)|<br />

≤ γkξk − y<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle<br />

≤ γkvk. (8.38)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!