24.02.2013 Views

Optimality

Optimality

Optimality

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

for positive α, and for α = 0 as<br />

d0(g, f) = lim<br />

α→0 dα(g, f) =<br />

Asymptotics of the MDPDE 335<br />

�<br />

X<br />

g(x)log[g(x)/f(x;θ)]dx.<br />

Note that when α = 1, the DPD becomes<br />

�<br />

d1(g, f) = [g(x)−f(x;θ)] 2 dx.<br />

X<br />

Thus when α = 0 the DPD is the Kullback–Leibler divergence, for α = 1 it is the<br />

L 2 metric, and for 0 < α < 1 it is a smooth bridge between these two quantities.<br />

For α > 0 fixed, we make the fundamental assumption that there exists a unique<br />

point θ0∈ Θ corresponding to the density f closest to g according to the DPD. The<br />

point θ0 is defined as the target parameter. Let X1, . . . , Xn be a random sample<br />

from G. The minimum density power estimator (MDPDE) of θ0 is the point that<br />

minimizes the DPD between the probability mass function ˆgn associated with the<br />

empirical distribution of the sample and f. Replacing g by ˆgn in the definition of<br />

the DPD, dα(g, f), and eliminating terms that do not involve θ, the MDPDE ˆ θα,n<br />

is the value that minimizes<br />

�<br />

f 1+α (x; θ)dx−<br />

X<br />

�<br />

1 + 1<br />

�<br />

1<br />

α n<br />

n�<br />

f α (Xi;θ)<br />

over Θ. In this parametric framework the density f(·;θ0) can be interpreted as the<br />

projection of the true density g on the parametric family. If, on the other hand, g<br />

is a member of the family then g = f(·; θ0).<br />

Consider the score function and the information matrix of f(x;θ), S(x; θ) and<br />

i(x; θ), respectively. Define the p×p matrices Kα(θ) and Jα(θ) by<br />

�<br />

(2.2) Kα(θ) = S(x;θ)S t (x;θ)f 2α (x; θ)g(x)dx−Uα(θ)U t α(θ),<br />

where<br />

and<br />

(2.3)<br />

Jα(θ) =<br />

�<br />

X<br />

�<br />

Uα(θ) =<br />

X<br />

i=1<br />

S(x;θ)f α (x; θ)g(x)dx<br />

S(x;θ)S<br />

X<br />

t (x;θ)f 1+α (x; θ)dx<br />

�<br />

+<br />

X<br />

� i(x; θ)−αS(x;θ)S t (x; θ) � × [g(x)−f(x;θ)]f α (x; θ)dx.<br />

Basu et al. [1] show that, under certain regularity conditions, there exists a sequence<br />

ˆ θα,n of MDPDEs that is consistent for θ0 and the asymptotic distribution of<br />

√ n( ˆ θα,n−θ0) is multivariate normal with mean vector zero and variance-covariance<br />

matrix Jα(θ0) −1 Kα(θ0)Jα(θ0) −1 . The next section shows this result under assumptions<br />

different from those of Basu et al. [1].<br />

3. Asymptotic Behavior of the MDPDE<br />

Fix α > 0 and define the function m :X× Θ→R as<br />

(3.1)<br />

�<br />

m(x, θ) = 1 + 1<br />

�<br />

f<br />

α<br />

α �<br />

(x;θ)− f 1+α (x;θ)dx<br />

X

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!