06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

116 1 Probability <strong>Theory</strong><br />

Pro<strong>of</strong>. Exercise. Compare this with Theorem 1.13 on page 27, in which the<br />

corresponding solution is g0(X) = E(Y |a) = E(Y ).<br />

By the general definition <strong>of</strong> projection (Definition 0.0.9 on page 631), we<br />

see that conditional expectation can be viewed as a projection in a linear space<br />

defined by the square-integrable random variables over a given probability<br />

space and the inner product 〈Y, X〉 = E(Y X) and its induced norm. (In fact,<br />

some people define conditional expectation this way instead <strong>of</strong> the way we<br />

have in Definitions 1.44 and 1.45.)<br />

In regression applications in statistics using least squares, as we discuss on<br />

page 434, “Y ”, or the “predicted” Y given X, that is, E(Y |X) is the projection<br />

<strong>of</strong> Y onto X. For given fixed values <strong>of</strong> Y and X the predicted Y given X is<br />

the vector projection, in the sense <strong>of</strong> Definition 0.0.9.<br />

We now formally define projection for random variables in a manner analogous<br />

to Definition 0.0.9. Note that the random variable space is the range <strong>of</strong><br />

the functions in G in Theorem 1.63.<br />

Definition 1.46 (projection <strong>of</strong> a random variable onto a space <strong>of</strong> random variables)<br />

Let Y be a random variable and let X be a random variable space defined on<br />

the same probability space. A random variable Xp ∈ X such that<br />

is called a projection <strong>of</strong> Y onto X.<br />

E(Y − Xp2) ≤ E(Y − X2) ∀X ∈ X (1.245)<br />

The most interesting random variable spaces are linear spaces, and in the<br />

following we will assume that X is a linear space, and hence the norm arises<br />

from an inner product so that the terms in inequality (1.245) involve variances<br />

and covariances.<br />

*** existence, closure <strong>of</strong> space in second norm (see page 35).<br />

*** treat vector variables differently: E(Y − E(Y )2) is not the variance****<br />

make this distinction earlier<br />

When X is a linear space, we have the following result for projections.<br />

Theorem 1.64<br />

Let X be a linear space <strong>of</strong> random variables with finite second moments. Then<br />

Xp is a projection <strong>of</strong> Y onto X iff Xp ∈ X and<br />

Pro<strong>of</strong>.<br />

For any X, Xp ∈ X we have<br />

E (Y − Xp) T X = 0 ∀X ∈ X. (1.246)<br />

E((Y − X) T (Y − X)) = E((Y − Xp) T (Y − Xp))<br />

+2E((Y − Xp) T (Xp − X))<br />

+E((Xp − X) T (Xp − X))<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!