06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

• monotone convergence:<br />

for 0 ≤ X1 ≤ X2 · · · a.s.<br />

• Fatou’s lemma:<br />

Xn<br />

1.5 Conditional Probability 113<br />

a.s.<br />

→ X ⇒ E(Xn|A) a.s.<br />

→ E(X|A). (1.236)<br />

0 ≤ Xn ∀ n ⇒ E(lim n inf Xn|A) ≤ lim n inf E(Xn|A) a.s.. (1.237)<br />

• dominated convergence:<br />

given a fixed Y with E(Y |A) < ∞,<br />

|Xn| ≤ Y ∀ n and Xn a.s.<br />

→ X ⇒ E(Xn|A) a.s.<br />

→ E(X|A). (1.238)<br />

Another useful fact is that if Y is A-measurable and |XY | and |X| are integrable<br />

(notice this latter is stronger than what is required to define E(X|A)),<br />

then<br />

E(XY |A) = Y E(X|A) a.s. (1.239)<br />

Some Useful Conditional Expectations<br />

There are some conditional expectations that arise <strong>of</strong>ten, and which we should<br />

immediately recognize. The simplest one is<br />

E E(Y |X) = E(Y ). (1.240)<br />

Note that the expectation operator is based on a probability distribution,<br />

and so anytime we see “E”, we need to ask “with respect to what probability<br />

distribution?” In notation such as that above, the distributions are implicit<br />

and all relate to the same probability space. The inner expectation on the left<br />

is with respect to the conditional distribution <strong>of</strong> Y given X, and so is a function<br />

<strong>of</strong> X. The outer expectation is with respect to the marginal distribution<br />

<strong>of</strong> X.<br />

Approaching this slightly differently, we consider a random variable Z that<br />

is a function <strong>of</strong> the random variables X and Y :<br />

Z = f(X, Y ).<br />

We have<br />

<br />

E(f(X, Y )) = EY EX|Y (f(X, Y )|Y ) <br />

= EX EY |X(f(X, Y )|X) . (1.241)<br />

Another useful conditional expectation relates adjusted variances to “total”<br />

variances:<br />

V(Y ) = V E(Y |X) + E V(Y |X) . (1.242)<br />

This is intuitive, although you should be able to prove it formally.The intuitive<br />

explanation is: the total variation in Y is the sum <strong>of</strong> the variation <strong>of</strong> its mean<br />

given X and its average variation about X (or given X). (Think <strong>of</strong> SST =<br />

SSR + SSE in regression analysis.)<br />

This equality implies the Rao-Blackwell inequality (drop the second term<br />

on the right).<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!