06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

838 Appendix B. Useful Inequalities in Probability<br />

• V(Y ) and V (E(Y |X)), e.g., Rao-Blackwell<br />

Any special case <strong>of</strong> these involves an appropriate definition <strong>of</strong> A or f (e.g.,<br />

nonnegative, convex, etc.)<br />

A more general case <strong>of</strong> the inequalities is to replace distributions, and<br />

hence expected values, by conditioning on a sub-σ-field, A.<br />

For each type <strong>of</strong> inequality there is an essentially straightforward method<br />

<strong>of</strong> pro<strong>of</strong>.<br />

Some <strong>of</strong> these inequalities involve absolute values <strong>of</strong> the random variable.<br />

To work with these inequalities, it is useful to recall the triangle inequality<br />

for the absolute value <strong>of</strong> real numbers:<br />

|x + y| ≤ |x| + |y|. (B.2)<br />

We can prove this merely by considering all four cases for the signs <strong>of</strong> x and<br />

y.<br />

This inequality generalizes immediately to | xi| ≤ |xi|.<br />

Expectations <strong>of</strong> absolute values <strong>of</strong> functions <strong>of</strong> random variables are functions<br />

<strong>of</strong> norms. (A norm is a function <strong>of</strong> x that (1) is positive unless x = 0<br />

a.e., that (2) is equivariant to scalar multiplication, and that (3) satisfies the<br />

triangle inequality.) The important form (E(|X| p )) 1/p for 1 ≤ p is an Lp norm,<br />

Xp. Some <strong>of</strong> the inequalities given below involving expectations <strong>of</strong> absolute<br />

values <strong>of</strong> random variables are essentially triangle inequalities and their truth<br />

establishes the expectation as a norm.<br />

Some <strong>of</strong> the expectations discussed below are recognizable as familiar<br />

norms over vector spaces. For example, the expectation in Minkowski’s inequality<br />

is essentially the Lp norm <strong>of</strong> a vector, which is defined for an n-vector<br />

x in a finite-dimensional vector space as xp ≡ ( |xi| p ) 1/p . Minkowski’s inequality<br />

in this case is x + yp ≤ xp + yp. For p = 1, this is the triangle<br />

inequality for absolute values given above.<br />

B.2 Pr(X ∈ Ai) and Pr(X ∈ ∪Ai) or Pr(X ∈ ∩Ai)<br />

These inequalities are <strong>of</strong>ten useful in working with order statistics and in tests<br />

<strong>of</strong> multiple hypotheses. Instead <strong>of</strong> Pr(X ∈ Ai), we may write Pr(Ai).<br />

Theorem B.1 (Bonferroni’s inequality)<br />

Given events A1, . . ., An, we have<br />

Pr(∩Ai) ≥ Pr(Ai) − n + 1. (B.3)<br />

Pro<strong>of</strong>. We will use induction. For n = 1, we have Pr(A1) ≥ Pr(A1),<br />

and for n = 2, we have Pr(A1 ∩ A2) = Pr(A1) + Pr(A2) − Pr(A1 ∪ A2) ≥<br />

Pr(A1) + Pr(A2) − 1.<br />

Now assume Pr(∩k i=1Ai) ≥ k i=1 AiPr(Ai) − k + 1, and consider ˜ k =<br />

k + 1. We have (from the n = 2 case) Pr(∩k i=1Ai ∩ Ak+1) ≥ Pr(∩k i=1Ai) +<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!