08.10.2016 Views

Foundations of Data Science

2dLYwbK

2dLYwbK

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

The first integral is just the standard integral <strong>of</strong> the beta function and its value is s!(k−2−s)!<br />

(k−1)!<br />

.<br />

To bound the second integral, note that for z ∈ [1, µ], |z − 1| ≤ 1<br />

k−2 and<br />

z k−s−2 ≤ ( 1 + ( 1/(k − 2) )) k−s−2<br />

≤ e (k−s−2)/(k−2) ≤ e.<br />

(<br />

So, |E((x i − µ) s (k − 1)s!(k − 2 − s)! e(k − 1)<br />

1<br />

)| ≤ +<br />

(k − 1)! (k − 2) ≤ s!Var(y) s+1 k − 4 + e )<br />

≤ s!Var(x).<br />

3!<br />

Now, apply the first inequality <strong>of</strong> Theorem 12.5 with s <strong>of</strong> that theorem set to k − 2 or<br />

k − 3 whichever is even. Note that a = εE(x) ≤ √ 2nσ 2 (since ε ≤ 1/k 2 ). The present<br />

theorem follows by a calculation.<br />

12.7 Eigenvalues and Eigenvectors<br />

Let A be an n×n real matrix. The scalar λ is called an eigenvalue <strong>of</strong> A if there exists a<br />

nonzero vector x satisfying the equation Ax = λx. The vector x is called the eigenvector<br />

<strong>of</strong> A associated with λ. The set <strong>of</strong> all eigenvectors associated with a given eigenvalue form<br />

a subspace as seen from the fact that if Ax = λx and Ay = λy, then for any scalers c<br />

and d, A(cx + dy) = λ(cx + dy). The equation Ax = λx has a nontrivial solution only if<br />

det(A − λI) = 0. The equation det(A − λI) = 0 is called the characteristic equation and<br />

has n not necessarily distinct roots.<br />

Matrices A and B are similar if there is an invertible matrix P such that A = P −1 BP .<br />

Theorem 12.8 If A and B are similar, then they have the same eigenvalues.<br />

Pro<strong>of</strong>: Let A and B be similar matrices. Then there exists an invertible matrix P<br />

such that A = P −1 BP . For an eigenvector x <strong>of</strong> A with eigenvalue λ, Ax = λx, which<br />

implies P −1 BP x = λx or B(P x) = λ(P x). So, P x is an eigenvector <strong>of</strong> B with the same<br />

eigenvalue λ. Since the reverse also holds, the theorem follows.<br />

Even though two similar matrices, A and B, have the same eigenvalues, their eigenvectors<br />

are in general different.<br />

The matrix A is diagonalizable if A is similar to a diagonal matrix.<br />

Theorem 12.9 A is diagonalizable if and only if A has n linearly independent eigenvectors.<br />

Pro<strong>of</strong>:<br />

(only if) Assume A is diagonalizable. Then there exists an invertible matrix P<br />

and a diagonal matrix D such that D = P −1 AP . Thus, P D = AP . Let the diagonal<br />

elements <strong>of</strong> D be λ 1 , λ 2 , . . . , λ n and let p 1 , p 2 , . . . , p n be the columns <strong>of</strong> P . Then<br />

AP = [Ap 1 , Ap 2 , . . . , Ap n ] and P D = [λ 1 p 1 , λ 2 p 2 , . . . , λ n p n ] . Hence Ap i = λ i p i . That<br />

405

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!