08.10.2016 Views

Foundations of Data Science

2dLYwbK

2dLYwbK

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

The above theorem tells us that the maximum <strong>of</strong> x T Ax subject to the constraint that<br />

|x| 2 = 1 is λ 1 . Consider the problem <strong>of</strong> maximizing x T Ax subject to the additional restriction<br />

that x is orthogonal to the first eigenvector. This is equivalent to maximizing<br />

y t P t AP y subject to y being orthogonal to (1,0,. . . ,0), i.e. the first component <strong>of</strong> y being<br />

0. This maximum is clearly λ 2 and occurs for y = (0, 1, 0, . . . , 0). The corresponding x is<br />

the second column <strong>of</strong> P or the second eigenvector <strong>of</strong> A.<br />

Similarly the maximum <strong>of</strong> x T Ax for p 1 T x = p 2 T x = · · · p s T x = 0 is λ s+1 and is<br />

obtained for x = p s+1 .<br />

12.7.4 Eigenvalues <strong>of</strong> the Sum <strong>of</strong> Two Symmetric Matrices<br />

The min max theorem is useful in proving many other results. The following theorem<br />

shows how adding a matrix B to a matrix A changes the eigenvalues <strong>of</strong> A. The theorem<br />

is useful for determining the effect <strong>of</strong> a small perturbation on the eigenvalues <strong>of</strong> A.<br />

Theorem 12.13 Let A and B be n × n symmetric matrices. Let C=A+B. Let α i , β i ,<br />

and γ i denote the eigenvalues <strong>of</strong> A, B, and C respectively, where α 1 ≥ α 2 ≥ . . . α n and<br />

similarly for β i , γ i . Then α s + β 1 ≥ γ s ≥ α s + β n .<br />

Pro<strong>of</strong>: By the min max theorem we have<br />

α s =<br />

min<br />

max<br />

r 1 ,...,r s−1<br />

x<br />

r i ⊥x<br />

(x T Ax).<br />

Suppose r 1 , r 2 , . . . , r s−1 attain the minimum in the expression. Then using the min max<br />

theorem on C,<br />

(<br />

γ s ≤ max x T (A + B)x )<br />

x⊥r 1 ,r 2 ,...r s−1<br />

Therefore, γ s ≤ α s + β 1 .<br />

≤<br />

max (x T Ax) + max (x T Bx)<br />

x⊥r 1 ,r 2 ,...r s−1<br />

x⊥r 1 ,r 2 ,...r s−1<br />

≤ α s + max<br />

x (xT Bx) ≤ α s + β 1 .<br />

An application <strong>of</strong> the result to A = C + (−B), gives α s ≤ γ s − β n . The eigenvalues<br />

<strong>of</strong> -B are minus the eigenvalues <strong>of</strong> B and thus −β n is the largest eigenvalue. Hence<br />

γ s ≥ α s + β n and combining inequalities yields α s + β 1 ≥ γ s ≥ α s + β n .<br />

Lemma 12.14 Let A and B be n × n symmetric matrices. Let C=A+B. Let α i , β i ,<br />

and γ i denote the eigenvalues <strong>of</strong> A, B, and C respectively, where α 1 ≥ α 2 ≥ . . . α n and<br />

similarly for β i , γ i . Then γ r+s−1 ≤ α r + β s .<br />

411

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!