21.06.2014 Views

Numerical Methods Contents - SAM

Numerical Methods Contents - SAM

Numerical Methods Contents - SAM

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

9 ylabel ( ’ { \ b f c o n d i t i o n number } ’ , ’ f o n t s i z e ’ ,14) ;<br />

10 t i t l e ( ’ Condition numbers o f row t r a n s f o r m a t i o n matrices ’ ) ;<br />

11 legend ( ’2−norm ’ , ’maximum norm ’ , ’1−norm ’ , ’ l o c a t i o n ’ , ’ southeast ’ ) ;<br />

12 p r i n t −depsc2 ’ . . / PICTURES/ rowtrfcond . eps ’ ;<br />

all rows/columns are pairwise orthogonal (w.r.t. Euclidean inner product),<br />

| detQ| = 1, and all eigenvalues ∈ {z ∈ C: |z| = 1}.<br />

‖QA‖ 2 = ‖A‖ 2 for any matrix A ∈ K n,m<br />

Observation: cond(T(µ)) large for large µ<br />

As explained in Sect. 2.2, Gaussian (forward) elimination can be viewed as successive multiplication<br />

with elimination matrices. If an elimination matrix has a large condition number, then small relative errors<br />

in the entries of intermediate matrices caused by earlier roundoff errors can experience massive<br />

amplification and, thus, spoil all further steps (➤ loss of numerical stability, Def. 2.5.5).<br />

Drawbacks of LU-factorization:<br />

often pivoting required (➞ destroys structure, Ex. 2.6.21, leads to fill-in)<br />

Possible (theoretical) instability of partial pivoting → Ex. 2.5.2<br />

Therefore, the entries of elimination matrices should be kept small, and this is the main rationale<br />

behind (partial) pivoting (→ Sect. 2.3), which ensures that multipliers have modulus ≤ 1 throughout<br />

forward elimination.<br />

Ôº½¿ ¾º<br />

✸<br />

Stability problems of Gaussian elimination without pivoting are due to the fact that row transformations<br />

can convert well-conditioned matrices to ill-conditioned matrices, cf. Ex. 2.5.2<br />

Which bijective row transformations preserve the Euclidean condition number of a matrix ?<br />

Ôº½ ¾º<br />

➣ transformations hat preserve the Euclidean norm of a vector !<br />

Recall from linear algebra:<br />

Investigate algorithms that use orthogonal/unitary row transformations to convert a matrix to<br />

upper triangular form.<br />

Definition 2.8.1 (Unitary and orthogonal matrices).<br />

• Q ∈ K n,n , n ∈ N, is unitary, if Q −1 = Q H .<br />

• Q ∈ R n,n , n ∈ N, is orthogonal, if Q −1 = Q T .<br />

✬<br />

Theorem 2.8.2 (Criteria for Unitarity).<br />

✩<br />

Goal:<br />

find unitary row transformation rendering certain matrix elements zero.<br />

⎛ ⎞ ⎛ ⎞<br />

Q⎜<br />

⎟<br />

⎝ ⎠ = ⎜<br />

⎝ 0<br />

⎟<br />

⎠ with QH = Q −1 .<br />

Q ∈ C n,n unitary ⇔ ‖Qx‖ 2 = ‖x‖ 2 ∀x ∈ K n .<br />

✫<br />

Q unitary ⇒ cond(Q) = 1<br />

If Q ∈ K n,n unitary, then<br />

(2.8.1)<br />

➤<br />

unitary transformations enhance (numerical) stability<br />

all rows/columns (regarded as vectors ∈ K n ) have Euclidean norm = 1,<br />

✪<br />

Ôº½ ¾º<br />

This “annihilation of column entries” is the key operation in Gaussian forward elimination, where it<br />

is achieved by means of non-unitary row transformations, see Sect. 2.2. Now we want to find a<br />

counterpart of Gaussian elimination based on unitary row transformations on behalf of numerical<br />

stability.<br />

In 2D: two possible orthogonal transformations make 2nd component of a ∈ R 2 vanish:<br />

Ôº½ ¾º

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!