22.01.2014 Views

download searchable PDF of Circuit Design book - IEEE Global ...

download searchable PDF of Circuit Design book - IEEE Global ...

download searchable PDF of Circuit Design book - IEEE Global ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

134 Gradient Optimization<br />

Hessian, A-I Consider the function in (5.30) and its gradient in (5.33) when<br />

H=A. When this gradient is zero, dx in (5.34) corresponds to the location <strong>of</strong><br />

the minimum value. Substituting this in (5.30) yields<br />

(5.50)<br />

This is the amount by which F(p) exceeds its minimum value.<br />

The most popular optimization algorithm is the Fletcher-Powell method,<br />

which was first described by Davidon (1959). It is also known as the variable<br />

metric method, and it is worthwhile to observe that the latter name comes<br />

directly from (5.50). Davidon noted that the matrix A-I in (5.50) associates a<br />

squared length to any gradient. Therefore, he considered the inverse Hessian<br />

matrix for any nonlinear function as its metric or measure <strong>of</strong> standard length.<br />

His optimization method starts with a guess for H- 1 , usually the unit matrix<br />

U. This produces the steepest descent move according to (5.34). Following<br />

each iteration, Davidon "updates~' the estimate <strong>of</strong> the inverse Hessian, so that<br />

it is exact when a minimum'is finally found. In the interim, Davidon's metric<br />

varies, thus the name. There is also some statistical significance to the inverse<br />

Hessian for least-squares analysis (see Davidon, 1959).<br />

Variable metric methods in N dimensions require the storage <strong>of</strong> N(N + 1)/2<br />

elements <strong>of</strong> the symmetric, estimated inverse Hessian matrix; so they are not<br />

considered here for personal computers, although such methods converge<br />

rapidly near minima. There are many variable metric algorithms, but Dixon<br />

(1971) showed that most <strong>of</strong> these, which belong to a very large class <strong>of</strong><br />

algorithms, would produce equivalent results if the linear searches were<br />

absolutely accurate. Instead, another kind <strong>of</strong> conjugate gradient algorithm will<br />

be described, because it requires only 3N storage registers; it converges<br />

rapidly to good engineering accuracy, but lacks the ultimate convergence<br />

properties <strong>of</strong> variable metric methods. It is the Fletcher-Reeves conjugate<br />

gradient algorithm, which was originally suggested for very large problems<br />

(e.g., 1000 variables) on large computers. It is very effective for many<br />

problems (e.g., up to 25 variables) on desktop computers. The nature <strong>of</strong> the<br />

conjugate gradient search direction is described next, followed by a description<br />

<strong>of</strong> the Fletcher-Reeves algorithm.<br />

5.2.4. Fletcher-Reeves Conjugate Gradient Search Directions. Two vectors,<br />

x and y, are said to be orthogonal (perpendicular) if their inner product is<br />

zero, I.e.,<br />

xTy=O=XTUy, (5.51)<br />

where the unit matrix has been introduced to emphasize the following concept.<br />

The vectors are said to be conjugate if<br />

x T Ay=O, (5.52)<br />

where A is a positive-definite matrix. Conjugacy requires that the vec'tors are<br />

not parallel. More remarkably, conjugate vectors relate to A-quadratic forms<br />

as depicted in Figure 5.19. Just as illustrated for ellipsoids without cross terms,

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!