27.06.2013 Views

Evolution and Optimum Seeking

Evolution and Optimum Seeking

Evolution and Optimum Seeking

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Multidimensional Strategies 69<br />

devised a variant of the procedure (see also Sections 3.2.1.5 <strong>and</strong> 3.2.1.6). Lowe (1964)<br />

has gathered together the various schemes of trial steps for the EVOP strategy. The<br />

philosophy oftheEVOP strategy is treated in detail by Box <strong>and</strong> Draper (1969). Some<br />

examples of applications are given by Kenworthy (1967). The e ciency of methods of<br />

determining the gradient in the case of stochastic perturbations is dealt with by Mlynski<br />

(1964a,b, 1966a,b), Sergiyevskiy <strong>and</strong> Ter-Saakov (1970), <strong>and</strong> others.<br />

3.2.2.1 Strategy of Powell: Conjugate Directions<br />

The most important idea for overcoming the convergence di culties of the gradient strategy<br />

is due to Hestenes <strong>and</strong> Stiefel (1952), <strong>and</strong> again comes from the eld of linear algebra<br />

(see also Ginsburg, 1963 Beckman, 1967). It trades under the names conjugate directions<br />

or conjugate gradients. The directions fvi i = 1(1)ng are said to be conjugate with<br />

respect to a positive de nite n n matrix A if (Hestenes, 1956)<br />

v T<br />

i Avj =0 for all i j = 1(1)ni 6= j<br />

A further property of conjugate directions is their linear independence, i.e.,<br />

nX<br />

i=1<br />

i vi =0<br />

only holds if all the constants f i i =1(1)ng are zero. If A is replaced by the unit matrix,<br />

A = I, thenthevi are mutually orthogonal. With A = r 2 F (x) (Hessian matrix) the<br />

minimum of a quadratic function is obtained exactly in n line searches in the directions<br />

vi. This is a factor two better than the gradient Partan method. For general non-linear<br />

problems the convergence rate cannot be speci ed. As it is frequently assumed, however,<br />

that many problems behave roughly quadratically near the optimum, it seems worthwhile<br />

to use conjugate directions. The quadratic convergence of the search with conjugate<br />

directions comes about because second order properties of the objective function are<br />

taken into account. In this respect it is not, in fact, a rst order gradient method, but<br />

a second order procedure. If all the n rst <strong>and</strong> n<br />

(n +1) second partial derivatives are<br />

2<br />

available, the conjugate directions can be generated in one process corresponding to the<br />

Gram-Schmidt orthogonalization (Kowalik <strong>and</strong> Osborne, 1968). It calls for expensive<br />

matrix operations. Conjugate directions can, however, be constructed without knowledge<br />

of the second derivatives: for example, from the changes in the gradient vector in the<br />

course of the iterations (Fletcher <strong>and</strong> Reeves, 1964). Because of this implicit exploitation<br />

of second order properties, conjugate directions has been classi ed as a gradient method.<br />

The conjugate gradients method of Fletcher <strong>and</strong> Reeves consists of a sequence of line<br />

searches with Hermitian interpolation (see Sect. 3.1.2.3.4). As a rst search direction v (0)<br />

at the starting point x (0) , the simple gradient direction<br />

v (0) = ;rF (x (0) )<br />

is used. The recursion formula for the subsequent iterations is<br />

v (k) = (k) v (k;1) ;rF (x (k) ) for all k = 1(1)n (3.25)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!