27.06.2013 Views

Evolution and Optimum Seeking

Evolution and Optimum Seeking

Evolution and Optimum Seeking

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Multidimensional Strategies 83<br />

(1970b,c), Shanno (1970a,b), <strong>and</strong> Shanno <strong>and</strong> Kettler (1970) give criteria for choosing suitable<br />

(k) .However, the mixed correction, also known as BFS or Broyden-Fletcher-Shanno<br />

formula, cannot guarantee quadratic convergence either unless line searches are carried<br />

out. It can be proved that there will merely be a monotonic decrease in the eigenvalues of<br />

the matrix H (k) .From numerical tests, however, it turns out that the increased number<br />

of iterations is usually more than compensated for by thesaving in function calls made<br />

by dropping the one dimensional optimizations (Fletcher, 1970a). Fielding (1970) has designed<br />

an ALGOL program following Broyden's work with line searches (Broyden, 1965).<br />

With regard to the number of function calls it is usually inferior to the DFP method but<br />

it sometimes also converges where the variable metric method fails. Dixon (1973) de nes<br />

a correction to the chosen directions,<br />

where<br />

<strong>and</strong><br />

v (k) = ;H (k) rF (x (k) )+w (k)<br />

w (0) =0<br />

w (k+1) = w (k) + (x(k+1) ; x (k) ) T rF (x (k+1) )<br />

(x (k+1) ; x (k) ) T z (k)<br />

(x (k+1) ; x (k) )<br />

by which, together with a matrix correction as given by Equation (3.35), quadratic convergence<br />

can be achieved without line searches. He shows that at most n +2 function<br />

calls <strong>and</strong> gradient calculations are required each time if, after arriving at v (k) = 0, an<br />

iteration<br />

x (k+1) = x (k) ; H (k) rF (x (k) )<br />

is included. Nearly all the procedures de ned assume that at least the rst partial derivatives<br />

are speci ed as functions of the variables <strong>and</strong> are therefore exact to the signi cant<br />

gure accuracy of the computer used. The more costly matrix computations should<br />

wherever possible be executed with double precision in order to keep down the e ect of<br />

rounding errors.<br />

Just two more suggestions for derivative-free quasi-Newton methods will be mentioned<br />

here: those of Greenstadt (1972) <strong>and</strong> of Cullum (1972). While Cullum's algorithm, like<br />

Stewart's, approximates the gradient vector by function value di erences, Greenstadt attempts<br />

to get away from this. Analogously to Davidon's idea of approximating the Hessian<br />

matrix during the course of the iterations from knowledge of the gradients, Greenstadt<br />

proposes approximating the gradients by using information from objective function values<br />

over several subiterations. Only at the starting point must a di erence scheme for the rst<br />

partial derivatives be applied. Another interesting variable metric technique described by<br />

Elliott <strong>and</strong> Sworder (1969a,b, 1970) combines the concept of the stochastic approximation<br />

for the sequence of step lengths with the direction algorithms of the quasi-Newton<br />

strategy.<br />

Quasi-Newton strategies of degree one are especially suitable if the objective function<br />

is a sum of squares (Bard, 1970). Problems of minimizing a sum of squares arise<br />

for example from the problem of solving systems of simultaneous, non-linear equations,

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!