27.06.2013 Views

Evolution and Optimum Seeking

Evolution and Optimum Seeking

Evolution and Optimum Seeking

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

66 Hill climbing Strategies<br />

A distinction is sometimes drawn between short step methods, which evaluate the gradients<br />

again after a small step in the direction rF (x (k) ) (for maximization) or ;rF (x (k) )<br />

(for minimization), <strong>and</strong> their equivalent long step methods. Since for nite step lengths<br />

s (k) it is not certain whether the new variable vector is really better than the old, after the<br />

step the value of the objective function must be tested again. Working with small steps<br />

increases the number of objective function calls <strong>and</strong> gradientevaluations. Besides F (x) n<br />

partial derivatives must be evaluated. Even if the slopes can be obtained analytically <strong>and</strong><br />

can be speci ed as functions, there is no reason to suppose that the number of computational<br />

operations per function call is much less than for the objective function itself.<br />

Except in special cases, the total cost increases roughly as the weighting factor (n +1)<br />

<strong>and</strong> the number of objective function calls. This also holds if the partial derivatives are<br />

approximated by di erential quotients obtained by means of trial steps<br />

Fxi(x) = @F(x)<br />

=<br />

@xi<br />

F (x + ei) ; F (x) 2<br />

+ O( ) for all i =1(1)n<br />

Additional di culties arise here since for values of that are too small the subtraction<br />

is subject to rounding error, while for trial steps that are too large the neglect of terms<br />

O( 2 ) leads to incorrect values. The choice of suitable deviations requires special care<br />

in all cases (Hildebr<strong>and</strong>, 1956 Curtis <strong>and</strong> Reid, 1974).<br />

Cauchy (1847), Kantorovich (1940, 1945), Levenberg (1944), <strong>and</strong> Curry (1944) are the<br />

originators of the gradient strategy, which started life as a method of solving equations<br />

<strong>and</strong> systems of equations. It is rst referred to as an aid to solving variational problems<br />

by Hadamard (1908) <strong>and</strong> Courant (1943). Whereas Cauchy works with xed step lengths<br />

s (k) , Curry tries to determine the distance covered in the (not normalized) direction<br />

v (k) = ;rF (x (k) ) so as to reach a relative minimum (see also Brown, 1959). In principle,<br />

any one of the one dimensional search methods of Section 3.1 can be called upon to nd<br />

the optimal value for s (k) :<br />

F (x (k) + s (k) v (k) ) = min<br />

s fF (x (k) + sv (k) )g<br />

This variant of the basic strategy could thus be called a longest step procedure. It is<br />

better known however under the name optimum gradient method, ormethod ofsteepest<br />

descent (for maximization, ascent). Theoretical investigations of convergence <strong>and</strong> rate<br />

of convergence of the method can be found, e.g., in Akaike (1960), Goldstein (1962),<br />

Ostrowski (1967), Forsythe (1968), Elkin (1968), Zangwill (1969), <strong>and</strong> Wolfe (1969, 1970,<br />

1971). Zangwill proves convergence based on the assumptions that the line searches are<br />

exact <strong>and</strong> the objective function is continuously twice di erentiable. Exactness of the one<br />

dimensional minimization is not, however, a necessary assumption (Wolfe, 1969). It is<br />

signi cant that one can only establish theoretically that a stationary point willbereached<br />

(rF (x) = 0) or approached (krF (x)k < "" > 0). The stationary point is a minimum,<br />

only if F (x) is convex <strong>and</strong> three times di erentiable (Akaike, 1960). Zellnik, Sondak, <strong>and</strong><br />

Davis (1962), however, show that saddle points are in practice an obstacle, only if the<br />

search is started at one, or on a straight gradient trajectory passing through one. In other<br />

cases numerical rounding errors ensure that the path to a saddle point is unstable.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!