27.06.2013 Views

Evolution and Optimum Seeking

Evolution and Optimum Seeking

Evolution and Optimum Seeking

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

78 Hill climbing Strategies<br />

converges quadratically, i.e., after a nite number (maximum n) of steps the minimum of<br />

a quadratic function is located. Myers (1968) <strong>and</strong> Huang (1970) show that, if the same<br />

starting point ischosen <strong>and</strong> the objective function is of second order, the DFP algorithm<br />

generates the same iteration points as the conjugate gradient method of Fletcher <strong>and</strong><br />

Reeves.<br />

All these observations are based on the assumption that the computational operations,<br />

including the line searches, are carried out exactly. Then H (k) always remains positivede<br />

nite if H (0) was positive-de nite <strong>and</strong> the minimum search is stable, i.e., the objective<br />

function is improved at each iteration. Numerical tests (e.g., Pearson, 1969 Tabak,<br />

1969 Huang <strong>and</strong> Levy, 1970 Murtagh <strong>and</strong> Sargent, 1970 Himmelblau, 1972a,b), <strong>and</strong><br />

theoretical considerations (Bard, 1968 Dixon, 1972b) show that rounding errors <strong>and</strong><br />

especially inaccuracies in the one dimensional minimization frequently cause stability<br />

problems the matrix H (k) can easily lose its positive-de niteness without this being due<br />

to a singularity intheinverse Hessian matrix. The simplest remedy for a singular matrix<br />

H (k) , or one of reduced rank, is to forget from time to time all the experience stored within<br />

H (k) <strong>and</strong> to begin again with the unit matrix <strong>and</strong> simple gradient directions (Bard, 1968<br />

McCormick <strong>and</strong> Pearson, 1969). To do so certainly increases the number of necessary<br />

iterations, but in optimization as in other activities it is wise to put safety before speed.<br />

Stewart (1967) makes use of this procedure. His algorithm is of very great practical<br />

interest since he obtains the required information about the rst partial derivatives from<br />

function values alone by means of a cleverly constructed di erence scheme.<br />

3.2.3.2 Strategy of Stewart:<br />

Derivative-free Variable Metric Method<br />

Stewart (1967) focuses his attentiononchoosing the length of the trial step d (k)<br />

i<br />

approximation<br />

g (k)<br />

i ' Fx i x (k) = @F(x)<br />

@xi x (k)<br />

for the<br />

to the rst partial derivatives in such away astominimize the in uence of rounding<br />

errors on the actual iteration process. Two di erence schemes are available:<br />

<strong>and</strong><br />

g (k)<br />

i<br />

g (k)<br />

i = 1<br />

d (k)<br />

i<br />

=<br />

1<br />

2 d (k)<br />

i<br />

h F (x (k) + d (k)<br />

i ei) ; F (x (k) ) i<br />

h F (x (k) + d (k)<br />

i ei) ; F (x (k) ; d (k)<br />

i ei) i<br />

(forward di erence)<br />

(central di erence) (3.32)<br />

Application of the one sided (forward) di erence (Equation (3.32)) is preferred, since<br />

it only involves one extra function evaluation. To simplify the computation, Stewart<br />

introduces the vector h (k) ,whichcontains the diagonal elements of the matrix (H (k) ) ;1<br />

representing information about the curvature of the objective function in the coordinate<br />

directions.<br />

The algorithm for determining the g (k)<br />

i<br />

i = 1(1)n runs as follows:

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!