11.07.2015 Views

View - Universidad de Almería

View - Universidad de Almería

View - Universidad de Almería

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

A Trust-Region Algorithm for Global Optimization 111098765432100 1 2 3 4 5 6 7 8 9 10Figure 2.Example of the effect of local minimization2. Summary2.1 Gaussian SmoothingWe introduce a smoothing scheme to solve (3). The material of this section largely follows[2], although we give a different emphasis. We apply smoothing to L(x) for two reasons.First, L(x) is a piecewise constant function for which <strong>de</strong>scent directions are difficult to <strong>de</strong>fine(first-or<strong>de</strong>r <strong>de</strong>rivatives, when <strong>de</strong>fined, are always zero). Second, we expect the smoothing toprovi<strong>de</strong> a more global view of the function. Given a real-valued function L : IR n → IR anda smoothing kernel g : IR → IR, which is a continuous, boun<strong>de</strong>d, nonnegative, symmetricfunction whose integral is one, we <strong>de</strong>fine the g–transform of L(x) as∫〈L〉 g (x) = L(y)g(‖y − x‖) dy. (5)nIRThe value 〈L〉 g (x) is an average of the values of L(x) in all the domain; in particular, the closerthe points are to x, the higher is the contribution to the resulting value. Another importantproperty is that 〈L〉 g (x) is a differentiable function. Hence we can use standard smooth optimizationmethods to minimize it.The most wi<strong>de</strong>ly used kernel in the literature is the Gaussian kernelg(z) ∝ exp ( −z 2 /(2σ 2 ) ) ,where we use the symbol ∝ to avoid writing a multiplicative constant that plays no role in themethods we present here. Clearly, one cannot explicitly apply the smoothing operator as in(5) because this approach requires the approximation of an n-dimensional integral. Instead,we restrict our attention to a ball of radius ∆ around the current point x (B(x, ∆)) and weconstruct a discretized version of the smoothing of L(x),ˆL K B g (x) = ∑ g(‖y i − x‖)L(y i ) ∑ Ki=1 g(‖y i − x‖) , (6)i=1where y i , i = 1...K, are samples in B(x, ∆). This function has interesting properties: it is acontinuous function and a convex sum of the values of the samples. In particular, the weightassociated with each sample is larger if we are closer to the sample point. In other words,the more confi<strong>de</strong>nt we are in the sample value, the greater is the weight associated with it.In [2], the mo<strong>de</strong>l (6) is used to choose the new candidate point x + , starting from a point (say,

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!