04.02.2015 Views

Stochastic Programming - Index of

Stochastic Programming - Index of

Stochastic Programming - Index of

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

226 STOCHASTIC PROGRAMMING<br />

By assumption (9.3 i), ϕ(x) :=E˜ξF (x, ˜ξ) isconvexinx. If this function is<br />

also differentiable with respect to x at any arbitrary point z with the gradient<br />

g := ∇ϕ(z) =∇ x E˜ξF<br />

(z, ˜ξ), then −g is the direction <strong>of</strong> steepest descent <strong>of</strong><br />

ϕ(x) = E˜ξF<br />

(x, ˜ξ) inz, and we should probably like to choose −g as the<br />

search direction to decrease our objective. However, this does not seem to be<br />

a practical approach, since, as we know already, evaluating ϕ(x) =E˜ξF<br />

(x, ˜ξ),<br />

as well as ∇ϕ(z) =∇ x E˜ξF<br />

(z, ˜ξ), is a rather cumbersome task.<br />

In the differentiable case we know from Proposition 1.21 on page 81 that,<br />

for a convex function ϕ,<br />

(x − z) T ∇ϕ(z) ≤ ϕ(x) − ϕ(z) (9.6)<br />

has to hold ∀x, z (see Figure 27 in Chapter 1). But, even if the convex function<br />

ϕ is not differentiable at some point z, e.g.ifithasakinkthere,itisshown<br />

in convex analysis that there exists at least one vector g such that<br />

(x − z) T g ≤ ϕ(x) − ϕ(z) ∀x. (9.7)<br />

Any vector g satisfying (9.7) is called a subgradient <strong>of</strong> ϕ at z, and the set <strong>of</strong> all<br />

vectors satisfying (9.7) is called the subdifferential <strong>of</strong> ϕ at z and is denoted by<br />

∂ϕ(z). If ϕ is differentiable at z then ∂ϕ(z) ={∇ϕ(z)}; otherwise,i.e.inthe<br />

nondifferentiable case, ∂ϕ(z) may contain more than one element as shown<br />

for instance in Figure 32. Furthermore, in view <strong>of</strong> (9.7), it is easily seen that<br />

∂ϕ(z) is a convex set.<br />

If ϕ is convex and g ≠ 0 is a subgradient <strong>of</strong> ϕ at z then, by (9.7) for λ>0,<br />

it follows that<br />

ϕ(z + λg) ≥ ϕ(z)+g T (x − z)<br />

= ϕ(z)+g T (λg)<br />

= ϕ(z)+λ‖g‖ 2<br />

>ϕ(z).<br />

Hence any subgradient, g ∈ ∂ϕ, such that g ≠ 0 is a direction <strong>of</strong> ascent,<br />

although not necessarily the direction <strong>of</strong> steepest ascent as the gradient would<br />

be if ϕ were differentiable in z. However, in contrast to the differentiable case,<br />

−g need not be a direction <strong>of</strong> strict descent for ϕ in z. Consider for example<br />

the convex function in two variables<br />

ψ(u, v) :=|u| + |v|.<br />

Then for ẑ =(0, 3) T we have g =(1, 1) T ∈ ∂ψ(ẑ), since for all ε>0the<br />

gradient ∇ψ(ε, 3) exists and is equal to g. Hence, by (9.6), we have, for all<br />

(u, v),<br />

[( ) ( )] T ( ) T ( )<br />

u ε u − ε 1<br />

− g =<br />

v 3<br />

v − 3 1<br />

= u − ε + v − 3<br />

≤|u| + |v|−|ε|−|3|,

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!