Stochastic Programming - Index of
Stochastic Programming - Index of
Stochastic Programming - Index of
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
116 STOCHASTIC PROGRAMMING<br />
Pro<strong>of</strong> The unique meaning <strong>of</strong> “max” implies that we have that<br />
max ϕ 1(r 1 (z 1 ,x 1 ),ψ 2 (r 2 (z 2 ,x 2 ), ···,r T (z T ,x T )))<br />
{x t∈X t,t≥1}<br />
≥ ϕ 1 (r 1 (z 1 ,x 1 ), max [ψ 2(r 2 (z 2 ,x 2 ), ···,r T (z T ,x T ))]),<br />
{x t∈X t,t≥2}<br />
for all x 1 . Therefore this also holds when the right-hand side <strong>of</strong> this inequality<br />
is maximized with respect to x 1 .<br />
On the other hand, it is also obvious that<br />
max ψ 2(r 2 (z 2 ,x 2 ), ···,r T (z T ,x T ))<br />
{x t∈X t,t≥2}<br />
≥ ψ 2 (r 2 (z 2 ,x 2 ), ···,r T (z T ,x T )) ∀x t ∈ X t ,t≥ 2.<br />
Hence, by the assumed monotonicity <strong>of</strong> ϕ 1 with respect to ψ 2 ,wehavethat<br />
ϕ 1 (r 1 (z 1 ,x 1 ), max ψ 2(r 2 (z 2 ,x 2 ), ···,r T (z T ,x T )))<br />
{x t∈X t,t≥2}<br />
≥ ϕ 1 (r 1 (z 1 ,x 1 ),ψ 2 (r 2 (z 2 ,x 2 ), ···,r T (z T ,x T ))) ∀x t ∈ X t ,t≥ 1.<br />
Taking the maximum with respect to x t ,t≥ 2, on the right-hand side <strong>of</strong> this<br />
inequality and maximizing afterwards both sides with respect to x 1 ∈ X 1<br />
shows that the optimality principle (1.2) holds.<br />
✷<br />
Needless to say, all problems considered by Bellman in his first book on<br />
dynamic programming satisfied this proposition. In case (b) <strong>of</strong> our example,<br />
however, the monotonicity does not hold. The reason is that when “⊕”involves<br />
multiplication <strong>of</strong> possibly negative factors (i.e. negative immediate returns),<br />
the required monotonicity is lost. On the other hand, when “⊕” is summation,<br />
the required monotonicity is always satisfied.<br />
Let us add that the optimality principle applies to a much wider class <strong>of</strong><br />
problems than might seem to be the case from this brief sketch. For instance,<br />
if for finitely many states we denote by ρ t the vector having as ith component<br />
the immediate return r t (z t = i), and if we define the composition operation<br />
“⊕” such that, with a positive matrix S (i.e. all elements <strong>of</strong> S nonnegative),<br />
ρ t ⊕ ρ t+1 = ρ t + Sρ t+1 , t =1, ···,T − 1,<br />
then the monotonicity assumed for Proposition 2.2 follows immediately. This<br />
case is quite common in applications. Then S is the so-called transition matrix,<br />
which means that an element s ij represents the probability <strong>of</strong> entering state<br />
j at stage t + 1, given that the system is in state i at stage t. Iterating the<br />
above composition for T − 1,T − 2, ···, 1wegetthatF (ρ 1 , ···,ρ T )isthe<br />
vector <strong>of</strong> the expected total returns. The ith component gives the expected<br />
overall return if the system starts from state i at stage 1. Putting it this way<br />
we see that multistage stochastic programs with recourse (formula (4.13) in<br />
Chapter 1) belong to this class.