16.11.2012 Views

Wireless Network Design: Optimization Models and Solution ...

Wireless Network Design: Optimization Models and Solution ...

Wireless Network Design: Optimization Models and Solution ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

14 Cross Layer Scheduling in <strong>Wireless</strong> <strong>Network</strong>s 339<br />

Theorem 14.1. The optimal policy µ ∗ (s) = µ ∗ (q,x) is non decreasing in channel<br />

state x <strong>and</strong> non-decreasing in queue length q.<br />

The proof of this theorem is based on supermodularity <strong>and</strong> increasing differences<br />

properties of the value function. These properties establish the monotonicity of the<br />

optimal policy in channel state <strong>and</strong> buffer occupancy.<br />

The structural results imply that the optimal decision is to transmit a certain number<br />

of packets in a given slot where this number is an increasing function of the<br />

current queue length <strong>and</strong> channel state. Thus for a fixed channel state, the greater<br />

the queue length, the more will be the number of packets that will be transmitted.<br />

Similarly, for a fixed queue length, the better the channel, the more will be the number<br />

of packet transmissions. Thus, the optimal policy always transmits at the highest<br />

rate when channel condition is the most favorable <strong>and</strong> queue length is the largest.<br />

Similar structural results also hold for Markovian arrivals <strong>and</strong> Markovian channel<br />

state <strong>and</strong> continuous state processes. However, for these general state spaces,<br />

the dynamic programming equation (14.52) for the long run average cost problem<br />

requires to be rigorously justified. See bibliographic notes <strong>and</strong> [1] for a discussion<br />

of this.<br />

14.5.1.3 Learning Algorithm for Scheduling<br />

In this section, we discuss the computation of optimal policy discussed in the preceding<br />

section. For a fixed λ, the Relative Value Iteration Algorithm (RVIA) can<br />

be used for solving the dynamic programming equation in an iterative fashion. The<br />

average cost RVIA for determining the value function such that (14.52) is satisfied<br />

can be written as:<br />

Vn+1(s) = min<br />

u∈U(s) [c(λ,s,u) +∑ s ′<br />

p(s,u,s ′ )Vn(s ′ )] −Vn(s 0 ), (14.56)<br />

where s,s ′ ,s 0 ∈ S <strong>and</strong> s 0 is any fixed state. Vn(·) is an estimate of the value function<br />

after n iterations for a fixed LM λ.<br />

Due to the large size of the state space S, solution of this algorithm is computationally<br />

expensive. One approach for addressing this issue is to utilize the structural<br />

properties of the optimal policy outlined above to develop efficient heuristics that<br />

are computationally less expensive <strong>and</strong> hence can be implemented in a practical system.<br />

In [54], one such approach has been discussed. Techniques based on function<br />

approximation can also be employed for dimensionality reduction. The structural<br />

results of the policy may be used to choose the basis functions in function approximation.<br />

Another technique is to use ‘primal-dual’ type approach with conventional iteration<br />

schemes for the value function (primal) <strong>and</strong> Lagrange multiplier (dual) [13].<br />

However, even with this approach (<strong>and</strong> for that matter, with all other approaches<br />

discussed above for computing (14.56)), one major issue is that it requires the<br />

knowledge of transition probabilities p(s,u,s ′ ). Note that p(s,u,s ′ ) depends upon

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!