16.11.2012 Views

Wireless Network Design: Optimization Models and Solution ...

Wireless Network Design: Optimization Models and Solution ...

Wireless Network Design: Optimization Models and Solution ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

14 Cross Layer Scheduling in <strong>Wireless</strong> <strong>Network</strong>s 337<br />

Let cp(Sn,Un) = P(Xn,Un) be the ‘immediate’ cost in terms of power required<br />

in transmitting Un packets when the state is Sn. Let cq(Sn,Un) ∆ = Qn denote the<br />

‘immediate’ cost due to buffering. Let µ = {µ1, µ2,...} be the control policy. We<br />

would like to determine the policy µ that minimizes<br />

subject to<br />

¯P = limsup<br />

N→∞<br />

¯Q = limsup<br />

N→∞<br />

1<br />

N E<br />

1<br />

N E<br />

N<br />

∑<br />

n=1<br />

N<br />

∑<br />

n=1<br />

cp(Sn,Un), (14.44)<br />

cq(Sn,Un) ≤ ¯ δ. (14.45)<br />

It can be easily argued that this is a CMDP with average cost <strong>and</strong> finite state <strong>and</strong><br />

action spaces. For average cost CMDP with finite state <strong>and</strong> action space, it is well<br />

known that a optimal stationary r<strong>and</strong>omized policy exists. Let ¯P ∗ denote the optimal<br />

cost i.e.,<br />

¯P ∗ = min<br />

µ<br />

¯P µ , (14.46)<br />

where ¯P µ is the cost (14.44) under policy µ.<br />

Let µ(·|s) : s ∈ F be the probability measure on U. For each state s, µ(·|s) specifies<br />

the distribution with which the control in that state is applied. We assume that<br />

{Sn} is an ergodic Markov chain under such policies <strong>and</strong> thus has a unique stationary<br />

distribution ρ µ .<br />

Let E µ denote the expectation with respect to (w.r.t.) ρ µ . Under a r<strong>and</strong>omized<br />

policy µ, the costs in (14.44) can be expressed as:<br />

<strong>and</strong>,<br />

¯P µ ∆ = E µ �<br />

cp(Sn, µ(Sn))<br />

�<br />

= ∑ρ u,s<br />

µ (s)µ(u|s)cp(s, µ(s)), (14.47)<br />

¯Q µ ∆ µ<br />

= E �<br />

�<br />

cq(Sn, µ(Sn)) = ∑ρ u,s<br />

µ (s)µ(u|s)cq(s, µ(s)), (14.48)<br />

respectively. Then the scheduler objective can be stated as:<br />

Minimize ¯P µ subject to ¯Q µ ≤ δ. (14.49)<br />

We now demonstrate that the optimal average cost <strong>and</strong> policy can be determined<br />

using an unconstrained Markov Decision Process (MDP) problem <strong>and</strong> Lagrangian<br />

approach.<br />

14.5.1.1 The Lagrangian Approach<br />

Let λ ≥ 0 be a real number. Define c : R + × S × U → R as follows,<br />

c(λ,s,u) = cp(s,u) + λ(cq(s,u) − δ). (14.50)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!