06.06.2013 Views

STOCHASTIC

STOCHASTIC

STOCHASTIC

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

andom outcome, regret is defined to be the loss incurred by not following<br />

the best policy for the specific outcome. Thus, if the decision maker employed<br />

a reliable clairvoyant to inform him of the maximum price which would actually<br />

be realized, the optimal policy would be to sell the entire asset at the highest<br />

price. Failure to follow this policy would lead to regret, since profits would<br />

be lower than necessary. Not having access to a clairvoyant, the decision maker<br />

might choose instead to adopt a policy which leads to the smallest regret under<br />

worst possible conditions. By following such a policy the decision maker can<br />

be sure that even if events turn out badly, failure to adopt the policy would<br />

have exposed him to potentially worse outcomes. This is a conservative policy,<br />

but one which is clearly appropriate in cases where the decision maker is, for<br />

example, accountable for his actions after the fact. It is, in general, a much<br />

less conservative policy than always assuming the worst.<br />

One apparently common policy recommendation for asset selling is that of<br />

"dollar averaging." In such a policy, the decision maker sells equal quantities<br />

of the asset in each period. (In the case of security buying, equal quantities of<br />

dollars are "sold" in each period.) Pye shows that dollar averaging is an optimal<br />

nonsequential policy when the maximum possible single-period price increases<br />

and decreases are equal. In general, nonsequential minimax regret policies are<br />

not unique, and dollar averaging is merely one of an infinite number of optimal<br />

policies. In Exercise CR-13, another simple optimal nonsequential policy is<br />

outlined. Pye also derives the optimal nonsequential selling strategy for a policy<br />

of maximizing expected utility of total revenue. For any strictly concave utility<br />

function, the optimal policy is to sell all of the asset on the last day. In general,<br />

none of the infinitely many nonsequential minimax policies can have this form.<br />

Thus, an expected utility of revenue principle cannot lead to dollar averaging.<br />

Pye considers next the question of sequential minimax regret policies. Here<br />

the decision at each stage is conditioned on actual, realized past outcomes.<br />

Dynamic programming is used to derive the general form of an optimal policy.<br />

In the nomenclature of the Ziemba paper of Part I, the state variable at stage t<br />

consists of two dimensions: (1) the difference between the maximum price<br />

occurring up through time / and the price at time t, and (2) the quantity of<br />

asset still held at time t. The optimal return function at stage / is defined to<br />

be the regret at time / given that an optimal minimax sequential policy is<br />

followed at t and subsequently. A functional equation is derived for the optimal<br />

return function using backward induction. Pye shows that the return function<br />

is convex and decreasing in each state dimension, and for a fixed state, is<br />

decreasing in time. These properties are used to derive the form of the optimal<br />

policy, which is to sell at time t any portion of the asset in excess of a critical<br />

quantity. This critical value is a decreasing function of the first state dimension<br />

(i.e., the difference between maximum realized price up through t and the<br />

price at t). A numerical example is given to show that the critical value is not,<br />

442 PART V DYNAMIC MODELS

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!