23.02.2015 Views

Machine Learning - DISCo

Machine Learning - DISCo

Machine Learning - DISCo

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

CHAPTER 13 REINFORCEMENT LEARNJNG 375<br />

This recursive definition of Q provides the basis for algorithms that iteratively<br />

approximate Q (Watkins 1989). To describe the algorithm, we will use<br />

the symbol Q to refer to the learner's estimate, or hypothesis, of the actual Q<br />

function. In this algorithm the learner represents its hypothesis Q by a large table<br />

with a separate entry for each state-action pair. The table entry for the pair (s, a)<br />

stores the value for ~(s, a)-the learner's current hypothesis about the actual<br />

but unknown value Q(s, a). The table can be initially filled with random values<br />

(though it is easier to understand the algorithm if one assumes initial values of<br />

zero). The agent repeatedly observes its current state s, chooses some action a,<br />

executes this action, then observes the resulting reward r = r(s, a) and the new<br />

state s' = 6(s, a). It then updates the table entry for ~(s, a) following each such<br />

transition, according to the rule:<br />

Q(S, a) t r + y max &(st, a')<br />

a'<br />

(13.7)<br />

Note this training rule uses the agent's current Q values for the new state<br />

s' to refine its estimate of ~(s, a) for the previous state s. This training rule<br />

is motivated by Equation (13.6), although the training rule concerns the agent's<br />

approximation Q, whereas Equation (13.6) applies to the actual Q function. Note<br />

although Equation (13.6) describes Q in terms of the functions 6(s, a) and r(s, a),<br />

the agent does not need to know these general functions to apply the training<br />

rule of Equation (13.7). Instead it executes the action in its environment and<br />

then observes the resulting new state s' and reward r. Thus, it can be viewed as<br />

sampling these functions at the current values of s and a.<br />

The above Q learning algorithm for deterministic Markov decision processes<br />

is described more precisely in Table 13.1. Using this algorithm the agent's estimate<br />

Q converges in the limit to the actual Q function, provided the system can be<br />

modeled as a deterministic Markov decision process, the reward function r is<br />

Q learning algorithm<br />

For each s, a initialize the table entry ~(s, a) to zero.<br />

Observe the current state s<br />

Do forever:<br />

Select an action a and execute it<br />

Receive immediate reward r<br />

Observe the new state s'<br />

Update the table entry for ~ ( s a) , as follows:<br />

~(s,a) c r<br />

+ ymax&(s',af)<br />

a'<br />

S CS'<br />

TABLE 13.1<br />

Q learning algorithm, assuming deterministic rewards and actions. The discount factor y may be any<br />

constant such that 0 5 y < 1.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!