Dissertation - HQ
Dissertation - HQ
Dissertation - HQ
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
A general modelling framework for larval behaviour 125<br />
Numerical solution<br />
Given the description of the evolution of the state through transition<br />
matrices (Figure 6.2), solving the backward computation of gain and<br />
decisions (equation 6.4) is a matter of manipulating matrices. The<br />
final gain is a vector indexed by state. In the simple case presented in<br />
Figure 6.2, it is a column vector of length 12. Now, computing the value<br />
function (i.e. the optimal gain) at time T − 1 is a three step process:<br />
Multiply transition<br />
matrices and gain,<br />
backward in time<br />
1. Fill the matrices describing the transition probabilities from all<br />
states at t = T −1 to all other states at t = T , for the two decisions.<br />
These probabilities come from the description of the dynamical<br />
system (advection by currents, energy consumption, etc.) and this<br />
process has been detailed for three examples on page 121.<br />
2. Multiply each matrix by the final gain vector. For each initial state,<br />
this means multiplying the gain associated with every reachable<br />
final state by the probability to reach it and then summing all<br />
those products. These sums are therefore the mean gain at T − 1,<br />
for each state and each decision.<br />
3. For each state, compare the gain values in the two mean gain<br />
vectors (one for each decision), choose the maximum, and record<br />
to which decision it corresponds. This gives optimal mean gain<br />
and optimal decisions.<br />
To compute the value function at t = T − 2, repeat these steps with the<br />
optimal mean gain at t = T − 1 instead of the final gain. And similarly<br />
until t = 0.<br />
6.2.3 Example trajectories<br />
Given the description of the environment and the characteristics of<br />
the larva, optimal strategies (giving sequences of optimal decisions)<br />
and optimal trajectories (state trajectories for which the sequence of<br />
decisions is optimal) can be computed. There is no finite number of<br />
optimal trajectories. Indeed, in this version of the model, stochasticity<br />
is introduced by random survival.<br />
Two characteristic examples of optimal trajectories are presented in<br />
Figure 6.3. As remarked for the last two decisions, the behaviour of<br />
larvae is very intuitive. When it survives (left plot), the larva forages<br />
and lets itself be taken away by currents until it reaches its maximum<br />
energy resources. Then, it alternates swimming and foraging in order<br />
to keep energy close to its maximum at final time. In the right plot, the<br />
behaviour begins the same way but the larva dies at time step 25. We<br />
can conclude from the results of this simple model that the algorithm<br />
used to simulate larval behaviour and to solve the optimisation problem<br />
is correct.<br />
Predictable and sensible<br />
decisions: the problem<br />
is solved correctly