23.08.2015 Views

Here - Agents Lab - University of Nottingham

Here - Agents Lab - University of Nottingham

Here - Agents Lab - University of Nottingham

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

14. Džeroski, S., Raedt, L.D., Driessens, K.: Relational reinforcement learning. MachineLearning 43 (2001) 7–52 10.1023/A:1007694015589.15. Bordini, R.H., Dix, J., Dastani, M., Seghrouchni, A.E.F.: Multi-Agent Programming:Languages, Platforms and Applications. Volume 15 <strong>of</strong> Multiagent Systems,Artificial Societies, and Simulated Organizations. Springer (2005)16. Bordini, R.H., Dix, J., Dastani, M., Seghrouchni, A.E.F.: Multi-Agent Programming:Languages, Tools and Applications. Springer (2009)17. Broekens, J., Hindriks, K., Wiggers, P.: Reinforcement Learning as Heuristicfor Action-Rule Preferences. In: Programming Multi-Agent Systems (ProMAS).(2010)18. Hindriks, K.V., Riemsdijk, M.B.: Declarative agent languages and technologies vi.Springer-Verlag (2009) 215–23219. Bellman, R.E.: Dynamic Programming. Princeton <strong>University</strong> Press (1957)20. Watkins, C.J.: Learning from delayed rewards. PhD thesis, King’s College London(1989)21. Andre, D., Russell, S.J.: State abstraction for programmable reinforcement learningagents. In: Eighteenth national conference on Artificial intelligence, Menlo Park,CA, USA, American Association for Artificial Intelligence (2002) 119–12522. Pokahr, A., Braubach, L., Lamersdorf, W.: Jadex: A BDI reasoning engine. In:Multi-Agent Programming. Volume 15 <strong>of</strong> Multiagent Systems, Artificial Societies,And Simulated Organizations. Springer (2005) 149–17423. Subagdja, B., Sonenberg, L., Rahwan, I.: Intentional learning agent architecture.Autonomous <strong>Agents</strong> and Multi-Agent Systems 18 (2009) 417–47024. Singh, D., Sardina, S., Padgham, L.: Extending BDI plan selection to incorporatelearning from experience. Robotics and Autonomous Systems 58 (2010) 1067–107525. Singh, D., Sardina, S., Padgham, L., Airiau, S.: Learning context conditions forBDI plan selection. In: Proceedings <strong>of</strong> Autonomous <strong>Agents</strong> and Multi-Agent Systems(AAMAS). (May 2010) 325–33226. Singh, D., Sardina, S., Padgham, L., James, G.: Integrating learning into a BDIagent for environments with changing dynamics. In Toby Walsh, C.K., Sierra, C.,eds.: Proceedings <strong>of</strong> the International Joint Conference on Artificial Intelligence(IJCAI), Barcelona, Spain, AAAI Press (July 2011) 2525–253027. Anderson, J., Bothell, D., Byrne, M., Douglass, S., Lebiere, C., Qin, Y.: An integratedtheory <strong>of</strong> the mind. Psychological review 111(4) (2004) 103628. Fu, W., Anderson, J.: From recurrent choice to skill learning: A reinforcementlearningmodel. Journal <strong>of</strong> experimental psychology: General 135(2) (2006) 18429. Klahr, D., Langley, P., Neches, R.: Production system models <strong>of</strong> learning anddevelopment. The MIT Press (1987)30. Laird, J., Rosenbloom, P., Newell, A.: Chunking in soar: The anatomy <strong>of</strong> a generallearning mechanism. Machine Learning 1(1) (1986) 11–4631. Nason, S., Laird, J.: Soar-rl: Integrating reinforcement learning with soar. CognitiveSystems Research 6(1) (2005) 51–5932. Nejati, N., Langley, P., Konik, T.: Learning hierarchical task networks by observation.In: International Conference on Machine Learning, ACM Press (2006)665–67233. Slaney, J., Thiébaux, S.: Blocks world revisited. Artificial Intelligence 125(1-2)(2001) 119–15334. Gupta, N., Nau, D.: On the complexity <strong>of</strong> blocks-world planning. Artificial Intelligence56(2-3) (1992) 223–25435. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. TheJournal <strong>of</strong> Machine Learning Research 3 (2003) 1157–1182163

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!