Skip to content

Learning to deal with objects

by María Malfaz, Miguel A. Salichs
2009 IEEE 8th International Conference on Development and Learning, ICDL 2009 ()
Get full text at journal


In this paper, a modification of the standard learning algorithm Q-learning is presented: Object Q-learning (OQ-learning). An autonomous agent should be able to decide its own goals and behaviours in order to fulfil these goals. When the agent has no previous knowledge, it must learn what to do in every state (policy of behaviour). If the agent uses Q-learning, this implies that it learns the utility value Q of each action-state pair. Typically, an autonomous agent living in a complex environment has to interact with different objects present in that world. In this case, the number of states of the agent in relation to those objects may increase as the number of objects increases, making the learning process difficult to deal with. The proposed modification appears as a solution in order to cope with this problem. The experimental results prove the usefulness of the OQ-learning in this situation, in comparison with the standard Q-learning algorithm.

Cite this document (BETA)

Readership Statistics

9 Readers on Mendeley
by Discipline
56% Computer Science
22% Social Sciences
11% Medicine and Dentistry
by Academic Status
33% Student > Ph. D. Student
22% Researcher
22% Student > Master
by Country
22% Spain
11% Germany

Sign up today - FREE

Mendeley saves you time finding and organizing research. Learn more

  • All your research in one place
  • Add and import papers easily
  • Access it anywhere, anytime

Start using Mendeley in seconds!

Sign up & Download

Already have an account? Sign in