Learning a transferable world model by reinforcement agent in deterministic observable grid-world environments

1Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.

Abstract

Reinforcement-based agents have difficulties in transferring their acquired knowledge into new different environments due to the common identities-based percept representation and the lack of appropriate generalization capabilities. In this paper, the problem of knowledge transferability is addressed by proposing an agent dotted with decision tree induction and constructive induction capabilities and relying on decomposable properties-based percept representation. The agent starts without any prior knowledge of its environment and of the effects of its actions. It learns a world model (the set of decision trees) that corresponds to the set of explicit action definitions predicting action effects in terms of agent's percepts. Agent's planning component uses predictions of the world model to chain actions via a breadth-first search. The proposed agent was compared to the Q-learning and Adaptive Dynamic Programming based agents and demonstrated better ability to achieve goals in static observable deterministic gridworld environments different from those in which it has learnt its world model.

Cite

CITATION STYLE

APA

Kapociute-Dzikiene, J., & Raškinis, G. (2012). Learning a transferable world model by reinforcement agent in deterministic observable grid-world environments. Information Technology and Control, 41(4), 318–327. https://doi.org/10.5755/j01.itc.41.4.915

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free