Learning options for an MDP from demonstrations

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The options framework provides a foundation to use hierarchical actions in reinforcement learning. An agent using options, along with primitive actions, at any point in time can decide to perform a macro-action made out of many primitive actions rather than a primitive action. Such macro-actions can be hand-crafted or learned. There has been previous work on learning them by exploring the environment. Here we take a different perspective and present an approach to learn options from a set of experts demonstrations. Empirical results are also presented in a similar setting to the one used in other works in this area.

Author supplied keywords

Cite

CITATION STYLE

APA

Tamassia, M., Zambetta, F., Raffe, W., & Li, X. (2015). Learning options for an MDP from demonstrations. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8955, pp. 226–242). Springer Verlag. https://doi.org/10.1007/978-3-319-14803-8_18

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free