Model minimization in hierarchical reinforcement learning

45Citations
Citations of this article
36Readers
Mendeley users who have this article in their library.
Get full text

Abstract

When applied to real world problems Markov Decision Processes (MDPs) often exhibit considerable implicit redundancy, especially when there are symmetries in the problem. In this article we present an MDP minimization framework based on homomorphisms. The framework exploits redundancy and symmetry to derive smaller equivalent models of the problem. We then apply our minimization ideas to the options framework to derive relativized options—options defined without an absolute frame of reference. We demonstrate their utility empirically even in cases where the minimization criteria are not met exactly.

Cite

CITATION STYLE

APA

Ravindran, B., & Barto, A. G. (2002). Model minimization in hierarchical reinforcement learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2371, pp. 196–211). Springer Verlag. https://doi.org/10.1007/3-540-45622-8_15

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free