An overview of MAXQ hierarchical reinforcement learning

58Citations
Citations of this article
92Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Reinforcement learning addresses the problem of learning optimal policies for sequential decision-making problems involving stochastic operators and numerical reward functions rather than the more traditional deterministic operators and logical goal predicates. In many ways, reinforcement learning research is recapitulating the development of classical research in planning and problem solving. After studying the problem of solving "flat" problem spaces, researchers have recently turned their attention to hierarchical methods that incorporate subroutines and state abstractions. This paper gives an overview of the MAXQ value function decomposition and its support for state abstraction and action abstraction.

Cite

CITATION STYLE

APA

Dietterich, T. G. (2000). An overview of MAXQ hierarchical reinforcement learning. In Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science) (Vol. 1864, pp. 26–44). Springer Verlag. https://doi.org/10.1007/3-540-44914-0_2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free