Using localization and factorization to reduce the complexity of reinforcement learning

1Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

General reinforcement learning is a powerful framework for artificial intelligence that has seen much theoretical progress since introduced fifteen years ago.We have previously provided guarantees for cases with finitely many possible environments. Though the results are the best possible in general, a linear dependence on the size of the hypothesis class renders them impractical. However, we dramatically improved on these by introducing the concept of environments generated by combining laws. The bounds are then linear in the number of laws needed to generate the environment class. This number is identified as a natural complexity measure for classes of environments. The individual law might only predict some feature (factorization) and only in some contexts (localization). We here extend previous deterministic results to the important stochastic setting.

Cite

CITATION STYLE

APA

Sunehag, P., & Hutter, M. (2015). Using localization and factorization to reduce the complexity of reinforcement learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9205, pp. 177–186). Springer Verlag. https://doi.org/10.1007/978-3-319-21365-1_19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free