Lipschitz Lifelong Reinforcement Learning

22Citations
Citations of this article
55Readers
Mendeley users who have this article in their library.

Abstract

We consider the problem of knowledge transfer when an agent is facing a series of Reinforcement Learning (RL) tasks. We introduce a novel metric between Markov Decision Processes and establish that close MDPs have close optimal value functions. Formally, the optimal value functions are Lipschitz continuous with respect to the tasks space. These theoretical results lead us to a value-transfer method for Lifelong RL, which we use to build a PAC-MDP algorithm with improved convergence rate. Further, we show the method to experience no negative transfer with high probability. We illustrate the benefits of the method in Lifelong RL experiments.

Cite

CITATION STYLE

APA

Lecarpentier, E., Abel, D., Asadi, K., Jinnai, Y., Rachelson, E., & Littman, M. L. (2021). Lipschitz Lifelong Reinforcement Learning. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 9B, pp. 8270–8278). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i9.17006

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free