Scalable multitask policy gradient reinforcement learning

15Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

Abstract

Policy search reinforcement learning (RL) allows agents to learn autonomously with limited feedback. However, such methods typically require extensive experience for successful behavior due to their tabula rasa nature. Multitask RL is an approach, which aims to reduce data requirements by allowing knowledge transfer between tasks. Although successful, current multitask learning methods suffer from scalability issues when considering large number of tasks. The main reasons behind this limitation is the reliance on centralized solutions. This paper proposes to a novel distributed multitask RL framework, improving the scalability across many different types of tasks. Our framework maps multitask RL to an instance of general consensus and develops an efficient decentralized solver.We justify the correctness of the algorithm both theoretically and empirically: we first proof an improvement of convergence speed to an order of O(1/k) with k being the number of iterations, and then show our algorithm surpassing others on multiple dynamical system benchmarks.

Cite

CITATION STYLE

APA

El Bsat, S., Ammar, H. B., & Taylor, M. E. (2017). Scalable multitask policy gradient reinforcement learning. In 31st AAAI Conference on Artificial Intelligence, AAAI 2017 (pp. 1847–1853). AAAI press. https://doi.org/10.1609/aaai.v31i1.10942

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free