Generalization in transfer learning: robust control of robot locomotion

18Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we propose a set of robust training methods for deep reinforcement learning to transfer learning acquired in one control task to a set of previously unseen control tasks. We improve generalization in commonly used transfer learning benchmarks by a novel sample elimination technique, early stopping, and maximum entropy adversarial reinforcement learning. To generate robust policies, we use sample elimination during training via a method we call strict clipping. We apply early stopping, a method previously used in supervised learning, to deep reinforcement learning. Subsequently, we introduce maximum entropy adversarial reinforcement learning to increase the domain randomization during training for a better target task performance. Finally, we evaluate the robustness of these methods compared to previous work on simulated robots in target environments where the gravity, the morphology of the robot, and the tangential friction coefficient of the environment are altered.

Cite

CITATION STYLE

APA

Ada, S. E., Ugur, E., & Akin, H. L. (2022). Generalization in transfer learning: robust control of robot locomotion. Robotica, 40(11), 3811–3836. https://doi.org/10.1017/S0263574722000625

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free