Visual Navigation with Asynchronous Proximal Policy Optimization in Artificial Agents

12Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Vanilla policy gradient methods suffer from high variance, leading to unstable policies during training, where the policy's performance fluctuates drastically between iterations. To address this issue, we analyze the policy optimization process of the navigation method based on deep reinforcement learning (DRL) that uses asynchronous gradient descent for optimization. A variant navigation (asynchronous proximal policy optimization navigation, appoNav) is presented that can guarantee the policy monotonic improvement during the process of policy optimization. Our experiments are tested in DeepMind Lab, and the experimental results show that the artificial agents with appoNav perform better than the compared algorithm.

Cite

CITATION STYLE

APA

Zeng, F., & Wang, C. (2020). Visual Navigation with Asynchronous Proximal Policy Optimization in Artificial Agents. Journal of Robotics, 2020. https://doi.org/10.1155/2020/8702962

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free