Visual navigation using inverse reinforcement learning and an extreme learning machine

4Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we focus on the challenges of training efficiency, the designation of reward functions, and generalization in reinforcement learning for visual navigation and propose a regularized extreme learning machine-based inverse reinforcement learning approach (RELM-IRL) to improve the navigation performance. Our contributions are mainly three-fold: First, a framework combining extreme learning machine with inverse reinforcement learning is presented. This framework can improve the sample efficiency and obtain the reward function directly from the image information observed by the agent and improve the generation for the new target and the new envi-ronment. Second, the extreme learning machine is regularized by multi-response sparse regression and the leave-one-out method, which can further improve the generalization ability. Simulation experiments in the AI-THOR environment showed that the proposed approach outperformed previous end-to-end approaches, thus, demonstrating the effectiveness and efficiency of our approach.

Cite

CITATION STYLE

APA

Fang, Q., Zhang, W., & Wang, X. (2021). Visual navigation using inverse reinforcement learning and an extreme learning machine. Electronics (Switzerland), 10(16). https://doi.org/10.3390/electronics10161997

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free