Least absolute policy iteration - A robust approach to value function approximation

4Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Least-squares policy iteration is a useful reinforcement learning method in robotics due to its computational e?ciency. However, it tends to be sensitive to outliers in observed rewards. In this paper, we propose an alternative method that employs the absolute loss for enhancing robustness and reliability. The proposed method is formulated as a linear programming problem which can be solved eficiently by standard optimization software, so the computational advantage is not sacrificed for gaining robustness and reliability. We demonstrate the usefulness of the proposed approach through a simulated robot-control task. Copyright © 2010 The Institute of Electronics, Information and Communication Engineers.

Cite

CITATION STYLE

APA

Sugiyama, M., Hachiya, H., Kashima, H., & Mortmura, T. (2010). Least absolute policy iteration - A robust approach to value function approximation. IEICE Transactions on Information and Systems, E93-D(9), 2555–2565. https://doi.org/10.1587/transinf.E93.D.2555

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free