Learning on real robots from experience and simple user feedback

12Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

In this article we describe a novel algorithm that allows fast and continuous learning on a physical robot working in a real environment. The learning process is never stopped and new knowledge gained from robot-environment interactions can be incorporated into the controller at any time. Our algorithm lets a human observer control the reward given to the robot, hence avoiding the burden of defining a reward function. Despite the highly-non-deterministic reinforcement, through the experimental results described in this paper, we will see how the learning processes are never stopped and are able to achieve fast robot adaptation to the diversity of different situations the robot encounters while it is moving in several environments.

Cite

CITATION STYLE

APA

Quint́ia, P., Iglesias, R., Rodŕiguez, M. A., & Regueiro, C. V. (2013). Learning on real robots from experience and simple user feedback. Journal of Physical Agents, 7(1), 56–64. https://doi.org/10.14198/jopha.2013.7.1.08

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free