Learning Visuomotor Policies with Deep Movement Primitives

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we present a novel method to learn end-to-end visuomotor policies for robotic manipulators. The method computes state-action mappings in a supervised learning manner from video demonstrations and robot trajectories. We show that the robot learns to perform different tasks by associating image features with the corresponding movement primitives of different grasp poses. To evaluate the effectiveness of the proposed learning method, we conduct experiments with a PR2 robot in a simulation environment. The purpose of these experiments is to evaluate the system's ability to perform manipulation tasks.

Cite

CITATION STYLE

APA

Theofanidis, M., Bozcuoglu, A. K., Neumann, M., Cloud, J., Kyrarini, M., Makedon, F., & Beetz, M. (2021). Learning Visuomotor Policies with Deep Movement Primitives. In ACM International Conference Proceeding Series (pp. 140–146). Association for Computing Machinery. https://doi.org/10.1145/3453892.3453899

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free