Abstract
In this paper, we present a novel method to learn end-to-end visuomotor policies for robotic manipulators. The method computes state-action mappings in a supervised learning manner from video demonstrations and robot trajectories. We show that the robot learns to perform different tasks by associating image features with the corresponding movement primitives of different grasp poses. To evaluate the effectiveness of the proposed learning method, we conduct experiments with a PR2 robot in a simulation environment. The purpose of these experiments is to evaluate the system's ability to perform manipulation tasks.
Cite
CITATION STYLE
Theofanidis, M., Bozcuoglu, A. K., Neumann, M., Cloud, J., Kyrarini, M., Makedon, F., & Beetz, M. (2021). Learning Visuomotor Policies with Deep Movement Primitives. In ACM International Conference Proceeding Series (pp. 140–146). Association for Computing Machinery. https://doi.org/10.1145/3453892.3453899
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.