Learning visual representations for interactive systems

0Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We describe two quite different methods for associating action parameters to visual percepts. Our RLVC algorithm performs reinforcement learning directly on the visual input space. To make this very large space manageable, RLVC interleaves the reinforcement learner with a supervised classification algorithm that seeks to split perceptual states so as to reduce perceptual aliasing. This results in an adaptive discretization of the perceptual space based on the presence or absence of visual features. Its extension RLJC also handles continuous action spaces. In contrast to the minimalistic visual representations produced by RLVC and RLJC, our second method learns structural object models for robust object detection and pose estimation by probabilistic inference. To these models, the method associates grasp experiences autonomously learned by trial and error. These experiences form a nonparametric representation of grasp success likelihoods over gripper poses, which we call a grasp density. Thus, object detection in a novel scene simultaneously produces suitable grasping options. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Piater, J., Jodogne, S., Detry, R., Kraft, D., Krüger, N., Krömer, O., & Peters, J. (2011). Learning visual representations for interactive systems. In Springer Tracts in Advanced Robotics (Vol. 70, pp. 399–416). https://doi.org/10.1007/978-3-642-19457-3_24

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free