Using Human Gaze to Improve Robustness against Irrelevant Objects in Robot Manipulation Tasks

19Citations
Citations of this article
33Readers
Mendeley users who have this article in their library.

Abstract

Deep imitation learning enables the learning of complex visuomotor skills from raw pixel inputs. However, this approach suffers from the problem of overfitting to the training images. The neural network can easily be distracted by task-irrelevant objects. In this letter, we use the human gaze measured by a head-mounted eye tracking device to discard task-irrelevant visual distractions. We propose a mixture density network-based behavior cloning method that learns to imitate the human gaze. The model predicts gaze positions from raw pixel images and crops images around the predicted gazes. Only these cropped images are used to compute the output action. This cropping procedure can remove visual distractions because the gaze is rarely fixated on task-irrelevant objects. This robustness against irrelevant objects can improve the manipulation performance of robots in scenarios where task-irrelevant objects are present. We evaluated our model on four manipulation tasks designed to test the robustness of the model to irrelevant objects. The results indicate that the proposed model can predict the locations of task-relevant objects from gaze positions, is robust to task-irrelevant objects, and exhibits impressive manipulation performance especially in multi-object handling.

References Powered by Scopus

Deep residual learning for image recognition

178632Citations
N/AReaders
Get full text

You only look once: Unified, real-time object detection

38749Citations
N/AReaders
Get full text

Fast R-CNN

23561Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Human–robot collaboration in industrial environments: A literature review on non-destructive disassembly

158Citations
N/AReaders
Get full text

Transformer-based deep imitation learning for dual-arm robot manipulation

40Citations
N/AReaders
Get full text

Gaze-Based Dual Resolution Deep Imitation Learning for High-Precision Dexterous Robot Manipulation

22Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Kim, H., Ohmura, Y., & Kuniyoshi, Y. (2020). Using Human Gaze to Improve Robustness against Irrelevant Objects in Robot Manipulation Tasks. IEEE Robotics and Automation Letters, 5(3), 4415–4422. https://doi.org/10.1109/LRA.2020.2998410

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 10

71%

Researcher 3

21%

Professor / Associate Prof. 1

7%

Readers' Discipline

Tooltip

Engineering 13

68%

Computer Science 4

21%

Decision Sciences 1

5%

Medicine and Dentistry 1

5%

Save time finding and organizing research with Mendeley

Sign up for free