Object Localization and Depth Estimation for Eye-in-Hand Manipulator Using Mono Camera

16Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper proposes the object localization and depth estimation to select and set goals for robots via machine vision. An algorithm based on a deep region-based convolution neural network (R-CNN) will recognize targets and non-targets. After the targets are recognized, we employed both the k-nearest neighbors (kNN) and the fuzzy inference system (FIS) to localize two-dimension (2D) positions. Moreover, based on the field of view (FoV) and a disparity map, the depth is estimated by a mono camera mounted on the end-effector with an eye-in-hand manipulator structure. Although using a single mono camera, the system can easily find the camera baseline by only shifting the end-effector a few millimeters towards the x-axis. Thus, we can obtain and identify the depth of the layered environment in 3D points, which form a dataset to recognize the junction box covers on the table. Experimental tests confirmed that the algorithm could accurately distinguish junction box covers or non-targets and could estimate whether the targets are within the depth for grasping by three-finger grippers. Furthermore, the proposed optimized depth error of -0.0005%, and localization method could precisely position the junction box cover with recognizing and picking error rates 0.993 and 98.529% respectively.

Cite

CITATION STYLE

APA

Muslikhin, Horng, J. R., Yang, S. Y., & Wang, M. S. (2020). Object Localization and Depth Estimation for Eye-in-Hand Manipulator Using Mono Camera. IEEE Access, 8, 121765–121779. https://doi.org/10.1109/ACCESS.2020.3006843

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free