Abstract
Slip detection is essential for robots to make robust grasping and fine manipulation. In this paper, a novel dynamic vision-based finger system for slip detection and suppression is proposed. We also present a baseline and feature based approach to detect object slips under illumination and vibration uncertainty. A threshold method is devised to autonomously sample noise and object feature events in real-time to improve slip detection and suppression. Moreover, a fuzzy based suppression strategy using incipient slip feedback is proposed for regulating the grip force. A comprehensive experimental study of our proposed approaches under uncertainty and system for high-performance precision manipulation are presented. We also propose a slip metric to evaluate such performance quantitatively. For a class of objects, results indicate that the system can effectively detect incipient slip events at a sampling rate of 2kHz ( $\Delta t = 500\mu s$ ) and suppress them before a gross slip occurs. The event-based approach holds promises to high precision manipulation task requirement in industrial manufacturing and household services.
Author supplied keywords
Cite
CITATION STYLE
Muthusamy, R., Huang, X., Zweiri, Y., Seneviratne, L., & Gan, D. (2020). Neuromorphic Event-Based Slip Detection and Suppression in Robotic Grasping and Manipulation. IEEE Access, 8, 153364–153384. https://doi.org/10.1109/ACCESS.2020.3017738
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.