Purpose of Review: We investigate the first use of deep networks for victim identification in Urban Search and Rescue (USAR). Moreover, we provide the first experimental comparison of single-stage and two-stage networks for body part detection, for cases of partial occlusions and varying illumination, on a RGB-D dataset obtained by a mobile robot navigating cluttered USAR-like environments. Recent Findings: We considered the single-stage detectors Single Shot Multi-box Detector, You Only Look Once, and RetinaNet and the two-stage Feature Pyramid Network detector. Experimental results show that RetinaNet has the highest mean average precision (77.66%) and recall (86.98%) for detecting victims with body part occlusions in different lighting conditions. Summary: End-to-end deep networks can be used for finding victims in USAR by autonomously extracting RGB-D image features from sensory data. We show that RetinaNet using RGB-D is robust to body part occlusions and low-lighting conditions and outperforms other detectors regardless of the image input type.
CITATION STYLE
Fung, A., Wang, L. Y., Zhang, K., Nejat, G., & Benhabib, B. (2020). Using Deep Learning to Find Victims in Unknown Cluttered Urban Search and Rescue Environments. Current Robotics Reports, 1(3), 105–115. https://doi.org/10.1007/s43154-020-00011-8
Mendeley helps you to discover research relevant for your work.