Using Deep Learning to Find Victims in Unknown Cluttered Urban Search and Rescue Environments

  • Fung A
  • Wang L
  • Zhang K
  • et al.
N/ACitations
Citations of this article
28Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Purpose of Review: We investigate the first use of deep networks for victim identification in Urban Search and Rescue (USAR). Moreover, we provide the first experimental comparison of single-stage and two-stage networks for body part detection, for cases of partial occlusions and varying illumination, on a RGB-D dataset obtained by a mobile robot navigating cluttered USAR-like environments. Recent Findings: We considered the single-stage detectors Single Shot Multi-box Detector, You Only Look Once, and RetinaNet and the two-stage Feature Pyramid Network detector. Experimental results show that RetinaNet has the highest mean average precision (77.66%) and recall (86.98%) for detecting victims with body part occlusions in different lighting conditions. Summary: End-to-end deep networks can be used for finding victims in USAR by autonomously extracting RGB-D image features from sensory data. We show that RetinaNet using RGB-D is robust to body part occlusions and low-lighting conditions and outperforms other detectors regardless of the image input type.

Cite

CITATION STYLE

APA

Fung, A., Wang, L. Y., Zhang, K., Nejat, G., & Benhabib, B. (2020). Using Deep Learning to Find Victims in Unknown Cluttered Urban Search and Rescue Environments. Current Robotics Reports, 1(3), 105–115. https://doi.org/10.1007/s43154-020-00011-8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free