Robust 3D Object Detection from LiDAR Point Cloud Data with Spatial Information Aggregation

4Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Current 3D object detectors from Bird’s Eye View (BEV) LiDAR point cloud data rely on Convolutional Neural Networks (CNNs), which have originally been designed for camera images. Therefore, they look for the same target features, regardless of the position of the objects with respect to the sensor. Discarding this spatial information makes 3D object detection unreliable and not robust, because objects in LiDAR point clouds contain distance dependent features. The position of a group of points can be decisive to know if they represent an object or not. To solve this, we propose a network extension called FeatExt operation that enables the model to be aware of both the target objects features and their spatial location. FeatExt operation expands a group of feature maps extracted from a BEV representation to include the distance to a specific position of interest in the scene, in this case the distance with respect to the LiDAR. When adding the proposed operation to a baseline network in an intermediate fusion fashion, it shows up to an 8.9 average precision boost in the KITTI BEV benchmark. Our proposal can be easily added to improve existing object detection networks.

Cite

CITATION STYLE

APA

Aranjuelo, N., Engels, G., Unzueta, L., Arganda-Carreras, I., Nieto, M., & Otaegui, O. (2021). Robust 3D Object Detection from LiDAR Point Cloud Data with Spatial Information Aggregation. In Advances in Intelligent Systems and Computing (Vol. 1268 AISC, pp. 813–823). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-57802-2_78

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free