Convolution neural network application for road asset detection and classification in LiDAR point cloud

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Self-driving cars (or autonomous cars) can sense and navigate through an environment without any driver intervention. To achieve this task, they rely on vision sensors working in tandem with accurate algorithms to detect movable and non-movable objects around them. These vision sensors typically include cameras to identify static and non-static objects, Radio Detection and Ranging (RADAR) to detect the speed of the moving objects using Doppler effect and Light Detection and Ranging (LiDAR) to detect the distance to objects. In this paper, we explore a new usage of LiDAR data to classify static objects on the road. We present a pipeline to classify point cloud data grouped in volumetric pixels (voxels). We introduce a novel approach to point cloud data representation for processing within Convolution Neural Networks (CNN). Results show an accuracy exceeding 90% in the detection and classification of road edges, solid and broken lane markings, bike lanes, and lane center lines. Our data pipeline is capable of processing up to 20,000 points per 900ms on a server equipped with 2 Intel Xeon processors 8-core CPU with HyperThreading for a total of 32 threads and 2 NVIDIA Tesla K40 GPUs. Our model outperforms by 2% ResNet applied to camera images for the same road.

Cite

CITATION STYLE

APA

Sakr, G. E., Eido, L., & Maarawi, C. (2018). Convolution neural network application for road asset detection and classification in LiDAR point cloud. In Advances in Intelligent Systems and Computing (Vol. 868, pp. 86–100). Springer Verlag. https://doi.org/10.1007/978-3-030-01054-6_6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free