DNN Based Camera and Lidar Fusion Framework for 3D Object Recognition

1Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

A 3-stages deep neural network (DNN) based camera and lidar fusion framework for 3D objects recognition is proposed in this paper. First, to leverage the high resolution of camera and 3D spatial information of Lidar, region proposal network (RPN) is trained to generate proposals from RGB image feature maps and bird-view (BV) feature maps, these proposals are then lifted into 3D proposals. Then, a segmentation network is used to extract object points directly from points inside these 3D proposals. At last, 3D object bounding box instances are extracted from the interested object points by an estimation network followed after a translation by a light-weight TNet, which is a special supervised spatial transformer network (STN). Experiment results show that this proposed 3d object recognition framework can produce considerable result as the other leading methods on KITTI 3D object detection datasets.

Cite

CITATION STYLE

APA

Zhang, K., Wang, S. J., Ji, L., & Wang, C. (2020). DNN Based Camera and Lidar Fusion Framework for 3D Object Recognition. In Journal of Physics: Conference Series (Vol. 1518). Institute of Physics Publishing. https://doi.org/10.1088/1742-6596/1518/1/012044

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free