Uncertainty analysis of LIDAR and panoramic camera calibration

2Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Terrestrial platforms for 3D reconstruction typically combine several data acquisition systems such as lasers, cameras and inertial systems. However the geometrical combination of different sensors requires their calibration and data fusion. These topics are an important task for vision-based systems since it estimates the values of sensor model parameters, such as cameras. The uncertainty of these parameters must be known in order to evaluate the error of the final calibration and their applications. The aim of this paper is to present a method to compute the calibration of both sensors. A new calibration pattern, visible to both sensors is used. Correspondence is obtained between each laser point and its position in the image, the texture and color of each point of LIDAR can be know. Experimental results are presented for data collected with the platform integrated with a 3D laser scanner and a panoramic camera system. © Springer-Verlag Berlin Heidelberg 2014.

Cite

CITATION STYLE

APA

García-Moreno, A. I., Gonzalez-Barbosa, J. J., Ornelas-Rodriguez, F. J., & Hurtado-Ramos, J. B. (2014). Uncertainty analysis of LIDAR and panoramic camera calibration. In Lecture Notes in Electrical Engineering (Vol. 282 LNEE, pp. 317–327). Springer Verlag. https://doi.org/10.1007/978-3-642-41968-3_33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free