Terrestrial platforms for 3D reconstruction typically combine several data acquisition systems such as lasers, cameras and inertial systems. However the geometrical combination of different sensors requires their calibration and data fusion. These topics are an important task for vision-based systems since it estimates the values of sensor model parameters, such as cameras. The uncertainty of these parameters must be known in order to evaluate the error of the final calibration and their applications. The aim of this paper is to present a method to compute the calibration of both sensors. A new calibration pattern, visible to both sensors is used. Correspondence is obtained between each laser point and its position in the image, the texture and color of each point of LIDAR can be know. Experimental results are presented for data collected with the platform integrated with a 3D laser scanner and a panoramic camera system. © Springer-Verlag Berlin Heidelberg 2014.
CITATION STYLE
García-Moreno, A. I., Gonzalez-Barbosa, J. J., Ornelas-Rodriguez, F. J., & Hurtado-Ramos, J. B. (2014). Uncertainty analysis of LIDAR and panoramic camera calibration. In Lecture Notes in Electrical Engineering (Vol. 282 LNEE, pp. 317–327). Springer Verlag. https://doi.org/10.1007/978-3-642-41968-3_33
Mendeley helps you to discover research relevant for your work.