Spatial uncertainty model for visual features using a KinectTM sensor

65Citations
Citations of this article
78Readers
Mendeley users who have this article in their library.

Abstract

This study proposes a mathematical uncertainty model for the spatial measurement of visual features using Kinect™ sensors. This model can provide qualitative and quantitative analysis for the utilization of KinectTM sensors as 3D perception sensors. In order to achieve this objective, we derived the propagation relationship of the uncertainties between the disparity image space and the real Cartesian space with the mapping function between the two spaces. Using this propagation relationship, we obtained the mathematical model for the covariance matrix of the measurement error, which represents the uncertainty for spatial position of visual features from KinectTM sensors. In order to derive the quantitative model of spatial uncertainty for visual features, we estimated the covariance matrix in the disparity image space using collected visual feature data. Further, we computed the spatial uncertainty information by applying the covariance matrix in the disparity image space and the calibrated sensor parameters to the proposed mathematical model. This spatial uncertainty model was verified by comparing the uncertainty ellipsoids for spatial covariance matrices and the distribution of scattered matching visual features. We expect that this spatial uncertainty model and its analyses will be useful in various KinectTM sensor applications. © 2012 by the authors; licensee MDPI, Basel, Switzerland.

Cite

CITATION STYLE

APA

Park, J. H., Shin, Y. D., Bae, J. H., & Baeg, M. H. (2012). Spatial uncertainty model for visual features using a KinectTM sensor. Sensors (Switzerland), 12(7), 8640–8662. https://doi.org/10.3390/s120708640

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free