Visual Camera Re-Localization from RGB and RGB-D Images Using DSAC

86Citations
Citations of this article
96Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We describe a learning-based system that estimates the camera position and orientation from a single input image relative to a known environment. The system is flexible w.r.t. the amount of information available at test and at training time, catering to different applications. Input images can be RGB-D or RGB, and a 3D model of the environment can be utilized for training but is not necessary. In the minimal case, our system requires only RGB images and ground truth poses at training time, and it requires only a single RGB image at test time. The framework consists of a deep neural network and fully differentiable pose optimization. The neural network predicts so called scene coordinates, i.e., dense correspondences between the input image and 3D scene space of the environment. The pose optimization implements robust fitting of pose parameters using differentiable RANSAC (DSAC) to facilitate end-to-end training. The system, an extension of DSAC++ and referred to as DSAC*, achieves state-of-the-art accuracy on various public datasets for RGB-based re-localization, and competitive accuracy for RGB-D based re-localization.

Cite

CITATION STYLE

APA

Brachmann, E., & Rother, C. (2022). Visual Camera Re-Localization from RGB and RGB-D Images Using DSAC. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(9), 5847–5865. https://doi.org/10.1109/TPAMI.2021.3070754

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free