Visual-inertial 2D feature tracking based on an affine photometric model

3Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The robust tracking of point features throughout an image sequence is one fundamental stage in many different computer vision algorithms (e.g. visual modelling, object tracking, etc.). In most cases, this tracking is realised by means of a feature detection step and then a subsequent re-identification of the same feature point, based on some variant of a template matching algorithm. Without any auxiliary knowledge about the movement of the camera, actual tracking techniques are only robust for relatively moderate frame-to-frame feature displacements. This paper presents a framework for a visual-inertial feature tracking scheme, where images and measurements of an inertial measurement unit (IMU) are fused in order to allow a wider range of camera movements. The inertial measurements are used to estimate the visual appearance of a feature’s local neighbourhood based on a affine photometric warping model.

Cite

CITATION STYLE

APA

Aufderheide, D., Edwards, G., & Krybus, W. (2015). Visual-inertial 2D feature tracking based on an affine photometric model. Lecture Notes in Computational Vision and Biomechanics, 19, 297–317. https://doi.org/10.1007/978-3-319-13407-9_18

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free