ATDN vSLAM: An All-Through Deep Learning-Based Solution for Visual Simultaneous Localization and Mapping

2Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.

Abstract

In this paper, a novel solution is introduced for visual Simultaneous Localization and Mapping (vSLAM) that is built up of Deep Learning components. The proposed architecture is a highly modular framework in which each component offers state of the art results in their respective fields of vision-based Deep Learning solutions. The paper shows that with the synergic integration of these individual building blocks, a functioning and efficient all-through deep neural (ATDN) vSLAM system can be created. The Embedding Distance Loss function is introduced and using it the ATDN architecture is trained. The resulting system managed to achieve 4.4% translation and 0.0176 deg/m rotational error on a subset of the KITTI dataset. The proposed architecture can be used for efficient and lowlatency autonomous driving (AD) aiding database creation as well as a basis for autonomous vehicle (AV) control.

Cite

CITATION STYLE

APA

Szántó, M., Bogár, G. R., & Vajta, L. (2022). ATDN vSLAM: An All-Through Deep Learning-Based Solution for Visual Simultaneous Localization and Mapping. Periodica Polytechnica Electrical Engineering and Computer Science, 66(3), 236–247. https://doi.org/10.3311/PPee.20437

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free