Biologically inspired online learning of visual autonomous driving

4Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

While autonomously driving systems accumulate more and more sensors as well as highly specialized visual features and engineered solutions, the human visual system provides evidence that visual input and simple low level image features are sufficient for successful driving. In this paper we propose extensions (non-linear update and coherence weighting) to one of the simplest biologically inspired learning schemes (Hebbian learning). We show that this is sufficient for online learning of visual autonomous driving, where the system learns to directly map low level image features to control signals. After the initial training period, the system seamlessly continues autonomously. This extended Hebbian algorithm, qHebb, has constant bounds on time and memory complexity for training and evaluation, independent of the number of training samples presented to the system. Further, the proposed algorithm compares favorably to state of the art engineered batch learning algorithms.

Cite

CITATION STYLE

APA

Öfjäll, K., & Felsberg, M. (2014). Biologically inspired online learning of visual autonomous driving. In BMVC 2014 - Proceedings of the British Machine Vision Conference 2014. British Machine Vision Association, BMVA. https://doi.org/10.5244/c.28.94

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free