Multiple feature fusion based on co-training approach and time regularization for place classification in wearable video

6Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The analysis of video acquired with a wearable camera is a challenge that multimedia community is facing with the proliferation of such sensors in various applications. In this paper, we focus on the problem of automatic visual place recognition in a weakly constrained environment, targeting the indexing of video streams by topological place recognition. We propose to combine several machine learning approaches in a time regularized framework for image-based place recognition indoors. The framework combines the power of multiple visual cues and integrates the temporal continuity information of video. We extend it with computationally efficient semisupervised method leveraging unlabeled video sequences for an improved indexing performance. The proposed approach was applied on challenging video corpora. Experiments on a public and a real-world video sequence databases show the gain brought by the different stages of the method. © 2013 Vladislavs Dovgalecs et al.

Cite

CITATION STYLE

APA

Dovgalecs, V., Mégret, R., & Berthoumieu, Y. (2013). Multiple feature fusion based on co-training approach and time regularization for place classification in wearable video. Advances in Multimedia, 2013. https://doi.org/10.1155/2013/175064

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free