Real-time indoor scene description for the visually impaired using autoencoder fusion strategies with visible cameras

12Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

This paper describes three coarse image description strategies, which are meant to promote a rough perception of surrounding objects for visually impaired individuals, with application to indoor spaces. The described algorithms operate on images (grabbed by the user, by means of a chest-mounted camera), and provide in output a list of objects that likely exist in his context across the indoor scene. In this regard, first, different colour, texture, and shape-based feature extractors are generated, followed by a feature learning step by means of AutoEncoder (AE) models. Second, the produced features are fused and fed into a multilabel classifier in order to list the potential objects. The conducted experiments point out that fusing a set of AE-learned features scores higher classification rates with respect to using the features individually. Furthermore, with respect to reference works, our method: (i) yields higher classification accuracies, and (ii) runs (at least four times) faster, which enables a potential full real-time application.

Cite

CITATION STYLE

APA

Malek, S., Melgani, F., Mekhalfi, M. L., & Bazi, Y. (2017). Real-time indoor scene description for the visually impaired using autoencoder fusion strategies with visible cameras. Sensors (Switzerland), 17(11). https://doi.org/10.3390/s17112641

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free