Multi-modal unsupervised feature learning for RGB-D scene labeling

38Citations
Citations of this article
72Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Most of the existing approaches for RGB-D indoor scene labeling employ hand-crafted features for each modality independently and combine them in a heuristic manner. There has been some attempt on directly learning features from raw RGB-D data, but the performance is not satisfactory. In this paper, we adapt the unsupervised feature learning technique for RGB-D labeling as a multi-modality learning problem. Our learning framework performs feature learning and feature encoding simultaneously which significantly boosts the performance. By stacking basic learning structure, higher-level features are derived and combined with lower-level features for better representing RGB-D data. Experimental results on the benchmark NYU depth dataset show that our method achieves competitive performance, compared with state-of-the-art. © 2014 Springer International Publishing.

Cite

CITATION STYLE

APA

Wang, A., Lu, J., Wang, G., Cai, J., & Cham, T. J. (2014). Multi-modal unsupervised feature learning for RGB-D scene labeling. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8693 LNCS, pp. 453–467). Springer Verlag. https://doi.org/10.1007/978-3-319-10602-1_30

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free