Most of the existing approaches for RGB-D indoor scene labeling employ hand-crafted features for each modality independently and combine them in a heuristic manner. There has been some attempt on directly learning features from raw RGB-D data, but the performance is not satisfactory. In this paper, we adapt the unsupervised feature learning technique for RGB-D labeling as a multi-modality learning problem. Our learning framework performs feature learning and feature encoding simultaneously which significantly boosts the performance. By stacking basic learning structure, higher-level features are derived and combined with lower-level features for better representing RGB-D data. Experimental results on the benchmark NYU depth dataset show that our method achieves competitive performance, compared with state-of-the-art. © 2014 Springer International Publishing.
CITATION STYLE
Wang, A., Lu, J., Wang, G., Cai, J., & Cham, T. J. (2014). Multi-modal unsupervised feature learning for RGB-D scene labeling. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8693 LNCS, pp. 453–467). Springer Verlag. https://doi.org/10.1007/978-3-319-10602-1_30
Mendeley helps you to discover research relevant for your work.