Season-Invariant Semantic Segmentation with a Deep Multimodal Network

22Citations
Citations of this article
36Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Semantic scene understanding is a useful capability for autonomous vehicles operating in off-roads. While cameras are the most common sensor used for semantic classification, the performance of methods using camera imagery may suffer when there is significant variation between the train and testing sets caused by illumination, weather, and seasonal variations. On the other hand, 3D information from active sensors such as LiDAR is comparatively invariant to these factors, which motivates us to investigate whether it can be used to improve performance in this scenario. In this paper, we propose a novel multimodal Convolutional Neural Network (CNN) architecture consisting of two streams, 2D and 3D, which are fused by projecting 3D features to image space to achieve a robust pixelwise semantic segmentation. We evaluate our proposed method in a novel off-road terrain classification benchmark, and show a 25% improvement in mean Intersection over Union (IoU) of navigation-related semantic classes, relative to an image-only baseline.

Cite

CITATION STYLE

APA

Kim, D. K., Maturana, D., Uenoyama, M., & Scherer, S. (2018). Season-Invariant Semantic Segmentation with a Deep Multimodal Network. In Springer Proceedings in Advanced Robotics (Vol. 5, pp. 255–270). Springer Science and Business Media B.V. https://doi.org/10.1007/978-3-319-67361-5_17

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free