Polarization-driven semantic segmentation via efficient attention-bridged fusion

  • Xiang K
  • Yang K
  • Wang K
65Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Semantic segmentation (SS) is promising for outdoor scene perception in safety-critical applications like autonomous vehicles, assisted navigation and so on. However, traditional SS is primarily based on RGB images, which limits the reliability of SS in complex outdoor scenes, where RGB images lack necessary information dimensions to fully perceive unconstrained environments. As a preliminary investigation, we examine SS in an unexpected obstacle detection scenario, which demonstrates the necessity of multimodal fusion. Thereby, in this work, we present EAFNet, an Efficient Attention-bridged Fusion Network, to exploit complementary information coming from different optical sensors. Specifically, we incorporate polarization sensing to obtain supplementary information, considering its optical characteristics for robust representation of diverse materials. By using a single-shot polarization sensor, we build the first RGB-P dataset which consists of 394 annotated pixel-aligned RGB-polarization images. A comprehensive variety of experiments shows the effectiveness of EAFNet to fuse polarization and RGB information, as well as its flexibility to be adapted to other sensor combination scenarios.

Cite

CITATION STYLE

APA

Xiang, K., Yang, K., & Wang, K. (2021). Polarization-driven semantic segmentation via efficient attention-bridged fusion. Optics Express, 29(4), 4802. https://doi.org/10.1364/oe.416130

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free