Abstract
Salient object detection (SOD) aims to identify and locate the most attractive regions in an image, which has been widely used in various vision tasks. Recent years, with the development of RGBD sensor technology, depth information of scenes becomes available for image understanding. In this paper, we systematically investigate and evaluate on how to integrate depth cues in a pre-trained deep network and learn informative features for SOD. First, we propose a CNN-based cross-modal transfer learning, which learn knowledge from sufficient labeled RGB salient object datasets and guide the depth domain feature extraction. Then we design a feature fusion module to fuse the complementary features in a hierarchical manner. At last, the final saliency map is obtained by integrating multi-scale information step by step. Extensive experiments on five popular RGBD benchmark datasets demonstrate that our proposed approach achieves significant improvements and outperforms the state-of-the-art methods.
Author supplied keywords
Cite
CITATION STYLE
Xiao, F., Li, B., Peng, Y., Cao, C., Hu, K., & Gao, X. (2020). Multi-Modal Weights Sharing and Hierarchical Feature Fusion for RGBD Salient Object Detection. IEEE Access, 8, 26602–26611. https://doi.org/10.1109/ACCESS.2020.2971509
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.