MVSalNet: Multi-view Augmentation for RGB-D Salient Object Detection

17Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

RGB-D salient object detection (SOD) enjoys significant advantages in understanding 3D geometry of the scene. However, the geometry information conveyed by depth maps are mostly under-explored in existing RGB-D SOD methods. In this paper, we propose a new framework to address this issue. We augment the input image with multiple different views rendered using the depth maps, and cast the conventional single-view RGB-D SOD into a multi-view setting. Since different views captures complementary context of the 3D scene, the accuracy can be significantly improved through multi-view aggregation. We further design a multi-view saliency detection network (MVSalNet), which firstly performs saliency prediction for each view separately and incorporates multi-view outputs through a fusion model to produce final saliency prediction. A dynamic filtering module is also designed to facilitate more effective and flexible feature extraction. Extensive experiments on 6 widely used datasets demonstrate that our approach compares favorably against state-of-the-art approaches.

Cite

CITATION STYLE

APA

Zhou, J., Wang, L., Lu, H., Huang, K., Shi, X., & Liu, B. (2022). MVSalNet: Multi-view Augmentation for RGB-D Salient Object Detection. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13689 LNCS, pp. 270–287). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-19818-2_16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free