Monocular depth estimation with joint attention feature distillation and wavelet-based loss function

17Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

Depth estimation is a crucial component in many 3D vision applications. Monocular depth estimation is gaining increasing interest due to flexible use and extremely low system requirements, but inherently ill-posed and ambiguous characteristics still cause unsatisfactory estimation results. This paper proposes a new deep convolutional neural network for monocular depth estimation. The network applies joint attention feature distillation and wavelet-based loss function to recover the depth information of a scene. Two improvements were achieved, compared with previous methods. First, we combined feature distillation and joint attention mechanisms to boost feature modulation discrimination. The network extracts hierarchical features using a progressive feature distillation and refinement strategy and aggregates features using a joint attention operation. Second, we adopted a wavelet-based loss function for network training, which improves loss function effectiveness by obtaining more structural details. The experimental results on challenging indoor and outdoor benchmark datasets verified the proposed method’s superiority compared with current state-of-the-art methods.

Cite

CITATION STYLE

APA

Liu, P., Zhang, Z., Meng, Z., & Gao, N. (2021). Monocular depth estimation with joint attention feature distillation and wavelet-based loss function. Sensors (Switzerland), 21(1), 1–21. https://doi.org/10.3390/s21010054

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free