Unstructured Multi-view Depth Estimation Using Mask-Based Multiplane Representation

3Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper presents a novel method, MaskMVS, to solve depth estimation for unstructured multi-view image-pose pairs. In the plane-sweep procedure, the depth planes are sampled by histogram matching that ensures covering the depth range of interest. Unlike other plane-sweep methods, we do not rely on a cost metric to explicitly build the cost volume, but instead infer a multiplane mask representation which regularizes the learning. Compared to many previous approaches, we show that our method is lightweight and generalizes well without requiring excessive training. We outperform the current state-of-the-art and show results on the sun3d, scenes11, MVS, and RGBD test data sets.

Cite

CITATION STYLE

APA

Hou, Y., Solin, A., & Kannala, J. (2019). Unstructured Multi-view Depth Estimation Using Mask-Based Multiplane Representation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11482 LNCS, pp. 54–66). Springer Verlag. https://doi.org/10.1007/978-3-030-20205-7_5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free