Synthesizing light field from a single image with variable MPI and two network fusion

35Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.

Abstract

We propose a learning-based approach to synthesize a light field with a small baseline from a single image. We synthesize the novel view images by first using a convolutional neural network (CNN) to promote the input image into a layered representation of the scene. We extend the multiplane image (MPI) representation by allowing the disparity of the layers to be inferred from the input image. We show that, compared to the original MPI representation, our representation models the scenes more accurately. Moreover, we propose to handle the visible and occluded regions separately through two parallel networks. The synthesized images using these two networks are then combined through a soft visibility mask to generate the final results. To effectively train the networks, we introduce a large-scale light field dataset of over 2,000 unique scenes containing a wide range of objects. We demonstrate that our approach synthesizes high-quality light fields on a variety of scenes, better than the state-of-the-art methods.

Cite

CITATION STYLE

APA

Li, Q., & Kalantari, N. K. (2020). Synthesizing light field from a single image with variable MPI and two network fusion. ACM Transactions on Graphics, 39(6). https://doi.org/10.1145/3414685.3417785

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free