LightGAN: A Deep Generative Model for Light Field Reconstruction

23Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

A light field image captured by a plenoptic camera can be considered a sampling of light distribution within a given space. However, with the limited pixel count of the sensor, the acquisition of a high-resolution sample often comes at the expense of losing parallax information. In this work, we present a learning-based generative framework to overcome such tradeoff by directly simulating the light field distribution. An important module of our model is the high-dimensional residual block, which fully exploits the spatio-angular information. By directly learning the distribution, our approach can generate both high-quality sub-aperture images and densely-sampled light fields. Experimental results on both real-world and synthetic datasets demonstrate that the proposed method outperforms other state-of-the-art approaches and achieves visually more realistic results.

Cite

CITATION STYLE

APA

Meng, N., Ge, Z., Zeng, T., & Lam, E. Y. (2020). LightGAN: A Deep Generative Model for Light Field Reconstruction. IEEE Access, 8, 116052–116063. https://doi.org/10.1109/ACCESS.2020.3004477

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free