Light Field GAN-based View Synthesis using full 4D information

2Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Light Field (LF) technology offers a truly immersive experience having the potential to revolutionize entertainment, training, education, virtual and augmented reality, gaming, autonomous driving, and digital health. However, one of the main issues when working with LF is the amount of data needed to create a mesmerizing experience with realistic disparity, smooth motion parallax between views. In this paper, we introduce a learning based LF angular super-resolution approach for efficient view synthesis of novel virtual images. This is achieved by taking four corner views and then generating up to five in-between views. Our generative adversarial network approach uses LF spatial and angular information to ensure smooth disparity between the generated and original views. We consider plenoptic, synthetic LF content and camera array implementations which support different baseline settings. Experimental results show that our proposed method outperforms state-of-the-art light field view synthesis techniques, offering novel generated views with high visual quality.

Cite

CITATION STYLE

APA

Wafa, A., & Nasiopoulos, P. (2022). Light Field GAN-based View Synthesis using full 4D information. In Proceedings - CVMP 2022: 19th ACM SIGGRAPH European Conference on Visual Media Production. Association for Computing Machinery, Inc. https://doi.org/10.1145/3565516.3565519

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free