E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context

8Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, the image-wise implicit neural representation of videos, NeRV, has gained popularity for its promising results and swift speed compared to regular pixel-wise implicit representations. However, the redundant parameters within the network structure can cause a large model size when scaling up for desirable performance. The key reason of this phenomenon is the coupled formulation of NeRV, which outputs the spatial and temporal information of video frames directly from the frame index input. In this paper, we propose E-NeRV, which dramatically expedites NeRV by decomposing the image-wise implicit neural representation into separate spatial and temporal context. Under the guidance of this new formulation, our model greatly reduces the redundant model parameters, while retaining the representation ability. We experimentally find that our method can improve the performance to a large extent with fewer parameters, resulting in a more than 8 × faster speed on convergence. Code is available at https://github.com/kyleleey/E-NeRV.

Cite

CITATION STYLE

APA

Li, Z., Wang, M., Pi, H., Xu, K., Mei, J., & Liu, Y. (2022). E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13695 LNCS, pp. 267–284). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-19833-5_16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free