Video frame interpolation via deformable separable convolution

105Citations
Citations of this article
40Readers
Mendeley users who have this article in their library.

Abstract

Learning to synthesize non-existing frames from the original consecutive video frames is a challenging task. Recent kernel-based interpolation methods predict pixels with a single convolution process to replace the dependency of optical flow. However, when scene motion is larger than the predefined kernel size, these methods yield poor results even though they take thousands of neighboring pixels into account. To solve this problem in this paper, we propose to use deformable separable convolution (DSepConv) to adaptively estimate kernels, offsets and masks to allow the network to obtain information with much fewer but more relevant pixels. In addition, we show that the kernel-based methods and conventional flow-based methods are specific instances of the proposed DSepConv. Experimental results demonstrate that our method significantly outperforms the other kernel-based interpolation methods and shows strong performance on par or even better than the state-of-the-art algorithms both qualitatively and quantitatively.

Cite

CITATION STYLE

APA

Cheng, X., & Chen, Z. (2020). Video frame interpolation via deformable separable convolution. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 10607–10614). AAAI press. https://doi.org/10.1609/aaai.v34i07.6634

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free