Abstract
Video super-resolution (SR) aims at generating high-resolution (HR) frames from consecutive low-resolution (LR) frames. The challenge is how to make use of temporal coherence among neighbouring LR frames. Most previous works use motion estimation and compensation based models. However, their performance relies heavily on motion estimation accuracy. In this paper, we propose a multi-scale pyramid 3D convolutional (MP3D) network for video SR, where 3D convolution can explore temporal correlation directly without explicit motion compensation. Specifically, we first apply 3D convolution into a pyramid subnet to extractmulti-scale spatial and temporal features simultaneously from the LR frames, such that it can handle various sizes of motions. We then feed the fused feature maps into an SR reconstruction subnet, where a 3D sub-pixel convolution layer is used for up-sampling. Finally, we append a detail refinement subnet based on the encoder-decoder structure to further enhance texture details of the reconstructed HR frames. Extensive experiments on benchmark datasets and real-world cases show that the proposed MP3D model outperforms state-of-the-art video SR methods in terms of PSNR/SSIM values, visual quality and temporal consistency, respectively.
Author supplied keywords
Cite
CITATION STYLE
Luo, J., Huang, S., & Yuan, Y. (2020). Video Super-Resolution using Multi-scale Pyramid 3D Convolutional Networks. In MM 2020 - Proceedings of the 28th ACM International Conference on Multimedia (pp. 1882–1890). Association for Computing Machinery, Inc. https://doi.org/10.1145/3394171.3413587
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.