Abstract
Video inpainting is a task of filling missing regions in videos. In this task, it is important to efficiently use information from other frames and generate plausible results with sufficient temporal consistency. In this paper, we present a video inpainting method jointly using affine transformation and deformable convolutions for frame alignment. The former is responsible for frame-scale rough alignment and the latter performs pixel-level fine alignment. Our model does not depend on 3D convolutions, which limits the temporal window, or troublesome flow estimation. The proposed method achieves improved object removal results and better PSNR and SSIM values compared with previous learning-based methods.
Author supplied keywords
Cite
CITATION STYLE
Hara, Y., Wang, X., & Yamasaki, T. (2021). Video inpainting by frame alignment with deformable convolution. IEICE Transactions on Information and Systems, E104D(8), 1349–1358. https://doi.org/10.1587/transinf.2020EDP7194
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.