Abstract
Painters can successfully recover severely damaged objects, yet current inpainting algorithms still can not achieve this ability. Generally, painters will have a conjecture about the seriously missing image before restoring it, which can be expressed in a text description. This paper imitates the process of painters' conjecture, and proposes to introduce the text description into the image inpainting task for the first time, which provides abundant guidance information for image restoration through the fusion of multimodal features. We propose a multimodal fusion learning method for image inpainting (MMFL). To make better use of text features, we construct an image-adaptive word demand module to reasonably filter the effective text features. We introduce a text guided attention loss and a text-image matching loss to make the network pay more attention to the entities in the text description. Extensive experiments prove that our method can better predict the semantics of objects in the missing regions and generate fine grained textures.
Author supplied keywords
Cite
CITATION STYLE
Lin, Q., Yan, B., Li, J., & Tan, W. (2020). MMFL: Multimodal Fusion Learning for Text-Guided Image Inpainting. In MM 2020 - Proceedings of the 28th ACM International Conference on Multimedia (pp. 1094–1102). Association for Computing Machinery, Inc. https://doi.org/10.1145/3394171.3413982
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.