CanvasGAN: A Simple Baseline for Text to Image Generation by Incrementally Patching a Canvas

3Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We propose a new recurrent generative model for generating images from text captions while attending on specific parts of text captions. Our model creates images by incrementally adding patches on a “canvas” while attending on words from text caption at each timestep. Finally, the canvas is passed through an upscaling network to generate images. We also introduce a new method for generating visual-semantic sentence embeddings based on self-attention over text. We compare our model’s generated images with those generated by Reed’s model and show that our model is a stronger baseline for text to image generation tasks.

Cite

CITATION STYLE

APA

Singh, A., & Agrawal, S. (2020). CanvasGAN: A Simple Baseline for Text to Image Generation by Incrementally Patching a Canvas. In Advances in Intelligent Systems and Computing (Vol. 943, pp. 86–98). Springer Verlag. https://doi.org/10.1007/978-3-030-17795-9_7

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free