Text-to-Image Generation using Generative AI

  • S R
N/ACitations
Citations of this article
23Readers
Mendeley users who have this article in their library.

Abstract

Abstract—This survey reviews text-to-image generation by using different approaches. One of the approaches identified in this study is Cross-modal Semantic Matching Generative Adversarial Networks (CSM-GAN), which is used to increase semantic consistency between text descriptions and synthesised pictures for fine-grained text- to-image creation. This includes other two modules, Text Encoder Module and Textual-Visual Semantic Matching Module. We further discussed about Imagen which is a text- to-image diffusion model with photorealism and deep language understanding, which is used on the COCO dataset. Lastly, we discussed about Text to image synthesis used to automates image generation using conditional generative models and GAN, enhancing artificial intelligence and deep learning. Based on these approaches we present a review of text to image generation using generative AI. Keywords— Generative AI, Diffusion model, Text-to- image, Imagen, CSM-GAN

Cite

CITATION STYLE

APA

S, R. (2023). Text-to-Image Generation using Generative AI. INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT, 07(08). https://doi.org/10.55041/ijsrem25320

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free