Proceedings on Engineering Sciences (Dec 2024)

TEXT DESCRIPTION TO IMAGE GENERATION USING GENERATIVE ADVERSARIAL NETWORK

  • Kayal Padmanandam ,
  • Yeshasvi Mogula,
  • Nikitha Pitla

DOI
https://doi.org/10.24874/PES.SI.25.03b.015
Journal volume & issue
Vol. 6, no. 4
pp. 1829 – 1836

Abstract

Read online

The progression of translating text into images has been an imperative topic of research. The significant challenges arise from translating visual to textual information and vice versa. High-quality images can be generated from text using a Generative Adversarial Network (GAN), however, there are challenges associated with accurately portraying the content of the sentence provided to the model. Text-to-image conversion strategies can produce examples that closely reflect the descriptions' intended meaning. The user descriptions may however lack crucial details. To create the conditioned text descriptions, this study employs an Attention-Generative Adversarial Network to generate 256*256-pixel images that are image-sensitive. In the initial phase of GAN sketches, the input text descriptions serve solely to inform the basic form and color scheme of the devices. The information gleaned from the first stage, along with the textual descriptions, is fed into a GAN which generates images with high resolution and realistic detail. The conditional GAN training may be stabilized using conditioning augmentation, and the generated samples can have higher quality. Using Style based Generator, samples for each style of the image can be drawn. The proposed system can generate photorealistic visuals of an object when the user inputs the textual descriptions in the application’s GUI.

Keywords