Semantic-aware Mapping for Text-to-Image Synthesis
Main Article Content
Abstract
This study explores the fast-progressing domain of Text-to-Image (T2I) synthesis, which aims to bridge the gap between language and visual comprehension. The main emphasis is on the crucial significance of Generative Adversarial Networks (GANs), which have transformed the process of image formation, with a specific emphasis on the impact of conditional GANs. The conditional models enable controlled image generation, and their influence on the production of high-quality images is extensively analyzed. We propose a novel method of generating semantically aware embeddings from the input text description which learns better mapping to generate the output image. Moreover, the paper examines the crucial significance of datasets in T2I research and investigates the development of T2I approaches. Ultimately, the research highlights the persistent difficulties in assessing T2I models, with a particular emphasis on image quality measurements. It emphasizes the necessity for complete evaluation methods that take into account both visual realism and semantic coherence. Experimental results demonstrate that our approach yields considerable performance over existing approaches for text to image generation.