TextControlGAN: Text-to-Image Synthesis with Controllable Generative Adversarial Networks
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ku, Hyeeun | - |
dc.contributor.author | Lee, Minhyeok | - |
dc.date.accessioned | 2024-01-09T04:33:46Z | - |
dc.date.available | 2024-01-09T04:33:46Z | - |
dc.date.issued | 2023-04 | - |
dc.identifier.issn | 2076-3417 | - |
dc.identifier.issn | 2076-3417 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/69890 | - |
dc.description.abstract | Generative adversarial networks (GANs) have demonstrated remarkable potential in the realm of text-to-image synthesis. Nevertheless, conventional GANs employing conditional latent space interpolation and manifold interpolation (GAN-CLS-INT) encounter challenges in generating images that accurately reflect the given text descriptions. To overcome these limitations, we introduce TextControlGAN, a controllable GAN-based model specifically designed for text-to-image synthesis tasks. In contrast to traditional GANs, TextControlGAN incorporates a neural network structure, known as a regressor, to effectively learn features from conditional texts. To further enhance the learning performance of the regressor, data augmentation techniques are employed. As a result, the generator within TextControlGAN can learn conditional texts more effectively, leading to the production of images that more closely adhere to the textual conditions. Furthermore, by concentrating the discriminator’s training efforts on GAN training exclusively, the overall quality of the generated images is significantly improved. Evaluations conducted on the Caltech-UCSD Birds-200 (CUB) dataset demonstrate that TextControlGAN surpasses the performance of the cGAN-based GAN-INT-CLS model, achieving a 17.6% improvement in Inception Score (IS) and a 36.6% reduction in Fréchet Inception Distance (FID). In supplementary experiments utilizing 128 × 128 resolution images, TextControlGAN exhibits a remarkable ability to manipulate minor features of the generated bird images according to the given text descriptions. These findings highlight the potential of TextControlGAN as a powerful tool for generating high-quality, text-conditioned images, paving the way for future advancements in the field of text-to-image synthesis. | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | MDPI | - |
dc.title | TextControlGAN: Text-to-Image Synthesis with Controllable Generative Adversarial Networks | - |
dc.type | Article | - |
dc.identifier.doi | 10.3390/app13085098 | - |
dc.identifier.bibliographicCitation | Applied Sciences (Switzerland), v.13, no.8 | - |
dc.description.isOpenAccess | Y | - |
dc.identifier.wosid | 000977524100001 | - |
dc.identifier.scopusid | 2-s2.0-85156122917 | - |
dc.citation.number | 8 | - |
dc.citation.title | Applied Sciences (Switzerland) | - |
dc.citation.volume | 13 | - |
dc.type.docType | Article | - |
dc.publisher.location | 스위스 | - |
dc.subject.keywordAuthor | computer vision | - |
dc.subject.keywordAuthor | generative adversarial networks | - |
dc.subject.keywordAuthor | image generation | - |
dc.subject.keywordAuthor | text-to-image synthesis | - |
dc.relation.journalResearchArea | Chemistry | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Materials Science | - |
dc.relation.journalResearchArea | Physics | - |
dc.relation.journalWebOfScienceCategory | Chemistry, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Engineering, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Materials Science, Multidisciplinary | - |
dc.relation.journalWebOfScienceCategory | Physics, Applied | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
84, Heukseok-ro, Dongjak-gu, Seoul, Republic of Korea (06974)02-820-6194
COPYRIGHT 2019 Chung-Ang University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.