Paper ID | ARS-5.8 | ||
Paper Title | Text to Image Synthesis with Erudite Generative Adversarial Networks | ||
Authors | Zhiqiang Zhang, Hosei University, Japan; Wenxin Yu, Ning Jiang, Southwest University of Science and Technology, China; Jinjia Zhou, Hosei University, Japan | ||
Session | ARS-5: Image and Video Synthesis, Rendering and Visualization | ||
Location | Area I | ||
Session Time: | Tuesday, 21 September, 08:00 - 09:30 | ||
Presentation Time: | Tuesday, 21 September, 08:00 - 09:30 | ||
Presentation | Poster | ||
Topic | Image and Video Analysis, Synthesis, and Retrieval: Image & Video Synthesis, Rendering, and Visualization | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | In this paper, an Erudite Generative Adversarial Networks (EruditeGAN) is proposed for the text-to-image synthesis task. By introducing additional image distribution related to the original image into the network structure, the entire network can learn more about the image distribution and become more knowledgeable. In this case, it can be more clear about the distribution of the image that needs to be synthesized and finally synthesize high-quality results. Experiments well validate the effectiveness of our method and demonstrate the different effects of different distribution situations on the final results. According to the quantitative results of Fréchet Inception Distance (FID) and R-precision, our method's comprehensive score is the best, which reflects our results are closer to the real image effect. |