SOTAVerified

Semantics-Enhanced Adversarial Nets for Text-to-Image Synthesis

2019-10-01ICCV 2019Unverified0· sign in to hype

Hongchen Tan, Xiuping Liu, Xin Li, Yi Zhang, Baocai Yin

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper presents a new model, Semantics-enhanced Generative Adversarial Network (SEGAN), for fine-grained text-to-image generation. We introduce two modules, a Semantic Consistency Module (SCM) and an Attention Competition Module (ACM), to our SEGAN. The SCM incorporates image-level semantic consistency into the training of the Generative Adversarial Network (GAN), and can diversify the generated images and improve their structural coherence. A Siamese network and two types of semantic similarities are designed to map the synthesized image and the groundtruth image to nearby points in the latent semantic feature space. The ACM constructs adaptive attention weights to differentiate keywords from unimportant words, and improves the stability and accuracy of SEGAN. Extensive experiments demonstrate that our SEGAN significantly outperforms existing state-of-the-art methods in generating photo-realistic images. All source codes and models will be released for comparative study.

Tasks

Reproductions