Wavelet-based Unsupervised Label-to-Image Translation
George Eskandar, Mohamed Abdelsamad, Karim Armanious, Shuai Zhang, Bin Yang
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/GeorgeEskandar/USIS-Unsupervised-Semantic-Image-SynthesisOfficialIn paperpytorch★ 8
Abstract
Semantic Image Synthesis (SIS) is a subclass of image-to-image translation where a semantic layout is used to generate a photorealistic image. State-of-the-art conditional Generative Adversarial Networks (GANs) need a huge amount of paired data to accomplish this task while generic unpaired image-to-image translation frameworks underperform in comparison, because they color-code semantic layouts and learn correspondences in appearance instead of semantic content. Starting from the assumption that a high quality generated image should be segmented back to its semantic layout, we propose a new Unsupervised paradigm for SIS (USIS) that makes use of a self-supervised segmentation loss and whole image wavelet based discrimination. Furthermore, in order to match the high-frequency distribution of real images, a novel generator architecture in the wavelet domain is proposed. We test our methodology on 3 challenging datasets and demonstrate its ability to bridge the performance gap between paired and unpaired models.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| ADE20K Labels-to-Photos | USIS-Wavelet | mIoU | 16.95 | — | Unverified |
| Cityscapes Labels-to-Photo | USIS-Wavelet | mIoU | 42.32 | — | Unverified |
| COCO-Stuff Labels-to-Photos | USIS-Wavelet | FID | 28.6 | — | Unverified |