USIS: Unsupervised Semantic Image Synthesis
George Eskandar, Mohamed Abdelsamad, Karim Armanious, Bin Yang
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/GeorgeEskandar/USIS-Unsupervised-Semantic-Image-SynthesisOfficialpytorch★ 8
Abstract
Semantic Image Synthesis (SIS) is a subclass of image-to-image translation where a photorealistic image is synthesized from a segmentation mask. SIS has mostly been addressed as a supervised problem. However, state-of-the-art methods depend on a huge amount of labeled data and cannot be applied in an unpaired setting. On the other hand, generic unpaired image-to-image translation frameworks underperform in comparison, because they color-code semantic layouts and feed them to traditional convolutional networks, which then learn correspondences in appearance instead of semantic content. In this initial work, we propose a new Unsupervised paradigm for Semantic Image Synthesis (USIS) as a first step towards closing the performance gap between paired and unpaired settings. Notably, the framework deploys a SPADE generator that learns to output images with visually separable semantic classes using a self-supervised segmentation loss. Furthermore, in order to match the color and texture distribution of real images without losing high-frequency information, we propose to use whole image wavelet-based discrimination. We test our methodology on 3 challenging datasets and demonstrate its ability to generate multimodal photorealistic images with an improved quality in the unpaired setting.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| ADE20K Labels-to-Photos | USIS | mIoU | 17.38 | — | Unverified |
| Cityscapes Labels-to-Photo | USIS | mIoU | 44.78 | — | Unverified |
| COCO-Stuff Labels-to-Photos | USIS | FID | 27.8 | — | Unverified |