SOTAVerified

Dual Contradistinctive Generative Autoencoder

2020-11-19CVPR 2021Unverified0· sign in to hype

Gaurav Parmar, Dacheng Li, Kwonjoon Lee, Zhuowen Tu

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We present a new generative autoencoder model with dual contradistinctive losses to improve generative autoencoder that performs simultaneous inference (reconstruction) and synthesis (sampling). Our model, named dual contradistinctive generative autoencoder (DC-VAE), integrates an instance-level discriminative loss (maintaining the instance-level fidelity for the reconstruction/synthesis) with a set-level adversarial loss (encouraging the set-level fidelity for there construction/synthesis), both being contradistinctive. Extensive experimental results by DC-VAE across different resolutions including 32x32, 64x64, 128x128, and 512x512 are reported. The two contradistinctive losses in VAE work harmoniously in DC-VAE leading to a significant qualitative and quantitative performance enhancement over the baseline VAEs without architectural changes. State-of-the-art or competitive results among generative autoencoders for image reconstruction, image synthesis, image interpolation, and representation learning are observed. DC-VAE is a general-purpose VAE model, applicable to a wide variety of downstream tasks in computer vision and machine learning.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CelebA 128x128DC-VAEFID19.9Unverified
CelebA-HQ 256x256DC-VAEFID15.81Unverified
CIFAR-10DC-VAEFID17.9Unverified
LSUN Bedroom 128 x 128DC-VAEFID14.3Unverified
STL-10DC-VAEFID41.9Unverified

Reproductions