SOTAVerified

On gradient regularizers for MMD GANs

2018-05-29NeurIPS 2018Code Available0· sign in to hype

Michael Arbel, Danica J. Sutherland, Mikołaj Bińkowski, Arthur Gretton

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We propose a principled method for gradient-based regularization of the critic of GAN-like models trained by adversarially optimizing the kernel of a Maximum Mean Discrepancy (MMD). We show that controlling the gradient of the critic is vital to having a sensible loss function, and devise a method to enforce exact, analytical gradient constraints at no additional cost compared to existing approximate techniques based on additive regularizers. The new loss function is provably continuous, and experiments show that it stabilizes and accelerates training, giving image generation models that outperform state-of-the art methods on 160 160 CelebA and 64 64 unconditional ImageNet.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-10SN-SMMDGANFID25Unverified

Reproductions