SOTAVerified

StyleGAN-XL: Scaling StyleGAN to Large Diverse Datasets

2022-02-01Code Available2· sign in to hype

Axel Sauer, Katja Schwarz, Andreas Geiger

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Computer graphics has experienced a recent surge of data-centric approaches for photorealistic and controllable content creation. StyleGAN in particular sets new standards for generative modeling regarding image quality and controllability. However, StyleGAN's performance severely degrades on large unstructured datasets such as ImageNet. StyleGAN was designed for controllability; hence, prior works suspect its restrictive design to be unsuitable for diverse datasets. In contrast, we find the main limiting factor to be the current training strategy. Following the recently introduced Projected GAN paradigm, we leverage powerful neural network priors and a progressive growing strategy to successfully train the latest StyleGAN3 generator on ImageNet. Our final model, StyleGAN-XL, sets a new state-of-the-art on large-scale image synthesis and is the first to generate images at a resolution of 1024^2 at such a dataset scale. We demonstrate that this model can invert and edit images beyond the narrow domain of portraits or specific object classes.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
FFHQ 1024 x 1024StyleGAN-XLFID2.02Unverified
FFHQ 256 x 256StyleGAN-XLFID2.19Unverified
FFHQ 256 x 256StyleGAN-XL (DINOv2)FD240.07Unverified
FFHQ 512 x 512StyleGAN-XLFID2.41Unverified
ImageNet 128x128StyleGAN-XLFID1.81Unverified
ImageNet 256x256StyleGAN-XLFID2.3Unverified
ImageNet 32x32StyleGAN-XLFID1.1Unverified
ImageNet 512x512StyleGAN-XLFID2.4Unverified
ImageNet 64x64StyleGAN-XLFID1.51Unverified
Pokemon 1024x1024StyleGAN-XLFID25.47Unverified
Pokemon 256x256StyleGAN-XLFID23.97Unverified

Reproductions