SOTAVerified

Soft Truncation: A Universal Training Technique of Score-based Diffusion Model for High Precision Score Estimation

2021-06-10Code Available1· sign in to hype

Dongjun Kim, Seungjae Shin, Kyungwoo Song, Wanmo Kang, Il-Chul Moon

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Recent advances in diffusion models bring state-of-the-art performance on image generation tasks. However, empirical results from previous research in diffusion models imply an inverse correlation between density estimation and sample generation performances. This paper investigates with sufficient empirical evidence that such inverse correlation happens because density estimation is significantly contributed by small diffusion time, whereas sample generation mainly depends on large diffusion time. However, training a score network well across the entire diffusion time is demanding because the loss scale is significantly imbalanced at each diffusion time. For successful training, therefore, we introduce Soft Truncation, a universally applicable training technique for diffusion models, that softens the fixed and static truncation hyperparameter into a random variable. In experiments, Soft Truncation achieves state-of-the-art performance on CIFAR-10, CelebA, CelebA-HQ 256x256, and STL-10 datasets.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CelebA 64x64DDPM++ (VP, NLL) + STFID2.9Unverified
CelebA 64x64UNCSN++ (RVE) + STbits/dimension1.97Unverified
CelebA 64x64DDPM++ (VP, FID) + STFID1.9Unverified
CelebA-HQ 256x256UNCSN++ (RVE) + STFID7.16Unverified
FFHQ 256 x 256UDM (RVE) + STFID5.54Unverified
ImageNet 32x32DDPM++ (VP, NLL) + STFID8.42Unverified
LSUN Bedroom 256 x 256UDM (RVE) + STFID4.57Unverified
STL-10UNCSN++ (RVE) + STFID7.71Unverified

Reproductions