Denoising Diffusion Probabilistic Models
Jonathan Ho, Ajay Jain, Pieter Abbeel
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/hojonathanho/diffusionOfficialIn papertf★ 5,105
- github.com/labmlai/annotated_deep_learning_paper_implementationspytorch★ 66,103
- github.com/yang-song/score_sde_pytorchpytorch★ 2,088
- github.com/yang-song/score_sdepytorch★ 1,810
- github.com/Janspiry/Palette-Image-to-Image-Diffusion-Modelspytorch★ 1,807
- github.com/glouppe/info8010-deep-learningpytorch★ 1,279
- github.com/arpitbansal297/cold-diffusion-modelspytorch★ 1,120
- github.com/lmnt-com/diffwavepytorch★ 889
- github.com/teapearce/conditional_diffusion_mnistpytorch★ 825
- github.com/teticio/audio-diffusionpytorch★ 789
Abstract
We present high quality image synthesis results using diffusion probabilistic models, a class of latent variable models inspired by considerations from nonequilibrium thermodynamics. Our best results are obtained by training on a weighted variational bound designed according to a novel connection between diffusion probabilistic models and denoising score matching with Langevin dynamics, and our models naturally admit a progressive lossy decompression scheme that can be interpreted as a generalization of autoregressive decoding. On the unconditional CIFAR10 dataset, we obtain an Inception score of 9.46 and a state-of-the-art FID score of 3.17. On 256x256 LSUN, we obtain sample quality similar to ProgressiveGAN. Our implementation is available at https://github.com/hojonathanho/diffusion
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| CIFAR-10 | DDPM | NLL (bits/dim) | 3.69 | — | Unverified |