SOTAVerified

Long and Short Guidance in Score identity Distillation for One-Step Text-to-Image Generation

2024-06-03Code Available2· sign in to hype

Mingyuan Zhou, Zhendong Wang, Huangjie Zheng, Hai Huang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Diffusion-based text-to-image generation models trained on extensive text-image pairs have shown the capacity to generate photorealistic images consistent with textual descriptions. However, a significant limitation of these models is their slow sample generation, which requires iterative refinement through the same network. In this paper, we enhance Score identity Distillation (SiD) by developing long and short classifier-free guidance (LSG) to efficiently distill pretrained Stable Diffusion models without using real training data. SiD aims to optimize a model-based explicit score matching loss, utilizing a score-identity-based approximation alongside the proposed LSG for practical computation. By training exclusively with fake images synthesized with its one-step generator, SiD equipped with LSG rapidly improves FID and CLIP scores, achieving state-of-the-art FID performance while maintaining a competitive CLIP score. Specifically, its data-free distillation of Stable Diffusion 1.5 achieves a record low FID of 8.15 on the COCO-2014 validation set, with a CLIP score of 0.304 at an LSG scale of 1.5, and an FID of 9.56 with a CLIP score of 0.313 at an LSG scale of 2. Our code and distilled one-step text-to-image generators are available at https://github.com/mingyuanzhou/SiD-LSG.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
COCO (Common Objects in Context)SiD-LSG (Data-free distillation, zero-shot FID)FID8.15Unverified

Reproductions