SOTAVerified

FasterDiT: Towards Faster Diffusion Transformers Training without Architecture Modification

2024-10-14Code Available5· sign in to hype

Jingfeng Yao, Wang Cheng, Wenyu Liu, Xinggang Wang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Diffusion Transformers (DiT) have attracted significant attention in research. However, they suffer from a slow convergence rate. In this paper, we aim to accelerate DiT training without any architectural modification. We identify the following issues in the training process: firstly, certain training strategies do not consistently perform well across different data. Secondly, the effectiveness of supervision at specific timesteps is limited. In response, we propose the following contributions: (1) We introduce a new perspective for interpreting the failure of the strategies. Specifically, we slightly extend the definition of Signal-to-Noise Ratio (SNR) and suggest observing the Probability Density Function (PDF) of SNR to understand the essence of the data robustness of the strategy. (2) We conduct numerous experiments and report over one hundred experimental results to empirically summarize a unified accelerating strategy from the perspective of PDF. (3) We develop a new supervision method that further accelerates the training process of DiT. Based on them, we propose FasterDiT, an exceedingly simple and practicable design strategy. With few lines of code modifications, it achieves 2.30 FID on ImageNet 256 resolution at 1000k iterations, which is comparable to DiT (2.27 FID) but 7 times faster in training.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ImageNet 256x256FasterDiT-XL/2FID2.03Unverified

Reproductions