SOTAVerified

DIVE: Inverting Conditional Diffusion Models for Discriminative Tasks

2025-04-24Code Available0· sign in to hype

Yinqi Li, Hong Chang, Ruibing Hou, Shiguang Shan, Xilin Chen

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Diffusion models have shown remarkable progress in various generative tasks such as image and video generation. This paper studies the problem of leveraging pretrained diffusion models for performing discriminative tasks. Specifically, we extend the discriminative capability of pretrained frozen generative diffusion models from the classification task to the more complex object detection task, by "inverting" a pretrained layout-to-image diffusion model. To this end, a gradient-based discrete optimization approach for replacing the heavy prediction enumeration process, and a prior distribution model for making more accurate use of the Bayes' rule, are proposed respectively. Empirical results show that this method is on par with basic discriminative object detection baselines on COCO dataset. In addition, our method can greatly speed up the previous diffusion-based method for classification without sacrificing accuracy. Code and models are available at https://github.com/LiYinqi/DIVE .

Tasks

Reproductions