SOTAVerified

Image Generation

Image Generation (synthesis) is the task of generating new images from an existing dataset.

  • Unconditional generation refers to generating samples unconditionally from the dataset, i.e. $p(y)$
  • Conditional image generation (subtask) refers to generating samples conditionally from the dataset, based on a label, i.e. $p(y|x)$.

In this section, you can find state-of-the-art leaderboards for unconditional generation. For conditional generation, and other types of image generations, refer to the subtasks.

( Image credit: StyleGAN )

Papers

Showing 551600 of 6689 papers

TitleStatusHype
SEED-Bench-2: Benchmarking Multimodal Large Language ModelsCode2
Text-Driven Image Editing via Learnable RegionsCode2
LLMGA: Multimodal Large Language Model based Generation AssistantCode2
Flow-Guided Diffusion for Video InpaintingCode2
MVControl: Adding Conditional Control to Multi-view Diffusion for Controllable Text-to-3D GenerationCode2
Diffusion360: Seamless 360 Degree Panoramic Image Generation based on Diffusion ModelsCode2
The Chosen One: Consistent Characters in Text-to-Image Diffusion ModelsCode2
Matryoshka Diffusion ModelsCode2
A Pytorch Reproduction of Masked Generative Image TransformerCode2
LAMP: Learn A Motion Pattern for Few-Shot-Based Video GenerationCode2
PonderV2: Pave the Way for 3D Foundation Model with A Universal Pre-training ParadigmCode2
ScaleCrafter: Tuning-free Higher-Resolution Visual Generation with Diffusion ModelsCode2
DrivingDiffusion: Layout-Guided multi-view driving scene video generation with latent diffusion modelCode2
Mini-DALLE3: Interactive Text to Image by Prompting Large Language ModelsCode2
Aligning Text-to-Image Diffusion Models with Reward BackpropagationCode2
MiniGPT-5: Interleaved Vision-and-Language Generation via Generative VokensCode2
Direct Inversion: Boosting Diffusion-based Editing with 3 Lines of CodeCode2
Consistency Trajectory Models: Learning Probability Flow ODE Trajectory of DiffusionCode2
InstructCV: Instruction-Tuned Text-to-Image Diffusion Models as Vision GeneralistsCode2
Denoising Diffusion Bridge ModelsCode2
Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction TuningCode2
Relay Diffusion: Unifying diffusion process across resolutions for image synthesisCode2
Residual Denoising Diffusion ModelsCode2
Dense Text-to-Image Generation with Attention ModulationCode2
Bayesian Flow NetworksCode2
DatasetDM: Synthesizing Data with Perception Annotations Using Diffusion ModelsCode2
Taming the Power of Diffusion Models for High-Quality Virtual Try-On with Appearance FlowCode2
ConceptLab: Creative Concept Generation using VLM-Guided Diffusion Prior ConstraintsCode2
A Systematic Survey of Prompt Engineering on Vision-Language Foundation ModelsCode2
Subject-Diffusion:Open Domain Personalized Text-to-Image Generation without Test-time Fine-tuningCode2
BoxDiff: Text-to-Image Synthesis with Training-Free Box-Constrained DiffusionCode2
Flow Matching in Latent SpaceCode2
Planting a SEED of Vision in Large Language ModelCode2
T2I-CompBench: A Comprehensive Benchmark for Open-world Compositional Text-to-image GenerationCode2
SDXL: Improving Latent Diffusion Models for High-Resolution Image SynthesisCode2
MVDiffusion: Enabling Holistic Multi-view Image Generation with Correspondence-Aware DiffusionCode2
DreamDiffusion: Generating High-Quality Images from Brain EEG SignalsCode2
Human Preference Score v2: A Solid Benchmark for Evaluating Human Preferences of Text-to-Image SynthesisCode2
Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion modelsCode2
VideoComposer: Compositional Video Synthesis with Motion ControllabilityCode2
StyleDrop: Text-to-Image Generation in Any StyleCode2
Intelligent Grimm -- Open-ended Visual Storytelling via Latent Diffusion ModelsCode2
Differential Diffusion: Giving Each Pixel Its StrengthCode2
ViCo: Plug-and-play Visual Condition for Personalized Text-to-image GenerationCode2
STEVE-1: A Generative Model for Text-to-Behavior in MinecraftCode2
Wuerstchen: An Efficient Architecture for Large-Scale Text-to-Image Diffusion ModelsCode2
Tree-Ring Watermarks: Fingerprints for Diffusion Images that are Invisible and RobustCode2
Cones 2: Customizable Image Synthesis with Multiple SubjectsCode2
GPT4Tools: Teaching Large Language Model to Use Tools via Self-instructionCode2
Conditional Diffusion Models for Semantic 3D Brain MRI SynthesisCode2
Show:102550
← PrevPage 12 of 134Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Improved DDPMFID12.3Unverified
2ADMFID11.84Unverified
3BigGAN-deepFID8.1Unverified
4Polarity-BigGANFID6.82Unverified
5VQGAN+Transformer (k=mixed, p=1.0, a=0.005)FID6.59Unverified
6MaskGITFID6.18Unverified
7VQGAN+Transformer (k=600, p=1.0, a=0.05)FID5.2Unverified
8CDMFID4.88Unverified
9ADM-GFID4.59Unverified
10RINFID4.51Unverified
#ModelMetricClaimedVerifiedStatus
1PresGANFID52.2Unverified
2RESFLOWFID48.29Unverified
3Residual FlowFID46.37Unverified
4GLF+perceptual loss (ours)FID44.6Unverified
5ProdPoly no activation functionsFID40.45Unverified
6ProdPoly no activation functionsFID36.77Unverified
7ACGANFID35.47Unverified
8DenseFlow-74-10FID34.9Unverified
9NVAE w/ flowFID32.53Unverified
10QSNGANFID31.97Unverified
#ModelMetricClaimedVerifiedStatus
1GLIDE + CLSFID30.87Unverified
2GLIDE + CLIPFID30.46Unverified
3GLIDE + CLS-FREEFID29.22Unverified
4GLIDE + CLIP + CLS + CLS-FREEFID29.18Unverified
5PGMGANFID21.73Unverified
6CLR-GANFID20.27Unverified
7FMFID14.45Unverified
8CT (Direct Generation, NFE=1)FID13Unverified
9CT (Direct Generation, NFE=2)FID11.1Unverified
10GLIDE +CLSKID7.95Unverified