SOTAVerified

Image Generation

Image Generation (synthesis) is the task of generating new images from an existing dataset.

  • Unconditional generation refers to generating samples unconditionally from the dataset, i.e. $p(y)$
  • Conditional image generation (subtask) refers to generating samples conditionally from the dataset, based on a label, i.e. $p(y|x)$.

In this section, you can find state-of-the-art leaderboards for unconditional generation. For conditional generation, and other types of image generations, refer to the subtasks.

( Image credit: StyleGAN )

Papers

Showing 43514400 of 6689 papers

TitleStatusHype
CLIP2GAN: Towards Bridging Text with the Latent Space of GANs0
Hand-Object Interaction Image Generation0
The Myth of Culturally Agnostic AI Models0
Conditional Progressive Generative Adversarial Network for satellite image generation0
Diffusion Probabilistic Model Made Slim0
Traditional Classification Neural Networks are Good Generators: They are Competitive with DDPMs and GANs0
Cross-domain Microscopy Cell Counting by Disentangled Transfer Learning0
Efficient Video Prediction via Sparsely Conditioned Flow Matching0
ILSGAN: Independent Layer Synthesis for Unsupervised Foreground-Background SegmentationCode0
SpaText: Spatio-Textual Representation for Controllable Image Generation0
Unifying conditional and unconditional semantic image synthesis with OCO-GAN0
CHIMLE: Conditional Hierarchical IMLE for Multimodal Conditional Image SynthesisCode1
Learning Detailed Radiance Manifolds for High-Fidelity and 3D-Consistent Portrait Synthesis from Monocular Image0
Efficient Feature Extraction for High-resolution Video Frame InterpolationCode1
More comprehensive facial inversion for more effective expression recognitionCode0
Shifted Diffusion for Text-to-image GenerationCode1
Diffusion-SDF: Conditional Generative Modeling of Signed Distance Functions0
ReCo: Region-Controlled Text-to-Image Generation0
CGOF++: Controllable 3D Face Synthesis with Conditional Generative Occupancy Fields0
Latent Video Diffusion Models for High-Fidelity Long Video GenerationCode2
TetraDiffusion: Tetrahedral Diffusion Models for 3D Shape GenerationCode1
Inversion-Based Style Transfer with Diffusion ModelsCode2
Paint by Example: Exemplar-based Image Editing with Diffusion ModelsCode3
Rethinking Implicit Neural Representations for Vision Learners0
Retrieval-Augmented Multimodal Language Modeling0
Plug-and-Play Diffusion Features for Text-Driven Image-to-Image TranslationCode2
The Euclidean Space is Evil: Hyperbolic Attribute Editing for Few-shot Image GenerationCode1
Person Image Synthesis via Denoising Diffusion ModelCode2
Human Evaluation of Text-to-Image Models on a Multi-Task Benchmark0
SinDiffusion: Learning a Diffusion Model from a Single Natural ImageCode2
DreamArtist++: Controllable One-Shot Text-to-Image Generation via Positive-Negative Adapter0
Diffusion-Based Scene Graph to Image Generation with Masked Contrastive Pre-TrainingCode1
SceneComposer: Any-Level Semantic Image Synthesis0
TimbreCLIP: Connecting Timbre to Text and Images0
Exploring the Effectiveness of Mask-Guided Feature Modulation as a Mechanism for Localized Style Editing of Real Images0
VectorFusion: Text-to-SVG by Abstracting Pixel-Based Diffusion Models0
Exploring Discrete Diffusion Models for Image CaptioningCode1
IC3D: Image-Conditioned 3D Diffusion for Shape Generation0
Single Stage Multi-Pose Virtual Try-On0
Potential Auto-driving Threat: Universal Rain-removal Attack0
GLAMI-1M: A Multilingual Image-Text Fashion DatasetCode1
UMFuse: Unified Multi View Fusion for Human Editing applications0
RenderDiffusion: Image Diffusion for 3D Reconstruction, Inpainting and GenerationCode2
Null-text Inversion for Editing Real Images using Guided Diffusion ModelsCode4
Conffusion: Confidence Intervals for Diffusion ModelsCode1
MAGE: MAsked Generative Encoder to Unify Representation Learning and Image SynthesisCode2
A Creative Industry Image Generation Dataset Based on Captions0
Will Large-scale Generative Models Corrupt Future Datasets?Code0
Versatile Diffusion: Text, Images and Variations All in One Diffusion ModelCode6
Extreme Generative Image Compression by Learning Text Embedding from Diffusion Models0
Show:102550
← PrevPage 88 of 134Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Improved DDPMFID12.3Unverified
2ADMFID11.84Unverified
3BigGAN-deepFID8.1Unverified
4Polarity-BigGANFID6.82Unverified
5VQGAN+Transformer (k=mixed, p=1.0, a=0.005)FID6.59Unverified
6MaskGITFID6.18Unverified
7VQGAN+Transformer (k=600, p=1.0, a=0.05)FID5.2Unverified
8CDMFID4.88Unverified
9ADM-GFID4.59Unverified
10RINFID4.51Unverified
#ModelMetricClaimedVerifiedStatus
1PresGANFID52.2Unverified
2RESFLOWFID48.29Unverified
3Residual FlowFID46.37Unverified
4GLF+perceptual loss (ours)FID44.6Unverified
5ProdPoly no activation functionsFID40.45Unverified
6ProdPoly no activation functionsFID36.77Unverified
7ACGANFID35.47Unverified
8DenseFlow-74-10FID34.9Unverified
9NVAE w/ flowFID32.53Unverified
10QSNGANFID31.97Unverified
#ModelMetricClaimedVerifiedStatus
1GLIDE + CLSFID30.87Unverified
2GLIDE + CLIPFID30.46Unverified
3GLIDE + CLS-FREEFID29.22Unverified
4GLIDE + CLIP + CLS + CLS-FREEFID29.18Unverified
5PGMGANFID21.73Unverified
6CLR-GANFID20.27Unverified
7FMFID14.45Unverified
8CT (Direct Generation, NFE=1)FID13Unverified
9CT (Direct Generation, NFE=2)FID11.1Unverified
10GLIDE +CLSKID7.95Unverified