SOTAVerified

Image to Video Generation

Image to Video Generation refers to the task of generating a sequence of video frames based on a single still image or a set of still images. The goal is to produce a video that is coherent and consistent in terms of appearance, motion, and style, while also being temporally consistent, meaning that the generated video should look like a coherent sequence of frames that are temporally ordered. This task is typically tackled using deep generative models, such as Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs), that are trained on large datasets of videos. The models learn to generate plausible video frames that are conditioned on the input image, as well as on any other auxiliary information, such as a sound or text track.

Papers

Showing 5185 of 85 papers

TitleStatusHype
Identifying and Solving Conditional Image Leakage in Image-to-Video Diffusion Model0
TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video GenerationCode1
CamCo: Camera-Controllable 3D-Consistent Image-to-Video Generation0
CamViG: Camera Aware Image-to-Video Generation with Multimodal Transformers0
Dance Any Beat: Blending Beats with Visuals in Dance Video Generation0
TI2V-Zero: Zero-Shot Image Conditioning for Text-to-Video Diffusion ModelsCode2
AniClipart: Clipart Animation with Text-to-Video Priors0
TRIP: Temporal Residual Learning with Image Noise Prior for Image-to-Video Diffusion Models0
Champ: Controllable and Consistent Human Image Animation with 3D Parametric GuidanceCode7
AnyV2V: A Tuning-Free Framework For Any Video-to-Video Editing TasksCode4
Mora: Enabling Generalist Video Generation via A Multi-Agent FrameworkCode5
Follow-Your-Click: Open-domain Regional Image Animation via Short PromptsCode4
Tuning-Free Noise Rectification for High Fidelity Image-to-Video Generation0
AtomoVideo: High Fidelity Image-to-Video Generation0
ConsistI2V: Enhancing Visual Consistency for Image-to-Video GenerationCode3
Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling0
Kandinsky 3.0 Technical ReportCode2
DreamVideo: High-Fidelity Image-to-Video Generation with Image Retention and Text Guidance0
MagDiff: Multi-Alignment Diffusion for High-Fidelity Video Generation and EditingCode1
Stable Video Diffusion: Scaling Latent Video Diffusion Models to Large DatasetsCode0
Decouple Content and Motion for Conditional Image-to-Video Generation0
AnimateAnything: Fine-Grained Open Domain Image Animation with Motion GuidanceCode2
MoVideo: Motion-Aware Video Generation with Diffusion Models0
Conditional Image-to-Video Generation with Latent Flow Diffusion ModelsCode2
A Method for Animating Children's Drawings of the Human FigureCode6
Dreamix: Video Diffusion Models are General Video Editors0
SceneRF: Self-Supervised Monocular 3D Scene Reconstruction with Radiance FieldsCode2
Collaborative Neural Rendering using Anime Character SheetsCode2
Make It Move: Controllable Image-to-Video Generation with Text DescriptionsCode1
Image-to-Video Generation via 3D Facial Dynamics0
TiVGAN: Text to Image to Video Generation with Step-by-Step Evolutionary Generator0
Self-Training for Domain Adaptive Scene Text Detection0
Lifespan Age Transformation SynthesisCode1
Video Generation from Single Semantic Label MapCode0
Learning to Forecast and Refine Residual Motion for Image-to-Video GenerationCode0
Show:102550
← PrevPage 2 of 2Next →

No leaderboard results yet.