SOTAVerified

Image to Video Generation

Image to Video Generation refers to the task of generating a sequence of video frames based on a single still image or a set of still images. The goal is to produce a video that is coherent and consistent in terms of appearance, motion, and style, while also being temporally consistent, meaning that the generated video should look like a coherent sequence of frames that are temporally ordered. This task is typically tackled using deep generative models, such as Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs), that are trained on large datasets of videos. The models learn to generate plausible video frames that are conditioned on the input image, as well as on any other auxiliary information, such as a sound or text track.

Papers

Showing 1120 of 85 papers

TitleStatusHype
Follow-Your-Click: Open-domain Regional Image Animation via Short PromptsCode4
Step-Video-TI2V Technical Report: A State-of-the-Art Text-Driven Image-to-Video Generation ModelCode3
FramePainter: Endowing Interactive Image Editing with Video Diffusion PriorsCode3
Robust Watermarking Using Generative Priors Against Image Editing: From Benchmarking to AdvancesCode3
PhysGen: Rigid-Body Physics-Grounded Image-to-Video GenerationCode3
ConsistI2V: Enhancing Visual Consistency for Image-to-Video GenerationCode3
Every Painting Awakened: A Training-free Framework for Painting-to-Animation GenerationCode2
TI2V-Zero: Zero-Shot Image Conditioning for Text-to-Video Diffusion ModelsCode2
Kandinsky 3.0 Technical ReportCode2
AnimateAnything: Fine-Grained Open Domain Image Animation with Motion GuidanceCode2
Show:102550
← PrevPage 2 of 9Next →

No leaderboard results yet.