SOTAVerified

Image to Video Generation

Image to Video Generation refers to the task of generating a sequence of video frames based on a single still image or a set of still images. The goal is to produce a video that is coherent and consistent in terms of appearance, motion, and style, while also being temporally consistent, meaning that the generated video should look like a coherent sequence of frames that are temporally ordered. This task is typically tackled using deep generative models, such as Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs), that are trained on large datasets of videos. The models learn to generate plausible video frames that are conditioned on the input image, as well as on any other auxiliary information, such as a sound or text track.

Papers

Showing 2130 of 85 papers

TitleStatusHype
MotionAgent: Fine-grained Controllable Video Generation via Motion Field Agent0
FramePainter: Endowing Interactive Image Editing with Video Diffusion PriorsCode3
Through-The-Mask: Mask-based Motion Trajectories for Image-to-Video Generation0
I2VGuard: Safeguarding Images against Misuse in Diffusion-based Image-to-Video Models0
LTX-Video: Realtime Video Latent DiffusionCode9
Open-Sora: Democratizing Efficient Video Production for AllCode13
Generative Video Propagation0
SubstationAI: Multimodal Large Model-Based Approaches for Analyzing Substation Equipment Faults0
TIV-Diffusion: Towards Object-Centric Movement for Text-driven Image to Video Generation0
AniSora: Exploring the Frontiers of Animation Video Generation in the Sora EraCode7
Show:102550
← PrevPage 3 of 9Next →

No leaderboard results yet.