SOTAVerified

Image to Video Generation

Image to Video Generation refers to the task of generating a sequence of video frames based on a single still image or a set of still images. The goal is to produce a video that is coherent and consistent in terms of appearance, motion, and style, while also being temporally consistent, meaning that the generated video should look like a coherent sequence of frames that are temporally ordered. This task is typically tackled using deep generative models, such as Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs), that are trained on large datasets of videos. The models learn to generate plausible video frames that are conditioned on the input image, as well as on any other auxiliary information, such as a sound or text track.

Papers

Showing 2650 of 85 papers

TitleStatusHype
Extrapolating and Decoupling Image-to-Video Generation Models: Motion Modeling is Easier Than You ThinkCode1
Lifespan Age Transformation SynthesisCode1
Make It Move: Controllable Image-to-Video Generation with Text DescriptionsCode1
MMTrail: A Multimodal Trailer Video Dataset with Language and Music DescriptionsCode1
MVOC: a training-free multiple video object composition method with diffusion modelsCode1
Object-Centric Image to Video Generation with Language GuidanceCode1
TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video GenerationCode1
MagDiff: Multi-Alignment Diffusion for High-Fidelity Video Generation and EditingCode1
GeoMan: Temporally Consistent Human Geometry Estimation using Image-to-Video Diffusion0
I2V3D: Controllable image-to-video generation with 3D guidance0
I2VGuard: Safeguarding Images against Misuse in Diffusion-based Image-to-Video Models0
Identifying and Solving Conditional Image Leakage in Image-to-Video Diffusion Model0
Dynamic-I2V: Exploring Image-to-Video Generaion Models via Multimodal LLM0
Image-to-Video Generation via 3D Facial Dynamics0
DreamVideo: High-Fidelity Image-to-Video Generation with Image Retention and Text Guidance0
TiVGAN: Text to Image to Video Generation with Step-by-Step Evolutionary Generator0
Fleximo: Towards Flexible Text-to-Human Motion Video Generation0
FrameBridge: Improving Image-to-Video Generation with Bridge Models0
EvAnimate: Event-conditioned Image-to-Video Generation for Human Animation0
Generative Video Propagation0
Hunyuan-Game: Industrial-grade Intelligent Game Creation Model0
Towards Physically Plausible Video Generation via VLM Planning0
LMP: Leveraging Motion Prior in Zero-Shot Video Generation with Diffusion Transformer0
Dreamix: Video Diffusion Models are General Video Editors0
MagicMotion: Controllable Video Generation with Dense-to-Sparse Trajectory Guidance0
Show:102550
← PrevPage 2 of 4Next →

No leaderboard results yet.