SOTAVerified

Image to Video Generation

Image to Video Generation refers to the task of generating a sequence of video frames based on a single still image or a set of still images. The goal is to produce a video that is coherent and consistent in terms of appearance, motion, and style, while also being temporally consistent, meaning that the generated video should look like a coherent sequence of frames that are temporally ordered. This task is typically tackled using deep generative models, such as Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs), that are trained on large datasets of videos. The models learn to generate plausible video frames that are conditioned on the input image, as well as on any other auxiliary information, such as a sound or text track.

Papers

Showing 5185 of 85 papers

TitleStatusHype
TRIP: Temporal Residual Learning with Image Noise Prior for Image-to-Video Diffusion Models0
MarDini: Masked Autoregressive Diffusion for Video Generation at Scale0
MG-Gen: Single Image to Motion Graphics Generation with Layer Decomposition0
Tuning-Free Noise Rectification for High Fidelity Image-to-Video Generation0
Decouple Content and Motion for Conditional Image-to-Video Generation0
MotionAgent: Fine-grained Controllable Video Generation via Motion Field Agent0
MotionCanvas: Cinematic Shot Design with Controllable Image-to-Video Generation0
Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling0
MotionPro: A Precise Motion Controller for Image-to-Video Generation0
MotionStone: Decoupled Motion Intensity Modulation with Diffusion Transformer for Image-to-Video Generation0
MoVideo: Motion-Aware Video Generation with Diffusion Models0
AniClipart: Clipart Animation with Text-to-Video Priors0
VidCRAFT3: Camera, Object, and Lighting Control for Image-to-Video Generation0
OmniDrag: Enabling Motion Control for Omnidirectional Image-to-Video Generation0
Dance Any Beat: Blending Beats with Visuals in Dance Video Generation0
OSV: One Step is Enough for High-Quality Image to Video Generation0
Cavia: Camera-controllable Multi-view Video Diffusion with View-Integrated Attention0
RealCam-I2V: Real-World Image-to-Video Generation with Interactive Complex Camera Control0
CamViG: Camera Aware Image-to-Video Generation with Multimodal Transformers0
CamCo: Camera-Controllable 3D-Consistent Image-to-Video Generation0
AtomoVideo: High Fidelity Image-to-Video Generation0
Self-Training for Domain Adaptive Scene Text Detection0
SG-I2V: Self-Guided Trajectory Control in Image-to-Video Generation0
ATI: Any Trajectory Instruction for Controllable Video Generation0
SubstationAI: Multimodal Large Model-Based Approaches for Analyzing Substation Equipment Faults0
Through-The-Mask: Mask-based Motion Trajectories for Image-to-Video Generation0
A Survey of Emerging Approaches and Advances in Video Generation0
TIP-I2V: A Million-Scale Real Text and Image Prompt Dataset for Image-to-Video Generation0
TIV-Diffusion: Towards Object-Centric Movement for Text-driven Image to Video Generation0
Learning to Forecast and Refine Residual Motion for Image-to-Video GenerationCode0
Factorized-Dreamer: Training A High-Quality Video Generator with Limited and Low-Quality DataCode0
GenRec: Unifying Video Generation and Recognition with Diffusion ModelsCode0
Video Generation from Single Semantic Label MapCode0
Magic 1-For-1: Generating One Minute Video Clips within One MinuteCode0
Stable Video Diffusion: Scaling Latent Video Diffusion Models to Large DatasetsCode0
Show:102550
← PrevPage 2 of 2Next →

No leaderboard results yet.