SOTAVerified

Learning Cross-Video Neural Representations for High-Quality Frame Interpolation

2022-02-28Code Available1· sign in to hype

Wentao Shangguan, Yu Sun, Weijie Gan, Ulugbek S. Kamilov

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper considers the problem of temporal video interpolation, where the goal is to synthesize a new video frame given its two neighbors. We propose Cross-Video Neural Representation (CURE) as the first video interpolation method based on neural fields (NF). NF refers to the recent class of methods for the neural representation of complex 3D scenes that has seen widespread success and application across computer vision. CURE represents the video as a continuous function parameterized by a coordinate-based neural network, whose inputs are the spatiotemporal coordinates and outputs are the corresponding RGB values. CURE introduces a new architecture that conditions the neural network on the input frames for imposing space-time consistency in the synthesized video. This not only improves the final interpolation quality, but also enables CURE to learn a prior across multiple videos. Experimental evaluations show that CURE achieves the state-of-the-art performance on video interpolation on several benchmark datasets.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
MSU Video Frame InterpolationCUREPSNR28.01Unverified
Nvidia Dynamic SceneCUREPSNR36.24Unverified
SNU-FILM (easy)CUREPSNR39.9Unverified
SNU-FILM (extreme)CUREPSNR25.44Unverified
SNU-FILM (hard)CUREPSNR30.66Unverified
SNU-FILM (medium)CUREPSNR35.94Unverified
UCF101CUREPSNR35.36Unverified
Vimeo90KCUREPSNR35.73Unverified
X4K1000FPS-2KCUREPSNR30.05Unverified
Xiph-4kCUREPSNR30.94Unverified

Reproductions