SOTAVerified

Generating Videos with Scene Dynamics

2016-09-08NeurIPS 2016Unverified0· sign in to hype

Carl Vondrick, Hamed Pirsiavash, Antonio Torralba

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We capitalize on large amounts of unlabeled video in order to learn a model of scene dynamics for both video recognition tasks (e.g. action classification) and video generation tasks (e.g. future prediction). We propose a generative adversarial network for video with a spatio-temporal convolutional architecture that untangles the scene's foreground from the background. Experiments suggest this model can generate tiny videos up to a second at full frame rate better than simple baselines, and we show its utility at predicting plausible futures of static images. Moreover, experiments and visualizations show the model internally learns useful features for recognizing actions with minimal supervision, suggesting scene dynamics are a promising signal for representation learning. We believe generative video models can impact many applications in video understanding and simulation.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
UCF-101 16 frames, 64x64, UnconditionalVGANInception Score8.18Unverified
UCF-101 16 frames, Unconditional, Single GPUVGANInception Score8.18Unverified

Reproductions