SOTAVerified

Adversarial Video Generation on Complex Datasets

2019-07-15Code Available0· sign in to hype

Aidan Clark, Jeff Donahue, Karen Simonyan

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Generative models of natural images have progressed towards high fidelity samples by the strong leveraging of scale. We attempt to carry this success to the field of video modeling by showing that large Generative Adversarial Networks trained on the complex Kinetics-600 dataset are able to produce video samples of substantially higher complexity and fidelity than previous work. Our proposed model, Dual Video Discriminator GAN (DVD-GAN), scales to longer and higher resolution videos by leveraging a computationally efficient decomposition of its discriminator. We evaluate on the related tasks of video synthesis and video prediction, and achieve new state-of-the-art Fr\'echet Inception Distance for prediction for Kinetics-600, as well as state-of-the-art Inception Score for synthesis on the UCF-101 dataset, alongside establishing a strong baseline for synthesis on Kinetics-600.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
BAIR Robot PushingDVD-GAN-FPFVD score109.8Unverified
Kinetics-600 12 frames, 128x128DVD-GANFID2.16Unverified
Kinetics-600 12 frames, 64x64DVD-GANFVD31.1Unverified
Kinetics-600 48 frames, 64x64DVD-GANFID12.92Unverified

Reproductions