SOTAVerified

Perceptual Losses for Real-Time Style Transfer and Super-Resolution

2016-03-27Code Available1· sign in to hype

Justin Johnson, Alexandre Alahi, Li Fei-Fei

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We consider image transformation problems, where an input image is transformed into an output image. Recent methods for such problems typically train feed-forward convolutional neural networks using a per-pixel loss between the output and ground-truth images. Parallel work has shown that high-quality images can be generated by defining and optimizing perceptual loss functions based on high-level features extracted from pretrained networks. We combine the benefits of both approaches, and propose the use of perceptual loss functions for training feed-forward networks for image transformation tasks. We show results on image style transfer, where a feed-forward network is trained to solve the optimization problem proposed by Gatys et al in real-time. Compared to the optimization-based method, our network gives similar qualitative results but is three orders of magnitude faster. We also experiment with single-image super-resolution, where replacing a per-pixel loss with a perceptual loss gives visually pleasing results.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
BSD100 - 4x upscalingPerceptual LossPSNR24.95Unverified

Reproductions