SOTAVerified

EnhanceNet: Single Image Super-Resolution Through Automated Texture Synthesis

2016-12-23ICCV 2017Code Available0· sign in to hype

Mehdi S. M. Sajjadi, Bernhard Schölkopf, Michael Hirsch

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Single image super-resolution is the task of inferring a high-resolution image from a single low-resolution input. Traditionally, the performance of algorithms for this task is measured using pixel-wise reconstruction measures such as peak signal-to-noise ratio (PSNR) which have been shown to correlate poorly with the human perception of image quality. As a result, algorithms minimizing these metrics tend to produce over-smoothed images that lack high-frequency textures and do not look natural despite yielding high PSNR values. We propose a novel application of automated texture synthesis in combination with a perceptual loss focusing on creating realistic textures rather than optimizing for a pixel-accurate reproduction of ground truth images during training. By using feed-forward fully convolutional neural networks in an adversarial training setting, we achieve a significant boost in image quality at high magnification ratios. Extensive experiments on a number of datasets show the effectiveness of our approach, yielding state-of-the-art results in both quantitative and qualitative benchmarks.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
BSD100 - 4x upscalingENet-EPSNR27.5Unverified
FFHQ 1024 x 1024 - 4x upscalingEnhanceNetFID19.07Unverified
FFHQ 256 x 256 - 4x upscalingEnhanceNetFID116.38Unverified
Set14 - 4x upscalingENet-EPSNR28.42Unverified
Urban100 - 4x upscalingENet-EPSNR25.66Unverified

Reproductions