SOTAVerified

A Large-scale Study of Representation Learning with the Visual Task Adaptation Benchmark

2019-10-01arXiv 2020Code Available0· sign in to hype

Xiaohua Zhai, Joan Puigcerver, Alexander Kolesnikov, Pierre Ruyssen, Carlos Riquelme, Mario Lucic, Josip Djolonga, Andre Susano Pinto, Maxim Neumann, Alexey Dosovitskiy, Lucas Beyer, Olivier Bachem, Michael Tschannen, Marcin Michalski, Olivier Bousquet, Sylvain Gelly, Neil Houlsby

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Representation learning promises to unlock deep learning for the long tail of vision tasks without expensive labelled datasets. Yet, the absence of a unified evaluation for general visual representations hinders progress. Popular protocols are often too constrained (linear classification), limited in diversity (ImageNet, CIFAR, Pascal-VOC), or only weakly related to representation quality (ELBO, reconstruction error). We present the Visual Task Adaptation Benchmark (VTAB), which defines good representations as those that adapt to diverse, unseen tasks with few examples. With VTAB, we conduct a large-scale study of many popular publicly-available representation learning algorithms. We carefully control confounders such as architecture and tuning budget. We address questions like: How effective are ImageNet representations beyond standard natural datasets? How do representations trained via generative and discriminative models compare? To what extent can self-supervision replace labels? And, how close are we to general visual representations?

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
VTAB-1kS4L-Exemplar-ResNet50-LargeHyperSweepTop-1 Accuracy72.7Unverified
VTAB-1kS4L-Rotation-ResNet50-LargeHyperSweepTop-1 Accuracy71.5Unverified
VTAB-1kImageNet-ResNet50-LargeHyperSweepTop-1 Accuracy71.2Unverified
VTAB-1kS4L-Rotation-ResNet50Top-1 Accuracy67.5Unverified
VTAB-1kS4L-Exemplar-ResNet50Top-1 Accuracy67Unverified
VTAB-1kImageNet-ResNet50Top-1 Accuracy65.6Unverified
VTAB-1kS4L-10%-Rotation-ResNet50Top-1 Accuracy64.8Unverified
VTAB-1kS4L-10%-Exemplar-ResNet50Top-1 Accuracy63.9Unverified
VTAB-1kImageNet-10%-ResNet50Top-1 Accuracy61.6Unverified
VTAB-1kSelfSup-Rotation-ResNet50Top-1 Accuracy59.5Unverified
VTAB-1kResNet50-LargeHyperSweepTop-1 Accuracy59.2Unverified
VTAB-1kBigBiGAN-ResNet50Top-1 Accuracy59.1Unverified
VTAB-1kSelfSup-Exemplar-ResNet50Top-1 Accuracy57.5Unverified
VTAB-1kSelfSup-Jigsaw-ResNet50Top-1 Accuracy51.1Unverified
VTAB-1kSelfSup-RelativePatchLoc-ResNet50Top-1 Accuracy50.8Unverified
VTAB-1kUnconditional-BigGAN-ResNet50Top-1 Accuracy44Unverified
VTAB-1kResNet50Top-1 Accuracy42.1Unverified
VTAB-1kVAETop-1 Accuracy37.5Unverified
VTAB-1kWAE-MMDTop-1 Accuracy37.3Unverified
VTAB-1kConditional-BigGANTop-1 Accuracy35.3Unverified
VTAB-1kWAE-GANTop-1 Accuracy32Unverified
VTAB-1kWAE-UKLTop-1 Accuracy31Unverified

Reproductions