SOTAVerified

Do Deep Networks Transfer Invariances Across Classes?

2022-03-18ICLR 2022Code Available1· sign in to hype

Allan Zhou, Fahim Tajwar, Alexander Robey, Tom Knowles, George J. Pappas, Hamed Hassani, Chelsea Finn

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

To generalize well, classifiers must learn to be invariant to nuisance transformations that do not alter an input's class. Many problems have "class-agnostic" nuisance transformations that apply similarly to all classes, such as lighting and background changes for image classification. Neural networks can learn these invariances given sufficient data, but many real-world datasets are heavily class imbalanced and contain only a few examples for most of the classes. We therefore pose the question: how well do neural networks transfer class-agnostic invariances learned from the large classes to the small ones? Through careful experimentation, we observe that invariance to class-agnostic transformations is still heavily dependent on class size, with the networks being much less invariant on smaller classes. This result holds even when using data balancing techniques, and suggests poor invariance transfer across classes. Our results provide one explanation for why classifiers generalize poorly on unbalanced and long-tailed distributions. Based on this analysis, we show how a generative approach for learning the nuisance transformations can help transfer invariances across classes and improve performance on a set of imbalanced image classification benchmarks. Source code for our experiments is available at https://github.com/AllanYangZhou/generative-invariance-transfer.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-100-LT (ρ=100)CE+DRS+GITError Rate55.65Unverified
CIFAR-10-LT (ρ=100)CE+DRS+GITError Rate21.24Unverified

Reproductions