SOTAVerified

Fishr: Invariant Gradient Variances for Out-of-Distribution Generalization

2021-09-07Code Available1· sign in to hype

Alexandre Rame, Corentin Dancette, Matthieu Cord

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Learning robust models that generalize well under changes in the data distribution is critical for real-world applications. To this end, there has been a growing surge of interest to learn simultaneously from multiple training domains - while enforcing different types of invariance across those domains. Yet, all existing approaches fail to show systematic benefits under controlled evaluation protocols. In this paper, we introduce a new regularization - named Fishr - that enforces domain invariance in the space of the gradients of the loss: specifically, the domain-level variances of gradients are matched across training domains. Our approach is based on the close relations between the gradient covariance, the Fisher Information and the Hessian of the loss: in particular, we show that Fishr eventually aligns the domain-level loss landscapes locally around the final weights. Extensive experiments demonstrate the effectiveness of Fishr for out-of-distribution generalization. Notably, Fishr improves the state of the art on the DomainBed benchmark and performs consistently better than Empirical Risk Minimization. Our code is available at https://github.com/alexrame/fishr.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
DomainNetFishr (ResNet-50)Average Accuracy41.8Unverified
Office-HomeFishr (ResNet-50)Average Accuracy68.2Unverified
PACSFishr(ResNet-50,DomainBed)Average Accuracy86.9Unverified
TerraIncognitaFishr(ResNet-50)Average Accuracy47.4Unverified
VLCSFishr (ResNet-50)Average Accuracy78.2Unverified

Reproductions