SOTAVerified

Domain Generalization by Mutual-Information Regularization with Pre-trained Models

2022-03-21Code Available1· sign in to hype

Junbum Cha, Kyungjae Lee, Sungrae Park, Sanghyuk Chun

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Domain generalization (DG) aims to learn a generalized model to an unseen target domain using only limited source domains. Previous attempts to DG fail to learn domain-invariant representations only from the source domains due to the significant domain shifts between training and test domains. Instead, we re-formulate the DG objective using mutual information with the oracle model, a model generalized to any possible domain. We derive a tractable variational lower bound via approximating the oracle model by a pre-trained model, called Mutual Information Regularization with Oracle (MIRO). Our extensive experiments show that MIRO significantly improves the out-of-distribution performance. Furthermore, our scaling experiments show that the larger the scale of the pre-trained model, the greater the performance improvement of MIRO. Source code is available at https://github.com/kakaobrain/miro.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
DomainNetMIRO (RegNetY-16GF, SWAD)Average Accuracy60.7Unverified
DomainNetMIRO (ResNet-50, SWAD)Average Accuracy47Unverified
Office-HomeMIRO (RegNetY-16GF, SWAD)Average Accuracy83.3Unverified
Office-HomeMIRO (ResNet-50, SWAD)Average Accuracy72.4Unverified
PACSMIRO (RegNetY-16GF, SWAD)Average Accuracy96.8Unverified
PACSMIRO (ResNet-50, SWAD)Average Accuracy88.4Unverified
TerraIncognitaMIRO (RegNetY-16GF, SWAD)Average Accuracy64.3Unverified
TerraIncognitaMIRO (ResNet-50, SWAD)Average Accuracy52.9Unverified
VLCSMIRO (RegNetY-16GF, SWAD)Average Accuracy81.7Unverified
VLCSMIRO (ResNet-50, SWAD)Average Accuracy79.6Unverified

Reproductions