Learning multiple visual domains with residual adapters
Sylvestre-Alvise Rebuffi, Hakan Bilen, Andrea Vedaldi
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/srebuffi/residual_adaptersOfficialpytorch★ 0
- github.com/YuWang24/MultiTunepytorch★ 0
Abstract
There is a growing interest in learning data representations that work well for many different types of problems and data. In this paper, we look in particular at the task of learning a single visual representation that can be successfully utilized in the analysis of very different types of images, from dog breeds to stop signs and digits. Inspired by recent work on learning networks that predict the parameters of another, we develop a tunable deep network architecture that, by means of adapter residual modules, can be steered on the fly to diverse visual domains. Our method achieves a high degree of parameter sharing while maintaining or even improving the accuracy of domain-specific representations. We also introduce the Visual Decathlon Challenge, a benchmark that evaluates the ability of representations to capture simultaneously ten very different visual domains and measures their ability to recognize well uniformly.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| visual domain decathlon (10 tasks) | Res. adapt. (large) | decathlon discipline (Score) | 3,131 | — | Unverified |
| visual domain decathlon (10 tasks) | Res. adapt. finetune all | decathlon discipline (Score) | 2,643 | — | Unverified |
| visual domain decathlon (10 tasks) | Res. adapt. decay | decathlon discipline (Score) | 2,621 | — | Unverified |
| visual domain decathlon (10 tasks) | Res. adapt. dom-pred | decathlon discipline (Score) | 2,503 | — | Unverified |
| visual domain decathlon (10 tasks) | Res. adapt. | decathlon discipline (Score) | 2,118 | — | Unverified |