SOTAVerified

Co-training 2^L Submodels for Visual Recognition

2022-12-09Unverified0· sign in to hype

Hugo Touvron, Matthieu Cord, Maxime Oquab, Piotr Bojanowski, Jakob Verbeek, Hervé Jégou

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We introduce submodel co-training, a regularization method related to co-training, self-distillation and stochastic depth. Given a neural network to be trained, for each sample we implicitly instantiate two altered networks, ``submodels'', with stochastic depth: we activate only a subset of the layers. Each network serves as a soft teacher to the other, by providing a loss that complements the regular loss provided by the one-hot label. Our approach, dubbed cosub, uses a single set of weights, and does not involve a pre-trained external model or temporal averaging. Experimentally, we show that submodel co-training is effective to train backbones for recognition tasks such as image classification and semantic segmentation. Our approach is compatible with multiple architectures, including RegNet, ViT, PiT, XCiT, Swin and ConvNext. Our training strategy improves their results in comparable settings. For instance, a ViT-B pretrained with cosub on ImageNet-21k obtains 87.4% top-1 acc. @448 on ImageNet-val.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ImageNetViT-H@224 (cosub)Top 1 Accuracy88Unverified
ImageNetViT-L@224 (cosub)Top 1 Accuracy87.5Unverified
ImageNetSwin-L@224 (cosub)Top 1 Accuracy87.1Unverified
ImageNetViT-B@224 (cosub)Top 1 Accuracy86.3Unverified
ImageNetSwin-B@224 (cosub)Top 1 Accuracy86.2Unverified
ImageNetConvNeXt-B@224 (cosub)Top 1 Accuracy85.8Unverified
ImageNetPiT-B@224 (cosub)Top 1 Accuracy85.8Unverified
ImageNetViT-M@224 (cosub)Top 1 Accuracy85Unverified
ImageNetRegnetY16GF@224 (cosub)Top 1 Accuracy84.2Unverified
ImageNetViT-S@224 (cosub)Top 1 Accuracy83.1Unverified

Reproductions