Class-Balanced Loss Based on Effective Number of Samples
Yin Cui, Menglin Jia, Tsung-Yi Lin, Yang song, Serge Belongie
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/richardaecn/class-balanced-lossOfficialIn papertf★ 0
- github.com/frgfm/Holocronpytorch★ 328
- github.com/bazinga699/nclpytorch★ 86
- github.com/lijian16/fccpytorch★ 23
- github.com/feidfoe/AdjustBnd4Imbalancepytorch★ 19
- github.com/statsu1990/yoto_class_balanced_losspytorch★ 0
- github.com/MindSpore-scientific/code-3/tree/main/Class-balanced-loss-pytorch-mastermindspore★ 0
- github.com/tiagoCuervo/JapaNettf★ 0
- github.com/vandit15/Class-balanced-loss-pytorchpytorch★ 0
- github.com/MindCode-4/code-11/tree/main/Class-balanced-loss-pytorch-mastermindspore★ 0
Abstract
With the rapid increase of large-scale, real-world datasets, it becomes critical to address the problem of long-tailed data distribution (i.e., a few classes account for most of the data, while most classes are under-represented). Existing solutions typically adopt class re-balancing strategies such as re-sampling and re-weighting based on the number of observations for each class. In this work, we argue that as the number of samples increases, the additional benefit of a newly added data point will diminish. We introduce a novel theoretical framework to measure data overlap by associating with each sample a small neighboring region rather than a single point. The effective number of samples is defined as the volume of samples and can be calculated by a simple formula (1-^n)/(1-), where n is the number of samples and [0,1) is a hyperparameter. We design a re-weighting scheme that uses the effective number of samples for each class to re-balance the loss, thereby yielding a class-balanced loss. Comprehensive experiments are conducted on artificially induced long-tailed CIFAR datasets and large-scale datasets including ImageNet and iNaturalist. Our results show that when trained with the proposed class-balanced loss, the network is able to achieve significant performance gains on long-tailed datasets.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| iNaturalist 2018 | ResNet-101 | Top-1 Accuracy | 67.98 | — | Unverified |
| iNaturalist 2018 | ResNet-152 | Top-1 Accuracy | 69.05 | — | Unverified |
| iNaturalist 2018 | ResNet-152 | Top-1 Accuracy | 69.08 | — | Unverified |
| iNaturalist 2018 | ResNet-101 | Top-1 Accuracy | 68.39 | — | Unverified |