SOTAVerified

Trustworthy Long-Tailed Classification

2021-11-17CVPR 2022Code Available1· sign in to hype

Bolian Li, Zongbo Han, Haining Li, Huazhu Fu, Changqing Zhang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Classification on long-tailed distributed data is a challenging problem, which suffers from serious class-imbalance and accordingly unpromising performance especially on tail classes. Recently, the ensembling based methods achieve the state-of-the-art performance and show great potential. However, there are two limitations for current methods. First, their predictions are not trustworthy for failure-sensitive applications. This is especially harmful for the tail classes where the wrong predictions is basically frequent. Second, they assign unified numbers of experts to all samples, which is redundant for easy samples with excessive computational cost. To address these issues, we propose a Trustworthy Long-tailed Classification (TLC) method to jointly conduct classification and uncertainty estimation to identify hard samples in a multi-expert framework. Our TLC obtains the evidence-based uncertainty (EvU) and evidence for each expert, and then combines these uncertainties and evidences under the Dempster-Shafer Evidence Theory (DST). Moreover, we propose a dynamic expert engagement to reduce the number of engaged experts for easy samples and achieve efficiency while maintaining promising performances. Finally, we conduct comprehensive experiments on the tasks of classification, tail detection, OOD detection and failure prediction. The experimental results show that the proposed TLC outperforms existing methods and is trustworthy with reliable uncertainty.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-100-LT (ρ=100)TLC (4 Experts)Error Rate50.2Unverified
CIFAR-10-LT (ρ=100)TLC (4 experts)Error Rate19.6Unverified

Reproductions