SOTAVerified

Pretrained models are active learners

2021-09-29Unverified0· sign in to hype

Alex Tamkin, Dat Nguyen, Salil Deshpande, Jesse Mu, Noah Goodman

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

An important barrier to the safe deployment of machine learning systems is the risk of task ambiguity, where multiple behaviors are consistent with the provided examples. We investigate whether pretrained models are better active learners, capable of asking for example labels that disambiguate between the possible tasks a user may be trying to specify. Across a range of image and text datasets with spurious correlations, latent minority groups, or domain shifts, finetuning pretrained models with data acquired through simple uncertainty sampling achieves the same accuracy with up to 6 fewer labels compared to random sampling. Moreover, the examples chosen by these models are preferentially minority classes or informative examples where the spurious feature and class label are decorrelated. Notably, gains from active learning are not seen in unpretrained models, which do not select such examples, suggesting that the ability to actively learn is an emergent property of the pretraining process.

Tasks

Reproductions