SOTAVerified

Uncertainty-Aware Bootstrap Learning for Joint Extraction on Distantly-Supervised Data

2023-05-05Code Available0· sign in to hype

Yufei Li, Xiao Yu, Yanchi Liu, Haifeng Chen, Cong Liu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Jointly extracting entity pairs and their relations is challenging when working on distantly-supervised data with ambiguous or noisy labels. To mitigate such impact, we propose uncertainty-aware bootstrap learning, which is motivated by the intuition that the higher uncertainty of an instance, the more likely the model confidence is inconsistent with the ground truths. Specifically, we first explore instance-level data uncertainty to create an initial high-confident examples. Such subset serves as filtering noisy instances and facilitating the model to converge fast at the early stage. During bootstrap learning, we propose self-ensembling as a regularizer to alleviate inter-model uncertainty produced by noisy labels. We further define probability variance of joint tagging probabilities to estimate inner-model parametric uncertainty, which is used to select and build up new reliable training instances for the next iteration. Experimental results on two large datasets reveal that our approach outperforms existing strong baselines and related methods.

Tasks

Reproductions