SOTAVerified

Label-Efficient Training for Next Response Selection

2020-11-01EMNLP (sustainlp) 2020Unverified0· sign in to hype

Seungtaek Choi, Myeongho Jeong, Jinyoung Yeo, Seung-won Hwang

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper studies label augmentation for training dialogue response selection. The existing model is trained by “observational” annotation, where one observed response is annotated as gold. In this paper, we propose “counterfactual augmentation” of pseudo-positive labels. We validate that the effectiveness of augmented labels are comparable to positives, such that ours outperform state-of-the-arts without augmentation.

Tasks

Reproductions