SOTAVerified

Zero-Shot Rationalization by Multi-Task Transfer Learning from Question Answering

2020-11-01Findings of the Association for Computational LinguisticsCode Available0· sign in to hype

Po-Nien Kung, Tse-Hsuan Yang, Yi-Cheng Chen, Sheng-Siang Yin, Yun-Nung Chen

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Extracting rationales can help human understand which information the model utilizes and how it makes the prediction towards better interpretability. However, annotating rationales requires much effort and only few datasets contain such labeled rationales, making supervised learning for rationalization difficult. In this paper, we propose a novel approach that leverages the benefits of both multi-task learning and transfer learning for generating rationales through question answering in a zero-shot fashion. For two benchmark rationalization datasets, the proposed method achieves comparable or even better performance of rationalization without any supervised signal, demonstrating the great potential of zero-shot rationalization for better interpretability.

Tasks

Reproductions