SOTAVerified

Ensure the Correctness of the Summary: Incorporate Entailment Knowledge into Abstractive Sentence Summarization

2018-08-01COLING 2018Unverified0· sign in to hype

Haoran Li, Junnan Zhu, Jiajun Zhang, Cheng-qing Zong

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this paper, we investigate the sentence summarization task that produces a summary from a source sentence. Neural sequence-to-sequence models have gained considerable success for this task, while most existing approaches only focus on improving the informativeness of the summary, which ignore the correctness, i.e., the summary should not contain unrelated information with respect to the source sentence. We argue that correctness is an essential requirement for summarization systems. Considering a correct summary is semantically entailed by the source sentence, we incorporate entailment knowledge into abstractive summarization models. We propose an entailment-aware encoder under multi-task framework (i.e., summarization generation and entailment recognition) and an entailment-aware decoder by entailment Reward Augmented Maximum Likelihood (RAML) training. Experiment results demonstrate that our models significantly outperform baselines from the aspects of informativeness and correctness.

Tasks

Reproductions