SOTAVerified

Length Control in Abstractive Summarization by Pretraining Information Selection

2022-05-01ACL 2022Code Available0· sign in to hype

Yizhu Liu, Qi Jia, Kenny Zhu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Previous length-controllable summarization models mostly control lengths at the decoding stage, whereas the encoding or the selection of information from the source document is not sensitive to the designed length. They also tend to generate summaries as long as those in the training data. In this paper, we propose a length-aware attention mechanism (LAAM) to adapt the encoding of the source based on the desired length. Our approach works by training LAAM on a summary length balanced dataset built from the original training data, and then fine-tuning as usual. Results show that this approach is effective in generating high-quality summaries with desired lengths and even those short lengths never seen in the original training set.

Tasks

Reproductions