Data-to-Text Generation with Content Selection and Planning
Ratish Puduppully, Li Dong, Mirella Lapata
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/ratishsp/data2text-plan-pyOfficialpytorch★ 0
- github.com/jugalw13/Red-Hat-Hacknone★ 0
Abstract
Recent advances in data-to-text generation have led to the use of large-scale datasets and neural network models which are trained end-to-end, without explicitly modeling what to say and in what order. In this work, we present a neural network architecture which incorporates content selection and planning without sacrificing end-to-end training. We decompose the generation task into two stages. Given a corpus of data records (paired with descriptive documents), we first generate a content plan highlighting which information should be mentioned and in which order and then generate the document while taking the content plan into account. Automatic and human-based evaluation experiments show that our model outperforms strong baselines improving the state-of-the-art on the recently released RotoWire dataset.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| RotoWire | Neural Content Planning + conditional copy | BLEU | 16.5 | — | Unverified |
| RotoWire (Content Ordering) | Neural Content Planning + conditional copy | DLD | 18.58 | — | Unverified |
| Rotowire (Content Selection) | Neural Content Planning + conditional copy | Precision | 34.18 | — | Unverified |
| RotoWire (Relation Generation) | Neural Content Planning + conditional copy | Precision | 87.47 | — | Unverified |