Data-to-text Generation with Variational Sequential Planning
Ratish Puduppully, Yao Fu, Mirella Lapata
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/ratishsp/data2text-seq-plan-pyOfficialIn paperpytorch★ 21
Abstract
We consider the task of data-to-text generation, which aims to create textual output from non-linguistic input. We focus on generating long-form text, i.e., documents with multiple paragraphs, and propose a neural model enhanced with a planning component responsible for organizing high-level information in a coherent and meaningful way. We infer latent plans sequentially with a structured variational model, while interleaving the steps of planning and generation. Text is generated by conditioning on previous variational decisions and previously generated text. Experiments on two data-to-text benchmarks (RotoWire and MLB) show that our model outperforms strong baselines and is sample efficient in the face of limited training data (e.g., a few hundred instances).
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| MLB Dataset | SeqPlan | BLEU | 14.29 | — | Unverified |
| MLB Dataset (Content Ordering) | SeqPlan | DLD | 22.7 | — | Unverified |
| MLB Dataset (Content Selection) | SeqPlan | Precision | 43.3 | — | Unverified |
| MLB Dataset (Relation Generation) | SeqPlan | Precision | 95.9 | — | Unverified |
| RotoWire (Relation Generation) | SeqPlan | Precision | 97.6 | — | Unverified |