Data-to-Text Generation
A classic problem in natural-language generation (NLG) involves taking structured data, such as a table, as input, and producing text that adequately and fluently describes this data as output. Unlike machine translation, which aims for complete transduction of the sentence to be translated, this form of NLG is usually taken to require addressing (at least) two separate challenges: what to say, the selection of an appropriate subset of the input data to discuss, and how to say it, the surface realization of a generation.
( Image credit: Data-to-Text Generation with Content Selection and Planning )
Papers
Showing 1–10 of 219 papers
All datasetsWebNLGE2E NLG ChallengeWebNLG FullCleaned E2E NLG ChallengeRotoWireRotoWire (Relation Generation)ToTToXAlignDARTMULTIWOZ 2.1RotoWire (Content Ordering)Rotowire (Content Selection)
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | Fact-aware embedding with mT5 | BLEU4 | 29.27 | — | Unverified |
| 2 | Bi-lingual mT5 | BLEU4 | 25.88 | — | Unverified |
| 3 | mT5 | BLEU4 | 25 | — | Unverified |
| 4 | Vanilla Transformer | BLEU4 | 19.9 | — | Unverified |
| 5 | Translate-Output mT5 | BLEU4 | 18.91 | — | Unverified |
| 6 | Graph Attention Network Encoder +Transformer Decoder | BLEU4 | 18.3 | — | Unverified |