Control Prefixes for Parameter-Efficient Text Generation
Jordan Clive, Kris Cao, Marek Rei
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/Yale-LILY/dartOfficialIn papernone★ 157
- github.com/jordiclive/ControlPrefixesOfficialpytorch★ 90
Abstract
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to a downstream application. However, it uses the same dataset-level tuned prompt for all examples in the dataset. We extend this idea and propose a dynamic method, Control Prefixes, which allows for the inclusion of conditional input-dependent information, combining the benefits of prompt tuning and controlled generation. The method incorporates attribute-level learnable representations into different layers of a pre-trained transformer, allowing for the generated text to be guided in a particular direction. We provide a systematic evaluation of the technique and apply it to five datasets from the GEM benchmark for natural language generation (NLG). Although the aim is to develop a parameter-efficient model, we show Control Prefixes can even outperform full fine-tuning methods. We present state-of-the-art results on several data-to-text datasets, including WebNLG.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| Cleaned E2E NLG Challenge | Control Prefixes (T5-large) | BLEU (Test set) | 44.15 | — | Unverified |
| WebNLG | Control Prefixes (A1, T5-large) | BLEU | 67.32 | — | Unverified |
| WebNLG | Control Prefixes (A1, A2, T5-large) | BLEU | 67.15 | — | Unverified |
| WebNLG Full | Control Prefixes (A1, T5-large) | BLEU | 61.94 | — | Unverified |
| WebNLG Full | Control Prefixes (A1, A2, T5-large) | BLEU | 62.27 | — | Unverified |