First Train to Generate, then Generate to Train: UnitedSynT5 for Few-Shot NLI
Sourav Banerjee, Anush Mahajan, Ayushi Agarwal, Eishkaran Singh
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Natural Language Inference (NLI) tasks require identifying the relationship between sentence pairs, typically classified as entailment, contradiction, or neutrality. While the current state-of-the-art (SOTA) model, Entailment Few-Shot Learning (EFL), achieves a 93.1% accuracy on the Stanford Natural Language Inference (SNLI) dataset, further advancements are constrained by the dataset's limitations. To address this, we propose a novel approach leveraging synthetic data augmentation to enhance dataset diversity and complexity. We present UnitedSynT5, an advanced extension of EFL that leverages a T5-based generator to synthesize additional premise-hypothesis pairs, which are rigorously cleaned and integrated into the training data. These augmented examples are processed within the EFL framework, embedding labels directly into hypotheses for consistency. We train a GTR-T5-XL model on this expanded dataset, achieving a new benchmark of 94.7% accuracy on the SNLI dataset, 94.0% accuracy on the E-SNLI dataset, and 92.6% accuracy on the MultiNLI dataset, surpassing the previous SOTA models. This research demonstrates the potential of synthetic data augmentation in improving NLI models, offering a path forward for further advancements in natural language understanding tasks.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| e-SNLI | UnitedSynT5 (3B) | Accuracy | 94 | — | Unverified |
| e-SNLI | UnitedSynT5 (335M) | Accuracy | 89.8 | — | Unverified |
| MultiNLI | UnitedSynT5 (3B) | Matched | 92.6 | — | Unverified |
| MultiNLI | UnitedSynT5 (335M) | Matched | 89.8 | — | Unverified |
| SNLI | UnitedSynT5 (3B) | % Test Accuracy | 94.7 | — | Unverified |
| SNLI | UnitedSynT5 (335M) | % Test Accuracy | 93.5 | — | Unverified |