Shortcut-Stacked Sentence Encoders for Multi-Domain Inference
Yixin Nie, Mohit Bansal
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/easonnie/multiNLI_encoderOfficialIn paperpytorch★ 0
- github.com/DorinK/Implementing-an-SNLI-Paperpytorch★ 0
- github.com/KhenAharon/Deep-Learning-SNLI-Residual-Stacked-Encoderspytorch★ 0
Abstract
We present a simple sequential sentence encoder for multi-domain natural language inference. Our encoder is based on stacked bidirectional LSTM-RNNs with shortcut connections and fine-tuning of word embeddings. The overall supervised model uses the above encoder to encode two input sentences into two vectors, and then uses a classifier over the vector combination to label the relationship between these two sentences as that of entailment, contradiction, or neural. Our Shortcut-Stacked sentence encoders achieve strong improvements over existing encoders on matched and mismatched multi-domain natural language inference (top non-ensemble single-model result in the EMNLP RepEval 2017 Shared Task (Nangia et al., 2017)). Moreover, they achieve the new state-of-the-art encoding result on the original SNLI dataset (Bowman et al., 2015).
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| SNLI | 600D Residual stacked encoders | % Test Accuracy | 86 | — | Unverified |
| SNLI | 300D Residual stacked encoders | % Test Accuracy | 85.7 | — | Unverified |